FAQ
Modified: hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_multiskew_1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_multiskew_1.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_multiskew_1.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_multiskew_1.q.out Sat Feb 21 02:36:54 2015
@@ -189,7 +189,7 @@ STAGE PLANS:
                          columns _col0
                          columns.types string
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -349,7 +349,7 @@ STAGE PLANS:
                          columns _col0,_col1
                          columns.types string:string
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -502,7 +502,7 @@ STAGE PLANS:
                          columns _col0
                          columns.types string
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -656,7 +656,7 @@ STAGE PLANS:
                          columns _col0,_col1
                          columns.types string:string
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_multiskew_2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_multiskew_2.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_multiskew_2.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_multiskew_2.q.out Sat Feb 21 02:36:54 2015
@@ -187,7 +187,7 @@ STAGE PLANS:
                          columns _col0,_col1
                          columns.types string:string
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -386,7 +386,7 @@ STAGE PLANS:
                          columns _col0
                          columns.types string
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -559,7 +559,7 @@ STAGE PLANS:
                          columns _col0,_col1
                          columns.types string:string
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_multiskew_3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_multiskew_3.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_multiskew_3.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_multiskew_3.q.out Sat Feb 21 02:36:54 2015
@@ -305,7 +305,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3
                          columns.types string:string:string:string
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -511,7 +511,7 @@ STAGE PLANS:
                    columns _col0
                    columns.types bigint
                    escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                    serialization.format 1
                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -614,7 +614,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3
                          columns.types string:string:string:string
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -768,7 +768,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3
                          columns.types string:string:string:string
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_oneskew_1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_oneskew_1.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_oneskew_1.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_oneskew_1.q.out Sat Feb 21 02:36:54 2015
@@ -242,7 +242,7 @@ STAGE PLANS:
                          columns _col0
                          columns.types int
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -389,7 +389,7 @@ STAGE PLANS:
                          columns _col0
                          columns.types int
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -532,7 +532,7 @@ STAGE PLANS:
                          columns _col0
                          columns.types int
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_oneskew_2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_oneskew_2.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_oneskew_2.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_oneskew_2.q.out Sat Feb 21 02:36:54 2015
@@ -259,7 +259,7 @@ STAGE PLANS:
                          columns _col0
                          columns.types int
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -426,7 +426,7 @@ STAGE PLANS:
                          columns _col0,_col1
                          columns.types int:string
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -653,7 +653,7 @@ STAGE PLANS:
                    columns _col0,_col1
                    columns.types string:bigint
                    escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                    serialization.format 1
                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -845,7 +845,7 @@ STAGE PLANS:
                    columns _col0,_col1
                    columns.types int:bigint
                    escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                    serialization.format 1
                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_oneskew_3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_oneskew_3.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_oneskew_3.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_query_oneskew_3.q.out Sat Feb 21 02:36:54 2015
@@ -270,7 +270,7 @@ STAGE PLANS:
                          columns _col0
                          columns.types int
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/llap/ql/src/test/results/clientpositive/load_dyn_part14.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/load_dyn_part14.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/load_dyn_part14.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/load_dyn_part14.q.out Sat Feb 21 02:36:54 2015
@@ -36,21 +36,21 @@ value string
  PREHOOK: query: explain
  insert overwrite table nzhang_part14 partition(value)
  select key, value from (
- select 'k1' as key, cast(null as string) as value from src limit 2
+ select * from (select 'k1' as key, cast(null as string) as value from src limit 2)a
    union all
- select 'k2' as key, '' as value from src limit 2
+ select * from (select 'k2' as key, '' as value from src limit 2)b
    union all
- select 'k3' as key, ' ' as value from src limit 2
+ select * from (select 'k3' as key, ' ' as value from src limit 2)c
  ) T
  PREHOOK: type: QUERY
  POSTHOOK: query: explain
  insert overwrite table nzhang_part14 partition(value)
  select key, value from (
- select 'k1' as key, cast(null as string) as value from src limit 2
+ select * from (select 'k1' as key, cast(null as string) as value from src limit 2)a
    union all
- select 'k2' as key, '' as value from src limit 2
+ select * from (select 'k2' as key, '' as value from src limit 2)b
    union all
- select 'k3' as key, ' ' as value from src limit 2
+ select * from (select 'k3' as key, ' ' as value from src limit 2)c
  ) T
  POSTHOOK: type: QUERY
  STAGE DEPENDENCIES:
@@ -256,22 +256,22 @@ STAGE PLANS:

  PREHOOK: query: insert overwrite table nzhang_part14 partition(value)
  select key, value from (
- select 'k1' as key, cast(null as string) as value from src limit 2
+ select * from (select 'k1' as key, cast(null as string) as value from src limit 2)a
    union all
- select 'k2' as key, '' as value from src limit 2
+ select * from (select 'k2' as key, '' as value from src limit 2)b
    union all
- select 'k3' as key, ' ' as value from src limit 2
+ select * from (select 'k3' as key, ' ' as value from src limit 2)c
  ) T
  PREHOOK: type: QUERY
  PREHOOK: Input: default@src
  PREHOOK: Output: default@nzhang_part14
  POSTHOOK: query: insert overwrite table nzhang_part14 partition(value)
  select key, value from (
- select 'k1' as key, cast(null as string) as value from src limit 2
+ select * from (select 'k1' as key, cast(null as string) as value from src limit 2)a
    union all
- select 'k2' as key, '' as value from src limit 2
+ select * from (select 'k2' as key, '' as value from src limit 2)b
    union all
- select 'k3' as key, ' ' as value from src limit 2
+ select * from (select 'k3' as key, ' ' as value from src limit 2)c
  ) T
  POSTHOOK: type: QUERY
  POSTHOOK: Input: default@src

Modified: hive/branches/llap/ql/src/test/results/clientpositive/louter_join_ppr.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/louter_join_ppr.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/louter_join_ppr.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/louter_join_ppr.q.out Sat Feb 21 02:36:54 2015
@@ -327,7 +327,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2,_col3
                        columns.types string:string:string:string
                        escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -803,7 +803,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2,_col3
                        columns.types string:string:string:string
                        escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1180,7 +1180,7 @@ STAGE PLANS:
                      columns _col0,_col1,_col2,_col3
                      columns.types string:string:string:string
                      escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                      serialization.format 1
                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1557,7 +1557,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2,_col3
                        columns.types string:string:string:string
                        escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/llap/ql/src/test/results/clientpositive/mapjoin_mapjoin.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/mapjoin_mapjoin.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/mapjoin_mapjoin.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/mapjoin_mapjoin.q.out Sat Feb 21 02:36:54 2015
@@ -159,7 +159,7 @@ STAGE PLANS:
                                columns _col0
                                columns.types string
                                escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/llap/ql/src/test/results/clientpositive/merge4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/merge4.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/merge4.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/merge4.q.out Sat Feb 21 02:36:54 2015
@@ -2762,34 +2762,51 @@ select * from (
      select '1' as key, '1' as value, 'file,' as hr from src limit 1) s
  POSTHOOK: type: QUERY
  STAGE DEPENDENCIES:
- Stage-9 is a root stage
- Stage-2 depends on stages: Stage-9
- Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6
- Stage-5
- Stage-0 depends on stages: Stage-5, Stage-4, Stage-7
- Stage-3 depends on stages: Stage-0
+ Stage-1 is a root stage
+ Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5
    Stage-4
- Stage-6
- Stage-7 depends on stages: Stage-6
+ Stage-0 depends on stages: Stage-4, Stage-3, Stage-6
+ Stage-2 depends on stages: Stage-0
+ Stage-3
+ Stage-5
+ Stage-6 depends on stages: Stage-5

  STAGE PLANS:
- Stage: Stage-9
+ Stage: Stage-1
      Map Reduce
        Map Operator Tree:
            TableScan
+ alias: srcpart
+ Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string), value (type: string), hr (type: string)
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE
+ Union
+ Statistics: Num rows: 1500 Data size: 140124 Basic stats: COMPLETE Column stats: COMPLETE
+ Limit
+ Number of rows: 1
+ Statistics: Num rows: 1 Data size: 259 Basic stats: COMPLETE Column stats: COMPLETE
+ Reduce Output Operator
+ sort order:
+ Statistics: Num rows: 1 Data size: 259 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string)
+ TableScan
              alias: src
              Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE
              Select Operator
                expressions: '1' (type: string), '1' (type: string), 'file,' (type: string)
                outputColumnNames: _col0, _col1, _col2
                Statistics: Num rows: 500 Data size: 129500 Basic stats: COMPLETE Column stats: COMPLETE
- Limit
- Number of rows: 1
- Statistics: Num rows: 1 Data size: 259 Basic stats: COMPLETE Column stats: COMPLETE
- Reduce Output Operator
- sort order:
+ Union
+ Statistics: Num rows: 1500 Data size: 140124 Basic stats: COMPLETE Column stats: COMPLETE
+ Limit
+ Number of rows: 1
                    Statistics: Num rows: 1 Data size: 259 Basic stats: COMPLETE Column stats: COMPLETE
- value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string)
+ Reduce Output Operator
+ sort order:
+ Statistics: Num rows: 1 Data size: 259 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string)
        Reduce Operator Tree:
          Select Operator
            expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), VALUE._col2 (type: string)
@@ -2800,47 +2817,17 @@ STAGE PLANS:
              Statistics: Num rows: 1 Data size: 259 Basic stats: COMPLETE Column stats: COMPLETE
              File Output Operator
                compressed: false
+ Statistics: Num rows: 1 Data size: 259 Basic stats: COMPLETE Column stats: COMPLETE
                table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
-
- Stage: Stage-2
- Map Reduce
- Map Operator Tree:
- TableScan
- alias: srcpart
- Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: key (type: string), value (type: string), hr (type: string)
- outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE
- Union
- Statistics: Num rows: 1001 Data size: 10883 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- Statistics: Num rows: 1001 Data size: 10883 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.nzhang_part
- TableScan
- Union
- Statistics: Num rows: 1001 Data size: 10883 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- Statistics: Num rows: 1001 Data size: 10883 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.nzhang_part
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.nzhang_part

- Stage: Stage-8
+ Stage: Stage-7
      Conditional Operator

- Stage: Stage-5
+ Stage: Stage-4
      Move Operator
        files:
            hdfs directory: true
@@ -2859,10 +2846,10 @@ STAGE PLANS:
                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                name: default.nzhang_part

- Stage: Stage-3
+ Stage: Stage-2
      Stats-Aggr Operator

- Stage: Stage-4
+ Stage: Stage-3
      Map Reduce
        Map Operator Tree:
            TableScan
@@ -2874,7 +2861,7 @@ STAGE PLANS:
                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                    name: default.nzhang_part

- Stage: Stage-6
+ Stage: Stage-5
      Map Reduce
        Map Operator Tree:
            TableScan
@@ -2886,7 +2873,7 @@ STAGE PLANS:
                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                    name: default.nzhang_part

- Stage: Stage-7
+ Stage: Stage-6
      Move Operator
        files:
            hdfs directory: true
@@ -2913,13 +2900,7 @@ POSTHOOK: Input: default@src
  POSTHOOK: Input: default@srcpart
  POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
  POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Output: default@nzhang_part@ds=2010-08-15/hr=11
-POSTHOOK: Output: default@nzhang_part@ds=2010-08-15/hr=12
  POSTHOOK: Output: default@nzhang_part@ds=2010-08-15/hr=file,
-POSTHOOK: Lineage: nzhang_part PARTITION(ds=2010-08-15,hr=11).key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: nzhang_part PARTITION(ds=2010-08-15,hr=11).value EXPRESSION [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: nzhang_part PARTITION(ds=2010-08-15,hr=12).key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: nzhang_part PARTITION(ds=2010-08-15,hr=12).value EXPRESSION [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
  POSTHOOK: Lineage: nzhang_part PARTITION(ds=2010-08-15,hr=file,).key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
  POSTHOOK: Lineage: nzhang_part PARTITION(ds=2010-08-15,hr=file,).value EXPRESSION [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
  PREHOOK: query: show partitions nzhang_part

Modified: hive/branches/llap/ql/src/test/results/clientpositive/metadataOnlyOptimizer.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/metadataOnlyOptimizer.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/metadataOnlyOptimizer.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/metadataOnlyOptimizer.q.out Sat Feb 21 02:36:54 2015
@@ -1332,24 +1332,23 @@ POSTHOOK: Input: default@src
  96
  97
  98
-PREHOOK: query: select k from (SELECT '1' as k from src limit 0 union all select key as k from src limit 1)tab
+PREHOOK: query: select k from (select * from (SELECT '1' as k from src limit 0)a union all select key as k from src limit 1)tab
  PREHOOK: type: QUERY
  PREHOOK: Input: default@src
  #### A masked pattern was here ####
-POSTHOOK: query: select k from (SELECT '1' as k from src limit 0 union all select key as k from src limit 1)tab
+POSTHOOK: query: select k from (select * from (SELECT '1' as k from src limit 0)a union all select key as k from src limit 1)tab
  POSTHOOK: type: QUERY
  POSTHOOK: Input: default@src
  #### A masked pattern was here ####
  238
-PREHOOK: query: select k from (SELECT '1' as k from src limit 1 union all select key as k from src limit 0)tab
+PREHOOK: query: select k from (select * from (SELECT '1' as k from src limit 1)a union all select key as k from src limit 0)tab
  PREHOOK: type: QUERY
  PREHOOK: Input: default@src
  #### A masked pattern was here ####
-POSTHOOK: query: select k from (SELECT '1' as k from src limit 1 union all select key as k from src limit 0)tab
+POSTHOOK: query: select k from (select * from (SELECT '1' as k from src limit 1)a union all select key as k from src limit 0)tab
  POSTHOOK: type: QUERY
  POSTHOOK: Input: default@src
  #### A masked pattern was here ####
-1
  PREHOOK: query: select max(ds) from srcpart
  PREHOOK: type: QUERY
  PREHOOK: Input: default@srcpart

Modified: hive/branches/llap/ql/src/test/results/clientpositive/metadataonly1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/metadataonly1.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/metadataonly1.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/metadataonly1.q.out Sat Feb 21 02:36:54 2015
@@ -57,7 +57,7 @@ STAGE PLANS:
                    columns _col0
                    columns.types string
                    escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                    serialization.format 1
                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -202,7 +202,7 @@ STAGE PLANS:
                    columns _col0
                    columns.types string
                    escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                    serialization.format 1
                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -343,7 +343,7 @@ STAGE PLANS:
                    columns _col0
                    columns.types bigint
                    escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                    serialization.format 1
                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -484,7 +484,7 @@ STAGE PLANS:
                    columns _col0
                    columns.types bigint
                    escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                    serialization.format 1
                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -934,7 +934,7 @@ STAGE PLANS:
                    columns _col0
                    columns.types bigint
                    escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                    serialization.format 1
                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1199,7 +1199,7 @@ STAGE PLANS:
                    columns _col0,_col1
                    columns.types string:bigint
                    escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                    serialization.format 1
                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1439,7 +1439,7 @@ STAGE PLANS:
                    columns _col0,_col1
                    columns.types string:bigint
                    escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                    serialization.format 1
                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1623,7 +1623,7 @@ STAGE PLANS:
                    columns _col0
                    columns.types string
                    escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                    serialization.format 1
                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1994,7 +1994,7 @@ STAGE PLANS:
                    columns _col0,_col1
                    columns.types string:bigint
                    escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                    serialization.format 1
                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/llap/ql/src/test/results/clientpositive/optimize_nullscan.q.out
URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/optimize_nullscan.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
==============================================================================
--- hive/branches/llap/ql/src/test/results/clientpositive/optimize_nullscan.q.out (original)
+++ hive/branches/llap/ql/src/test/results/clientpositive/optimize_nullscan.q.out Sat Feb 21 02:36:54 2015
@@ -67,7 +67,7 @@ STAGE PLANS:
                          columns _col0
                          columns.types string
                          escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -202,7 +202,7 @@ STAGE PLANS:
                      columns _col0
                      columns.types bigint
                      escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                      serialization.format 1
                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -656,7 +656,7 @@ STAGE PLANS:
                    columns _col0,_col1
                    columns.types string:string
                    escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                    serialization.format 1
                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -699,7 +699,7 @@ ABSTRACT SYNTAX TREE:
  TOK_QUERY
     TOK_FROM
        TOK_SUBQUERY
- TOK_UNION
+ TOK_UNIONALL
              TOK_QUERY
                 TOK_FROM
                    TOK_TABREF
@@ -865,7 +865,7 @@ STAGE PLANS:
                        columns _col0
                        columns.types bigint
                        escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -890,7 +890,7 @@ STAGE PLANS:
                        columns _col0
                        columns.types bigint
                        escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1630,7 +1630,7 @@ STAGE PLANS:
                    columns _col0,_col1
                    columns.types string:string
                    escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                    serialization.format 1
                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1674,7 +1674,7 @@ ABSTRACT SYNTAX TREE:
  TOK_QUERY
     TOK_FROM
        TOK_SUBQUERY
- TOK_UNION
+ TOK_UNIONALL
              TOK_QUERY
                 TOK_FROM
                    TOK_TABREF
@@ -1849,7 +1849,7 @@ STAGE PLANS:
                        columns _col0
                        columns.types string
                        escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1884,7 +1884,7 @@ STAGE PLANS:
                            columns _col0
                            columns.types string
                            escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                            serialization.format 1
                            serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -2134,7 +2134,7 @@ STAGE PLANS:
                      columns _col0,_col1,_col2,_col3
                      columns.types string:string:string:string
                      escape.delim \
- hive.serialization.extend.nesting.levels true
+ hive.serialization.extend.additional.nesting.levels true
                      serialization.format 1
                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Search Discussions

  • Prasanthj at Feb 21, 2015 at 2:37 am
    Modified: hive/branches/llap/ql/src/test/results/clientpositive/bucketmapjoin7.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/bucketmapjoin7.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/bucketmapjoin7.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/bucketmapjoin7.q.out Sat Feb 21 02:36:54 2015
    @@ -330,7 +330,7 @@ STAGE PLANS:
                          columns _col0,_col1
                          columns.types int:string
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/bucketmapjoin8.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/bucketmapjoin8.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/bucketmapjoin8.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/bucketmapjoin8.q.out Sat Feb 21 02:36:54 2015
    @@ -312,7 +312,7 @@ STAGE PLANS:
                        columns _col0
                        columns.types bigint
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -605,7 +605,7 @@ STAGE PLANS:
                        columns _col0
                        columns.types bigint
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/bucketmapjoin9.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/bucketmapjoin9.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/bucketmapjoin9.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/bucketmapjoin9.q.out Sat Feb 21 02:36:54 2015
    @@ -312,7 +312,7 @@ STAGE PLANS:
                        columns _col0
                        columns.types bigint
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -630,7 +630,7 @@ STAGE PLANS:
                        columns _col0
                        columns.types bigint
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out Sat Feb 21 02:36:54 2015
    @@ -268,7 +268,7 @@ STAGE PLANS:
                                columns _col0,_col1,_col2,_col3
                                columns.types string:string:string:string
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -457,7 +457,7 @@ STAGE PLANS:
                                columns _col0,_col1,_col2,_col3
                                columns.types string:string:string:string
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -635,7 +635,7 @@ STAGE PLANS:
                                columns _col0,_col1,_col2,_col3
                                columns.types string:string:string:string
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -816,7 +816,7 @@ STAGE PLANS:
                                columns _col0,_col1,_col2,_col3
                                columns.types string:string:string:string
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -997,7 +997,7 @@ STAGE PLANS:
                                columns _col0,_col1,_col2,_col3
                                columns.types string:string:string:string
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1178,7 +1178,7 @@ STAGE PLANS:
                                columns _col0,_col1,_col2,_col3
                                columns.types string:string:string:string
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1359,7 +1359,7 @@ STAGE PLANS:
                                columns _col0,_col1,_col2,_col3
                                columns.types string:string:string:string
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1540,7 +1540,7 @@ STAGE PLANS:
                                columns _col0,_col1,_col2,_col3
                                columns.types string:string:string:string
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1721,7 +1721,7 @@ STAGE PLANS:
                                columns _col0,_col1,_col2,_col3
                                columns.types string:string:string:string
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/cbo_union.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/cbo_union.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/cbo_union.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/cbo_union.q.out Sat Feb 21 02:36:54 2015
    @@ -1,7 +1,7 @@
      PREHOOK: query: -- SORT_QUERY_RESULTS

      -- 11. Union All
    -select * from cbo_t1 order by key, c_boolean, value, dt union all select * from cbo_t2 order by key, c_boolean, value, dt
    +select * from (select * from cbo_t1 order by key, c_boolean, value, dt)a union all select * from (select * from cbo_t2 order by key, c_boolean, value, dt)b
      PREHOOK: type: QUERY
      PREHOOK: Input: default@cbo_t1
      PREHOOK: Input: default@cbo_t1@dt=2014
    @@ -11,7 +11,7 @@ PREHOOK: Input: default@cbo_t2@dt=2014
      POSTHOOK: query: -- SORT_QUERY_RESULTS

      -- 11. Union All
    -select * from cbo_t1 order by key, c_boolean, value, dt union all select * from cbo_t2 order by key, c_boolean, value, dt
    +select * from (select * from cbo_t1 order by key, c_boolean, value, dt)a union all select * from (select * from cbo_t2 order by key, c_boolean, value, dt)b
      POSTHOOK: type: QUERY
      POSTHOOK: Input: default@cbo_t1
      POSTHOOK: Input: default@cbo_t1@dt=2014

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/char_union1.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/char_union1.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/char_union1.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/char_union1.q.out Sat Feb 21 02:36:54 2015
    @@ -95,9 +95,6 @@ POSTHOOK: type: QUERY
      POSTHOOK: Input: default@char_union1_ch1
      #### A masked pattern was here ####
      1 abc
    -1 abc
    -2 abc
    -3 abc
      PREHOOK: query: -- union char with different length char
      select * from (
        select * from char_union1_ch1
    @@ -119,9 +116,6 @@ POSTHOOK: Input: default@char_union1_ch1
      POSTHOOK: Input: default@char_union1_ch2
      #### A masked pattern was here ####
      1 abc
    -1 abc
    -2 abc
    -3 abc
      PREHOOK: query: -- union char with string
      select * from (
        select * from char_union1_ch1
    @@ -143,9 +137,6 @@ POSTHOOK: Input: default@char_union1_ch1
      POSTHOOK: Input: default@char_union1_str
      #### A masked pattern was here ####
      1 abc
    -1 abc
    -2 abc
    -3 abc
      PREHOOK: query: drop table char_union1_ch1
      PREHOOK: type: DROPTABLE
      PREHOOK: Input: default@char_union1_ch1

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/columnstats_partlvl.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/columnstats_partlvl.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/columnstats_partlvl.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/columnstats_partlvl.q.out Sat Feb 21 02:36:54 2015
    @@ -204,7 +204,7 @@ STAGE PLANS:
                          columns _col0,_col1
                          columns.types struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint>:double
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -404,7 +404,7 @@ STAGE PLANS:
                          columns _col0,_col1
                          columns.types struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint>:double
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/columnstats_tbllvl.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/columnstats_tbllvl.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/columnstats_tbllvl.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/columnstats_tbllvl.q.out Sat Feb 21 02:36:54 2015
    @@ -191,7 +191,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2
                        columns.types struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint>:struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint>:struct<columntype:string,min:double,max:double,countnulls:bigint,numdistinctvalues:bigint>
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/combine2.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/combine2.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/combine2.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/combine2.q.out Sat Feb 21 02:36:54 2015
    @@ -593,7 +593,7 @@ STAGE PLANS:
                        columns _col0
                        columns.types bigint
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/combine2_hadoop20.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/combine2_hadoop20.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/combine2_hadoop20.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/combine2_hadoop20.q.out Sat Feb 21 02:36:54 2015
    @@ -606,7 +606,7 @@ STAGE PLANS:
                          columns _col0
                          columns.types bigint
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/combine2_win.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/combine2_win.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/combine2_win.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/combine2_win.q.out Sat Feb 21 02:36:54 2015
    @@ -603,7 +603,7 @@ STAGE PLANS:
                          columns _col0
                          columns.types bigint
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                    TotalFiles: 1
                    GatherStats: false

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/constantPropagateForSubQuery.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/constantPropagateForSubQuery.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/constantPropagateForSubQuery.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/constantPropagateForSubQuery.q.out Sat Feb 21 02:36:54 2015
    @@ -233,7 +233,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3
                          columns.types string:string:string:string
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/correlationoptimizer12.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/correlationoptimizer12.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/correlationoptimizer12.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/correlationoptimizer12.q.out Sat Feb 21 02:36:54 2015
    @@ -32,8 +32,8 @@ STAGE PLANS:
                    outputColumnNames: _col0, _col1
                    Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
                    Reduce Output Operator
    - key expressions: _col0 (type: string), _col0 (type: string)
    - sort order: ++
    + key expressions: _col0 (type: string)
    + sort order: +
                      Map-reduce partition columns: _col0 (type: string)
                      Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
                      value expressions: _col1 (type: string)
    @@ -43,6 +43,24 @@ STAGE PLANS:
                outputColumnNames: _col0, _col1
                Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
                PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col0: string, _col1: string
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col0
    + partition by: _col0
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col1
    + name: count
    + window function: GenericUDAFCountEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
                  Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
                  Filter Operator
                    predicate: _col0 is not null (type: boolean)
    @@ -103,8 +121,8 @@ STAGE PLANS:
                    outputColumnNames: _col0, _col1
                    Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
                    Reduce Output Operator
    - key expressions: _col0 (type: string), _col0 (type: string)
    - sort order: ++
    + key expressions: _col0 (type: string)
    + sort order: +
                      Map-reduce partition columns: _col0 (type: string)
                      Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
                      value expressions: _col1 (type: string)
    @@ -114,6 +132,24 @@ STAGE PLANS:
                outputColumnNames: _col0, _col1
                Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
                PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col0: string, _col1: string
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col0
    + partition by: _col0
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col1
    + name: count
    + window function: GenericUDAFCountEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
                  Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
                  Filter Operator
                    predicate: _col0 is not null (type: boolean)

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/ctas_colname.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/ctas_colname.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/ctas_colname.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/ctas_colname.q.out Sat Feb 21 02:36:54 2015
    @@ -189,6 +189,25 @@ STAGE PLANS:
                outputColumnNames: _col0, _col1
                Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
                PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col0: string, _col1: string
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: _col0
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col1
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
                  Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
                  Select Operator
                    expressions: _col0 (type: string), _col1 (type: string), _wcol0 (type: int)
    @@ -336,6 +355,25 @@ STAGE PLANS:
                outputColumnNames: _col0, _col1
                Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
                PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col0: string, _col1: string
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: _col0
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col0, 1
    + name: lead
    + window function: GenericUDAFLeadEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
                  Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
                  Select Operator
                    expressions: _col0 (type: string), _col1 (type: string), _wcol0 (type: string)

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/display_colstats_tbllvl.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/display_colstats_tbllvl.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/display_colstats_tbllvl.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/display_colstats_tbllvl.q.out Sat Feb 21 02:36:54 2015
    @@ -207,7 +207,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2
                        columns.types struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint>:struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint>:struct<columntype:string,min:double,max:double,countnulls:bigint,numdistinctvalues:bigint>
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/encrypted/encryption_join_unencrypted_tbl.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/encrypted/encryption_join_unencrypted_tbl.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/encrypted/encryption_join_unencrypted_tbl.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/encrypted/encryption_join_unencrypted_tbl.q.out Sat Feb 21 02:36:54 2015
    @@ -738,7 +738,7 @@ STAGE PLANS:
                            columns _col0,_col1,_col2,_col3
                            columns.types string:string:int:string
                            escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                            serialization.format 1
                            serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/encrypted/encryption_join_with_different_encryption_keys.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/encrypted/encryption_join_with_different_encryption_keys.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/encrypted/encryption_join_with_different_encryption_keys.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/encrypted/encryption_join_with_different_encryption_keys.q.out Sat Feb 21 02:36:54 2015
    @@ -258,7 +258,7 @@ STAGE PLANS:
                            columns _col0,_col1,_col2,_col3
                            columns.types int:string:int:string
                            escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                            serialization.format 1
                            serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/explain_logical.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/explain_logical.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/explain_logical.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/explain_logical.q.out Sat Feb 21 02:36:54 2015
    @@ -229,7 +229,7 @@ ABSTRACT SYNTAX TREE:
      TOK_QUERY
         TOK_FROM
            TOK_SUBQUERY
    - TOK_UNION
    + TOK_UNIONALL
                  TOK_QUERY
                     TOK_FROM
                        TOK_TABREF

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/explode_null.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/explode_null.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/explode_null.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/explode_null.q.out Sat Feb 21 02:36:54 2015
    @@ -1,14 +1,14 @@
      PREHOOK: query: SELECT explode(col) AS myCol FROM
    - (SELECT array(1,2,3) AS col FROM src LIMIT 1
    + (select * from (SELECT array(1,2,3) AS col FROM src LIMIT 1)ea
           UNION ALL
    - SELECT IF(false, array(1,2,3), NULL) AS col FROM src LIMIT 1) a
    + select * from (SELECT IF(false, array(1,2,3), NULL) AS col FROM src LIMIT 1)eb) a
      PREHOOK: type: QUERY
      PREHOOK: Input: default@src
      #### A masked pattern was here ####
      POSTHOOK: query: SELECT explode(col) AS myCol FROM
    - (SELECT array(1,2,3) AS col FROM src LIMIT 1
    + (select * from (SELECT array(1,2,3) AS col FROM src LIMIT 1)ea
           UNION ALL
    - SELECT IF(false, array(1,2,3), NULL) AS col FROM src LIMIT 1) a
    + select * from (SELECT IF(false, array(1,2,3), NULL) AS col FROM src LIMIT 1)eb) a
      POSTHOOK: type: QUERY
      POSTHOOK: Input: default@src
      #### A masked pattern was here ####
    @@ -16,16 +16,16 @@ POSTHOOK: Input: default@src
      2
      3
      PREHOOK: query: SELECT explode(col) AS (myCol1,myCol2) FROM
    - (SELECT map(1,'one',2,'two',3,'three') AS col FROM src LIMIT 1
    + (select * from (SELECT map(1,'one',2,'two',3,'three') AS col FROM src LIMIT 1)ea
           UNION ALL
    - SELECT IF(false, map(1,'one',2,'two',3,'three'), NULL) AS col FROM src LIMIT 1) a
    + select * from (SELECT IF(false, map(1,'one',2,'two',3,'three'), NULL) AS col FROM src LIMIT 1)eb ) a
      PREHOOK: type: QUERY
      PREHOOK: Input: default@src
      #### A masked pattern was here ####
      POSTHOOK: query: SELECT explode(col) AS (myCol1,myCol2) FROM
    - (SELECT map(1,'one',2,'two',3,'three') AS col FROM src LIMIT 1
    + (select * from (SELECT map(1,'one',2,'two',3,'three') AS col FROM src LIMIT 1)ea
           UNION ALL
    - SELECT IF(false, map(1,'one',2,'two',3,'three'), NULL) AS col FROM src LIMIT 1) a
    + select * from (SELECT IF(false, map(1,'one',2,'two',3,'three'), NULL) AS col FROM src LIMIT 1)eb ) a
      POSTHOOK: type: QUERY
      POSTHOOK: Input: default@src
      #### A masked pattern was here ####

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/filter_join_breaktask.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/filter_join_breaktask.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/filter_join_breaktask.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/filter_join_breaktask.q.out Sat Feb 21 02:36:54 2015
    @@ -379,7 +379,7 @@ STAGE PLANS:
                          columns _col0,_col1
                          columns.types int:string
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/groupby3_map_skew.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/groupby3_map_skew.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/groupby3_map_skew.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/groupby3_map_skew.q.out Sat Feb 21 02:36:54 2015
    @@ -151,12 +151,14 @@ POSTHOOK: Lineage: dest1.c6 EXPRESSION [
      POSTHOOK: Lineage: dest1.c7 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
      POSTHOOK: Lineage: dest1.c8 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
      POSTHOOK: Lineage: dest1.c9 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
    -PREHOOK: query: SELECT c1, c2, c3, c4, c5, c6, c7, ROUND(c8, 5), ROUND(c9, 5) FROM dest1
    +PREHOOK: query: SELECT ROUND(c1, 1), ROUND(c2, 3), ROUND(c3, 5), ROUND(c4, 1), ROUND(c5, 1), ROUND(c6, 5),
    +ROUND(c7,5), ROUND(c8, 5), ROUND(c9, 5) FROM dest1
      PREHOOK: type: QUERY
      PREHOOK: Input: default@dest1
      #### A masked pattern was here ####
    -POSTHOOK: query: SELECT c1, c2, c3, c4, c5, c6, c7, ROUND(c8, 5), ROUND(c9, 5) FROM dest1
    +POSTHOOK: query: SELECT ROUND(c1, 1), ROUND(c2, 3), ROUND(c3, 5), ROUND(c4, 1), ROUND(c5, 1), ROUND(c6, 5),
    +ROUND(c7,5), ROUND(c8, 5), ROUND(c9, 5) FROM dest1
      POSTHOOK: type: QUERY
      POSTHOOK: Input: default@dest1
      #### A masked pattern was here ####
    -130091.0 260.182 256.10355987055016 98.0 0.0 142.9268095075238 143.06995106518906 20428.07288 20469.0109
    +130091.0 260.182 256.10356 98.0 0.0 142.92681 143.06995 20428.07288 20469.0109

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/groupby7_noskew_multi_single_reducer.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/groupby7_noskew_multi_single_reducer.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/groupby7_noskew_multi_single_reducer.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/groupby7_noskew_multi_single_reducer.q.out Sat Feb 21 02:36:54 2015
    @@ -20,13 +20,13 @@ POSTHOOK: Output: database:default
      POSTHOOK: Output: default@DEST2
      PREHOOK: query: EXPLAIN
      FROM SRC
    -INSERT OVERWRITE TABLE DEST1 SELECT SRC.key, sum(SUBSTR(SRC.value,5)) GROUP BY SRC.key limit 10
    -INSERT OVERWRITE TABLE DEST2 SELECT SRC.key, sum(SUBSTR(SRC.value,5)) GROUP BY SRC.key limit 10
    +INSERT OVERWRITE TABLE DEST1 SELECT SRC.key, sum(SUBSTR(SRC.value,5)) GROUP BY SRC.key ORDER BY SRC.key limit 10
    +INSERT OVERWRITE TABLE DEST2 SELECT SRC.key, sum(SUBSTR(SRC.value,5)) GROUP BY SRC.key ORDER BY SRC.key limit 10
      PREHOOK: type: QUERY
      POSTHOOK: query: EXPLAIN
      FROM SRC
    -INSERT OVERWRITE TABLE DEST1 SELECT SRC.key, sum(SUBSTR(SRC.value,5)) GROUP BY SRC.key limit 10
    -INSERT OVERWRITE TABLE DEST2 SELECT SRC.key, sum(SUBSTR(SRC.value,5)) GROUP BY SRC.key limit 10
    +INSERT OVERWRITE TABLE DEST1 SELECT SRC.key, sum(SUBSTR(SRC.value,5)) GROUP BY SRC.key ORDER BY SRC.key limit 10
    +INSERT OVERWRITE TABLE DEST2 SELECT SRC.key, sum(SUBSTR(SRC.value,5)) GROUP BY SRC.key ORDER BY SRC.key limit 10
      POSTHOOK: type: QUERY
      STAGE DEPENDENCIES:
        Stage-2 is a root stage
    @@ -63,44 +63,39 @@ STAGE PLANS:
                  mode: complete
                  outputColumnNames: _col0, _col1
                  Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
    - Limit
    - Number of rows: 10
    - Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
    - File Output Operator
    - compressed: true
    - table:
    - input format: org.apache.hadoop.mapred.SequenceFileInputFormat
    - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
    - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
    + File Output Operator
    + compressed: true
    + table:
    + input format: org.apache.hadoop.mapred.SequenceFileInputFormat
    + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
    + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
                Group By Operator
                  aggregations: sum(VALUE._col0)
                  keys: KEY._col0 (type: string)
                  mode: complete
                  outputColumnNames: _col0, _col1
                  Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
    - Limit
    - Number of rows: 10
    - Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
    - File Output Operator
    - compressed: true
    - table:
    - input format: org.apache.hadoop.mapred.SequenceFileInputFormat
    - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
    - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
    + File Output Operator
    + compressed: true
    + table:
    + input format: org.apache.hadoop.mapred.SequenceFileInputFormat
    + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
    + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe

        Stage: Stage-3
          Map Reduce
            Map Operator Tree:
                TableScan
                  Reduce Output Operator
    - sort order:
    - Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
    - value expressions: _col0 (type: string), _col1 (type: double)
    + key expressions: _col0 (type: string)
    + sort order: +
    + Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
    + value expressions: _col1 (type: double)
            Reduce Operator Tree:
              Select Operator
    - expressions: VALUE._col0 (type: string), VALUE._col1 (type: double)
    + expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: double)
                outputColumnNames: _col0, _col1
    - Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
                Limit
                  Number of rows: 10
                  Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
    @@ -135,14 +130,15 @@ STAGE PLANS:
            Map Operator Tree:
                TableScan
                  Reduce Output Operator
    - sort order:
    - Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
    - value expressions: _col0 (type: string), _col1 (type: double)
    + key expressions: _col0 (type: string)
    + sort order: +
    + Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
    + value expressions: _col1 (type: double)
            Reduce Operator Tree:
              Select Operator
    - expressions: VALUE._col0 (type: string), VALUE._col1 (type: double)
    + expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: double)
                outputColumnNames: _col0, _col1
    - Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
                Limit
                  Number of rows: 10
                  Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/groupby_grouping_window.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/groupby_grouping_window.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/groupby_grouping_window.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/groupby_grouping_window.q.out Sat Feb 21 02:36:54 2015
    @@ -93,6 +93,25 @@ STAGE PLANS:
                outputColumnNames: _col0, _col2, _col3
                Statistics: Num rows: 1 Data size: 6 Basic stats: COMPLETE Column stats: NONE
                PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col0: int, _col2: int, _col3: int
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col3
    + partition by: _col0
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col3
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
                  Statistics: Num rows: 1 Data size: 6 Basic stats: COMPLETE Column stats: NONE
                  Select Operator
                    expressions: _col0 (type: int), _col2 (type: int), _col3 (type: int), _wcol0 (type: int)

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/groupby_multi_single_reducer3.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/groupby_multi_single_reducer3.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/groupby_multi_single_reducer3.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/groupby_multi_single_reducer3.q.out Sat Feb 21 02:36:54 2015
    @@ -1,9 +1,15 @@
      PREHOOK: query: -- HIVE-3849 Aliased column in where clause for multi-groupby single reducer cannot be resolved
    +
    +-- SORT_QUERY_RESULTS
    +
      create table e1 (key string, count int)
      PREHOOK: type: CREATETABLE
      PREHOOK: Output: database:default
      PREHOOK: Output: default@e1
      POSTHOOK: query: -- HIVE-3849 Aliased column in where clause for multi-groupby single reducer cannot be resolved
    +
    +-- SORT_QUERY_RESULTS
    +
      create table e1 (key string, count int)
      POSTHOOK: type: CREATETABLE
      POSTHOOK: Output: database:default

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/groupby_resolution.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/groupby_resolution.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/groupby_resolution.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/groupby_resolution.q.out Sat Feb 21 02:36:54 2015
    @@ -677,6 +677,25 @@ STAGE PLANS:
                outputColumnNames: _col0, _col1
                Statistics: Num rows: 83 Data size: 881 Basic stats: COMPLETE Column stats: NONE
                PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col0: string, _col1: bigint
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: 0
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col1
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
                  Statistics: Num rows: 83 Data size: 881 Basic stats: COMPLETE Column stats: NONE
                  Select Operator
                    expressions: _col0 (type: string), _col1 (type: bigint), _wcol0 (type: int)

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/groupby_sort_1.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/groupby_sort_1.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/groupby_sort_1.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/groupby_sort_1.q.out Sat Feb 21 02:36:54 2015
    @@ -4335,7 +4335,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3,_col4
                          columns.types string:bigint:string:string:bigint
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out Sat Feb 21 02:36:54 2015
    @@ -2406,7 +2406,7 @@ ABSTRACT SYNTAX TREE:
      TOK_QUERY
         TOK_FROM
            TOK_SUBQUERY
    - TOK_UNION
    + TOK_UNIONALL
                  TOK_QUERY
                     TOK_FROM
                        TOK_TABREF
    @@ -2903,7 +2903,7 @@ ABSTRACT SYNTAX TREE:
      TOK_QUERY
         TOK_FROM
            TOK_SUBQUERY
    - TOK_UNION
    + TOK_UNIONALL
                  TOK_QUERY
                     TOK_FROM
                        TOK_TABREF
    @@ -4150,7 +4150,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2,_col3,_col4
                        columns.types string:bigint:string:string:bigint
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/groupby_sort_skew_1.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/groupby_sort_skew_1.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/groupby_sort_skew_1.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/groupby_sort_skew_1.q.out Sat Feb 21 02:36:54 2015
    @@ -4725,7 +4725,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3,_col4
                          columns.types string:bigint:string:string:bigint
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out Sat Feb 21 02:36:54 2015
    @@ -2731,7 +2731,7 @@ ABSTRACT SYNTAX TREE:
      TOK_QUERY
         TOK_FROM
            TOK_SUBQUERY
    - TOK_UNION
    + TOK_UNIONALL
                  TOK_QUERY
                     TOK_FROM
                        TOK_TABREF
    @@ -3228,7 +3228,7 @@ ABSTRACT SYNTAX TREE:
      TOK_QUERY
         TOK_FROM
            TOK_SUBQUERY
    - TOK_UNION
    + TOK_UNIONALL
                  TOK_QUERY
                     TOK_FROM
                        TOK_TABREF
    @@ -4605,7 +4605,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2,_col3,_col4
                        columns.types string:bigint:string:string:bigint
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/input23.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/input23.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/input23.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/input23.q.out Sat Feb 21 02:36:54 2015
    @@ -161,7 +161,7 @@ STAGE PLANS:
                            columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7
                            columns.types string:string:string:string:string:string:string:string
                            escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                            serialization.format 1
                            serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/input25.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/input25.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/input25.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/input25.q.out Sat Feb 21 02:36:54 2015
    @@ -22,16 +22,16 @@ POSTHOOK: Output: default@tst
      POSTHOOK: Output: default@tst@d=2009-02-02
      PREHOOK: query: explain
      select * from (
    - select * from tst x where x.d='2009-01-01' limit 10
    + select * from (select * from tst x where x.d='2009-01-01' limit 10)a
          union all
    - select * from tst x where x.d='2009-02-02' limit 10
    + select * from (select * from tst x where x.d='2009-02-02' limit 10)b
      ) subq
      PREHOOK: type: QUERY
      POSTHOOK: query: explain
      select * from (
    - select * from tst x where x.d='2009-01-01' limit 10
    + select * from (select * from tst x where x.d='2009-01-01' limit 10)a
          union all
    - select * from tst x where x.d='2009-02-02' limit 10
    + select * from (select * from tst x where x.d='2009-02-02' limit 10)b
      ) subq
      POSTHOOK: type: QUERY
      STAGE DEPENDENCIES:
    @@ -144,9 +144,9 @@ STAGE PLANS:
              ListSink

      PREHOOK: query: select * from (
    - select * from tst x where x.d='2009-01-01' limit 10
    + select * from (select * from tst x where x.d='2009-01-01' limit 10)a
          union all
    - select * from tst x where x.d='2009-02-02' limit 10
    + select * from (select * from tst x where x.d='2009-02-02' limit 10)b
      ) subq
      PREHOOK: type: QUERY
      PREHOOK: Input: default@tst
    @@ -154,9 +154,9 @@ PREHOOK: Input: default@tst@d=2009-01-01
      PREHOOK: Input: default@tst@d=2009-02-02
      #### A masked pattern was here ####
      POSTHOOK: query: select * from (
    - select * from tst x where x.d='2009-01-01' limit 10
    + select * from (select * from tst x where x.d='2009-01-01' limit 10)a
          union all
    - select * from tst x where x.d='2009-02-02' limit 10
    + select * from (select * from tst x where x.d='2009-02-02' limit 10)b
      ) subq
      POSTHOOK: type: QUERY
      POSTHOOK: Input: default@tst

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/input26.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/input26.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/input26.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/input26.q.out Sat Feb 21 02:36:54 2015
    @@ -1,15 +1,15 @@
      PREHOOK: query: explain
      select * from (
    - select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5
    + select * from (select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5)pa
          union all
    - select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5
    + select * from (select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5)pb
      )subq
      PREHOOK: type: QUERY
      POSTHOOK: query: explain
      select * from (
    - select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5
    + select * from (select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5)pa
          union all
    - select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5
    + select * from (select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5)pb
      )subq
      POSTHOOK: type: QUERY
      STAGE DEPENDENCIES:
    @@ -113,18 +113,18 @@ STAGE PLANS:
              ListSink

      PREHOOK: query: select * from (
    - select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5
    + select * from (select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5)pa
          union all
    - select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5
    + select * from (select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5)pb
      )subq
      PREHOOK: type: QUERY
      PREHOOK: Input: default@srcpart
      PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
      #### A masked pattern was here ####
      POSTHOOK: query: select * from (
    - select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5
    + select * from (select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5)pa
          union all
    - select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5
    + select * from (select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5)pb
      )subq
      POSTHOOK: type: QUERY
      POSTHOOK: Input: default@srcpart

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/input42.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/input42.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/input42.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/input42.q.out Sat Feb 21 02:36:54 2015
    @@ -1220,7 +1220,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2,_col3
                              columns.types string:string:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1792,7 +1792,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2,_col3
                              columns.types string:string:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/input_part7.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/input_part7.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/input_part7.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/input_part7.q.out Sat Feb 21 02:36:54 2015
    @@ -19,7 +19,7 @@ ABSTRACT SYNTAX TREE:
      TOK_QUERY
         TOK_FROM
            TOK_SUBQUERY
    - TOK_UNION
    + TOK_UNIONALL
                  TOK_QUERY
                     TOK_FROM
                        TOK_TABREF
    @@ -283,7 +283,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2,_col3
                        columns.types string:string:string:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/input_part9.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/input_part9.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/input_part9.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/input_part9.q.out Sat Feb 21 02:36:54 2015
    @@ -71,7 +71,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2,_col3
                              columns.types string:string:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/join34.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/join34.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/join34.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/join34.q.out Sat Feb 21 02:36:54 2015
    @@ -40,7 +40,7 @@ TOK_QUERY
         TOK_FROM
            TOK_JOIN
               TOK_SUBQUERY
    - TOK_UNION
    + TOK_UNIONALL
                     TOK_QUERY
                        TOK_FROM
                           TOK_TABREF

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/join35.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/join35.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/join35.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/join35.q.out Sat Feb 21 02:36:54 2015
    @@ -40,7 +40,7 @@ TOK_QUERY
         TOK_FROM
            TOK_JOIN
               TOK_SUBQUERY
    - TOK_UNION
    + TOK_UNIONALL
                     TOK_QUERY
                        TOK_FROM
                           TOK_TABREF

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/join_filters_overlap.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/join_filters_overlap.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/join_filters_overlap.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/join_filters_overlap.q.out Sat Feb 21 02:36:54 2015
    @@ -229,7 +229,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3,_col4,_col5
                          columns.types int:int:int:int:int:int
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -480,7 +480,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3,_col4,_col5
                          columns.types int:int:int:int:int:int
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -745,7 +745,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3,_col4,_col5
                          columns.types int:int:int:int:int:int
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1040,7 +1040,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7
                          columns.types int:int:int:int:int:int:int:int
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1329,7 +1329,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7
                          columns.types int:int:int:int:int:int:int:int
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/limit_pushdown.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/limit_pushdown.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/limit_pushdown.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/limit_pushdown.q.out Sat Feb 21 02:36:54 2015
    @@ -352,34 +352,34 @@ STAGE PLANS:
            Map Operator Tree:
                TableScan
                  alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                  Select Operator
                    expressions: cdouble (type: double)
                    outputColumnNames: _col0
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                    Group By Operator
                      keys: _col0 (type: double)
                      mode: hash
                      outputColumnNames: _col0
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                      Reduce Output Operator
                        key expressions: _col0 (type: double)
                        sort order: +
                        Map-reduce partition columns: _col0 (type: double)
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        TopN Hash Memory Usage: 0.3
            Reduce Operator Tree:
              Group By Operator
                keys: KEY._col0 (type: double)
                mode: mergepartial
                outputColumnNames: _col0
    - Statistics: Num rows: 6144 Data size: 188618 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 6144 Data size: 1320982 Basic stats: COMPLETE Column stats: NONE
                Limit
                  Number of rows: 20
    - Statistics: Num rows: 20 Data size: 600 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 20 Data size: 4300 Basic stats: COMPLETE Column stats: NONE
                  File Output Operator
                    compressed: false
    - Statistics: Num rows: 20 Data size: 600 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 20 Data size: 4300 Basic stats: COMPLETE Column stats: NONE
                    table:
                        input format: org.apache.hadoop.mapred.TextInputFormat
                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
    @@ -435,22 +435,22 @@ STAGE PLANS:
            Map Operator Tree:
                TableScan
                  alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                  Select Operator
                    expressions: ctinyint (type: tinyint), cdouble (type: double)
                    outputColumnNames: _col0, _col1
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                    Group By Operator
                      aggregations: count(DISTINCT _col1)
                      keys: _col0 (type: tinyint), _col1 (type: double)
                      mode: hash
                      outputColumnNames: _col0, _col1, _col2
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                      Reduce Output Operator
                        key expressions: _col0 (type: tinyint), _col1 (type: double)
                        sort order: ++
                        Map-reduce partition columns: _col0 (type: tinyint)
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        TopN Hash Memory Usage: 0.3
            Reduce Operator Tree:
              Group By Operator
    @@ -458,13 +458,13 @@ STAGE PLANS:
                keys: KEY._col0 (type: tinyint)
                mode: mergepartial
                outputColumnNames: _col0, _col1
    - Statistics: Num rows: 6144 Data size: 188618 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 6144 Data size: 1320982 Basic stats: COMPLETE Column stats: NONE
                Limit
                  Number of rows: 20
    - Statistics: Num rows: 20 Data size: 600 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 20 Data size: 4300 Basic stats: COMPLETE Column stats: NONE
                  File Output Operator
                    compressed: false
    - Statistics: Num rows: 20 Data size: 600 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 20 Data size: 4300 Basic stats: COMPLETE Column stats: NONE
                    table:
                        input format: org.apache.hadoop.mapred.TextInputFormat
                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
    @@ -522,22 +522,22 @@ STAGE PLANS:
            Map Operator Tree:
                TableScan
                  alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                  Select Operator
                    expressions: ctinyint (type: tinyint), cstring1 (type: string), cstring2 (type: string)
                    outputColumnNames: _col0, _col1, _col2
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                    Group By Operator
                      aggregations: count(DISTINCT _col1), count(DISTINCT _col2)
                      keys: _col0 (type: tinyint), _col1 (type: string), _col2 (type: string)
                      mode: hash
                      outputColumnNames: _col0, _col1, _col2, _col3, _col4
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                      Reduce Output Operator
                        key expressions: _col0 (type: tinyint), _col1 (type: string), _col2 (type: string)
                        sort order: +++
                        Map-reduce partition columns: _col0 (type: tinyint)
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        TopN Hash Memory Usage: 0.3
            Reduce Operator Tree:
              Group By Operator
    @@ -545,13 +545,13 @@ STAGE PLANS:
                keys: KEY._col0 (type: tinyint)
                mode: mergepartial
                outputColumnNames: _col0, _col1, _col2
    - Statistics: Num rows: 6144 Data size: 188618 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 6144 Data size: 1320982 Basic stats: COMPLETE Column stats: NONE
                Limit
                  Number of rows: 20
    - Statistics: Num rows: 20 Data size: 600 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 20 Data size: 4300 Basic stats: COMPLETE Column stats: NONE
                  File Output Operator
                    compressed: false
    - Statistics: Num rows: 20 Data size: 600 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 20 Data size: 4300 Basic stats: COMPLETE Column stats: NONE
                    table:
                        input format: org.apache.hadoop.mapred.TextInputFormat
                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_1.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_1.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_1.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_1.q.out Sat Feb 21 02:36:54 2015
    @@ -466,7 +466,7 @@ STAGE PLANS:
                              columns _col0,_col1
                              columns.types string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_11.q.java1.7.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_11.q.java1.7.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_11.q.java1.7.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_11.q.java1.7.out Sat Feb 21 02:36:54 2015
    @@ -338,7 +338,7 @@ STAGE PLANS:
                              columns _col0,_col1
                              columns.types string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_12.q.java1.7.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_12.q.java1.7.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_12.q.java1.7.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_12.q.java1.7.out Sat Feb 21 02:36:54 2015
    @@ -351,7 +351,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2,_col3,_col4,_col5,_col6
                              columns.types string:string:string:string:string:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -509,7 +509,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2,_col3,_col4,_col5,_col6
                              columns.types string:string:string:string:string:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_13.q.java1.7.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_13.q.java1.7.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_13.q.java1.7.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_13.q.java1.7.out Sat Feb 21 02:36:54 2015
    @@ -351,7 +351,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2,_col3,_col4,_col5,_col6
                              columns.types string:string:string:string:string:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_14.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_14.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_14.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_14.q.out Sat Feb 21 02:36:54 2015
    @@ -309,7 +309,7 @@ STAGE PLANS:
                              columns _col0,_col1
                              columns.types string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_2.q.java1.7.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_2.q.java1.7.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_2.q.java1.7.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_2.q.java1.7.out Sat Feb 21 02:36:54 2015
    @@ -437,7 +437,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2,_col3
                              columns.types string:string:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_3.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_3.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_3.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_3.q.out Sat Feb 21 02:36:54 2015
    @@ -406,7 +406,7 @@ STAGE PLANS:
                              columns _col0,_col1
                              columns.types string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_4.q.java1.7.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_4.q.java1.7.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_4.q.java1.7.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_4.q.java1.7.out Sat Feb 21 02:36:54 2015
    @@ -848,7 +848,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2,_col3
                              columns.types string:string:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_5.q.java1.7.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_5.q.java1.7.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_5.q.java1.7.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_5.q.java1.7.out Sat Feb 21 02:36:54 2015
    @@ -481,7 +481,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2,_col3
                              columns.types string:string:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_9.q.java1.7.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_9.q.java1.7.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_9.q.java1.7.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/list_bucket_dml_9.q.java1.7.out Sat Feb 21 02:36:54 2015
    @@ -848,7 +848,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2,_col3
                              columns.types string:string:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
  • Prasanthj at Feb 21, 2015 at 2:37 am
    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFPower.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFPower.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFPower.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFPower.java Sat Feb 21 02:36:54 2015
    @@ -41,7 +41,6 @@ extended = "Example:\n"
      @VectorizedExpressions({FuncPowerLongToDouble.class, FuncPowerDoubleToDouble.class})
      public class GenericUDFPower extends GenericUDF {
        private final String opName;
    - private final String opDisplayName;

        private transient PrimitiveObjectInspector baseOI;
        private transient PrimitiveObjectInspector powerOI;
    @@ -54,7 +53,6 @@ public class GenericUDFPower extends Gen

        public GenericUDFPower() {
          opName = getClass().getSimpleName();
    - opDisplayName = "power";
          resultOI = PrimitiveObjectInspectorFactory.writableDoubleObjectInspector;
        }

    @@ -103,7 +101,7 @@ public class GenericUDFPower extends Gen
        @Override
        public String getDisplayString(String[] children) {
          assert (children.length == 2);
    - return opDisplayName + "(" + children[0] + ", " + children[1] + ")";
    + return getStandardDisplayString("power", children);
        }

        @Override

    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFPrintf.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFPrintf.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFPrintf.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFPrintf.java Sat Feb 21 02:36:54 2015
    @@ -141,12 +141,6 @@ public class GenericUDFPrintf extends Ge
        @Override
        public String getDisplayString(String[] children) {
          assert (children.length >= 2);
    - StringBuilder sb = new StringBuilder();
    - sb.append("printf(");
    - for (int i = 0; i < children.length - 1; i++) {
    - sb.append(children[i]).append(", ");
    - }
    - sb.append(children[children.length - 1]).append(")");
    - return sb.toString();
    + return getStandardDisplayString("printf", children);
        }
      }

    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFRound.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFRound.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFRound.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFRound.java Sat Feb 21 02:36:54 2015
    @@ -22,6 +22,7 @@ import org.apache.hadoop.hive.common.typ
      import org.apache.hadoop.hive.ql.exec.Description;
      import org.apache.hadoop.hive.ql.exec.UDFArgumentException;
      import org.apache.hadoop.hive.ql.exec.UDFArgumentLengthException;
    +import org.apache.hadoop.hive.ql.exec.UDFArgumentTypeException;
      import org.apache.hadoop.hive.ql.exec.vector.VectorizedExpressions;
      import org.apache.hadoop.hive.ql.exec.vector.expressions.FuncRoundWithNumDigitsDecimalToDecimal;
      import org.apache.hadoop.hive.ql.exec.vector.expressions.RoundWithNumDigitsDoubleToDouble;
    @@ -83,38 +84,42 @@ public class GenericUDFRound extends Gen
                "ROUND requires one or two argument, got " + arguments.length);
          }

    - inputOI = (PrimitiveObjectInspector) arguments[0];
    - if (inputOI.getCategory() != Category.PRIMITIVE) {
    - throw new UDFArgumentException(
    - "ROUND input only takes primitive types, got " + inputOI.getTypeName());
    + if (arguments[0].getCategory() != Category.PRIMITIVE) {
    + throw new UDFArgumentTypeException(0,
    + "ROUND input only takes primitive types, got " + arguments[0].getTypeName());
          }
    + inputOI = (PrimitiveObjectInspector) arguments[0];

          if (arguments.length == 2) {
    + if (arguments[1].getCategory() != Category.PRIMITIVE) {
    + throw new UDFArgumentTypeException(1,
    + "ROUND second argument only takes primitive types, got " + arguments[1].getTypeName());
    + }
            PrimitiveObjectInspector scaleOI = (PrimitiveObjectInspector) arguments[1];
            switch (scaleOI.getPrimitiveCategory()) {
            case VOID:
              break;
            case BYTE:
              if (!(scaleOI instanceof WritableConstantByteObjectInspector)) {
    - throw new UDFArgumentException("ROUND second argument only takes constant");
    + throw new UDFArgumentTypeException(1, "ROUND second argument only takes constant");
              }
              scale = ((WritableConstantByteObjectInspector)scaleOI).getWritableConstantValue().get();
              break;
            case SHORT:
              if (!(scaleOI instanceof WritableConstantShortObjectInspector)) {
    - throw new UDFArgumentException("ROUND second argument only takes constant");
    + throw new UDFArgumentTypeException(1, "ROUND second argument only takes constant");
              }
              scale = ((WritableConstantShortObjectInspector)scaleOI).getWritableConstantValue().get();
              break;
            case INT:
              if (!(scaleOI instanceof WritableConstantIntObjectInspector)) {
    - throw new UDFArgumentException("ROUND second argument only takes constant");
    + throw new UDFArgumentTypeException(1, "ROUND second argument only takes constant");
              }
              scale = ((WritableConstantIntObjectInspector)scaleOI).getWritableConstantValue().get();
              break;
            case LONG:
              if (!(scaleOI instanceof WritableConstantLongObjectInspector)) {
    - throw new UDFArgumentException("ROUND second argument only takes constant");
    + throw new UDFArgumentTypeException(1, "ROUND second argument only takes constant");
              }
              long l = ((WritableConstantLongObjectInspector)scaleOI).getWritableConstantValue().get();
              if (l < Integer.MIN_VALUE || l > Integer.MAX_VALUE) {
    @@ -123,7 +128,7 @@ public class GenericUDFRound extends Gen
              scale = (int)l;
              break;
            default:
    - throw new UDFArgumentException("ROUND second argument only takes integer constant");
    + throw new UDFArgumentTypeException(1, "ROUND second argument only takes integer constant");
            }
          }

    @@ -151,8 +156,9 @@ public class GenericUDFRound extends Gen
            converterFromString = ObjectInspectorConverters.getConverter(inputOI, outputOI);
            break;
          default:
    - throw new UDFArgumentException("Only numeric data types are allowed for ROUND function. Got " +
    - inputType.name());
    + throw new UDFArgumentTypeException(0,
    + "Only numeric or string group data types are allowed for ROUND function. Got "
    + + inputType.name());
          }

          return outputOI;
    @@ -240,8 +246,9 @@ public class GenericUDFRound extends Gen
             }
             return round(doubleValue, scale);
           default:
    - throw new UDFArgumentException("Only numeric data types are allowed for ROUND function. Got " +
    - inputType.name());
    + throw new UDFArgumentTypeException(0,
    + "Only numeric or string group data types are allowed for ROUND function. Got "
    + + inputType.name());
          }
        }

    @@ -256,17 +263,7 @@ public class GenericUDFRound extends Gen

        @Override
        public String getDisplayString(String[] children) {
    - StringBuilder sb = new StringBuilder();
    - sb.append("round(");
    - if (children.length > 0) {
    - sb.append(children[0]);
    - for (int i = 1; i < children.length; i++) {
    - sb.append(", ");
    - sb.append(children[i]);
    - }
    - }
    - sb.append(")");
    - return sb.toString();
    + return getStandardDisplayString("round", children);
        }

      }

    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSentences.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSentences.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSentences.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSentences.java Sat Feb 21 02:36:54 2015
    @@ -128,14 +128,6 @@ public class GenericUDFSentences extends
        @Override
        public String getDisplayString(String[] children) {
          assert (children.length >= 1 && children.length <= 3);
    - String display = "sentences(" + children[0];
    - if(children.length > 1) {
    - display += ", " + children[1];
    - if(children.length > 2) {
    - display += ", " + children[2];
    - }
    - }
    - display += ")";
    - return display;
    + return getStandardDisplayString("sentences", children);
        }
      }

    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSize.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSize.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSize.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSize.java Sat Feb 21 02:36:54 2015
    @@ -79,6 +79,6 @@ public class GenericUDFSize extends Gene
        @Override
        public String getDisplayString(String[] children) {
          assert (children.length == 1);
    - return "size(" + children[0] + ")";
    + return getStandardDisplayString("size", children);
        }
      }

    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSortArray.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSortArray.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSortArray.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSortArray.java Sat Feb 21 02:36:54 2015
    @@ -116,6 +116,6 @@ public class GenericUDFSortArray extends
        @Override
        public String getDisplayString(String[] children) {
          assert (children.length == 1);
    - return "sort_array(" + children[0] + ")";
    + return getStandardDisplayString("sort_array", children);
       }
      }

    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSplit.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSplit.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSplit.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSplit.java Sat Feb 21 02:36:54 2015
    @@ -82,7 +82,7 @@ public class GenericUDFSplit extends Gen
        @Override
        public String getDisplayString(String[] children) {
          assert (children.length == 2);
    - return "split(" + children[0] + ", " + children[1] + ")";
    + return getStandardDisplayString("split", children);
        }

      }

    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFStringToMap.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFStringToMap.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFStringToMap.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFStringToMap.java Sat Feb 21 02:36:54 2015
    @@ -104,19 +104,7 @@ public class GenericUDFStringToMap exten

        @Override
        public String getDisplayString(String[] children) {
    - StringBuilder sb = new StringBuilder();
    - sb.append("str_to_map(");
          assert (children.length <= 3);
    - boolean firstChild = true;
    - for (String child : children) {
    - if (firstChild) {
    - firstChild = false;
    - } else {
    - sb.append(",");
    - }
    - sb.append(child);
    - }
    - sb.append(")");
    - return sb.toString();
    + return getStandardDisplayString("str_to_map", children, ",");
        }
      }

    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFStruct.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFStruct.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFStruct.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFStruct.java Sat Feb 21 02:36:54 2015
    @@ -59,15 +59,6 @@ public class GenericUDFStruct extends Ge

        @Override
        public String getDisplayString(String[] children) {
    - StringBuilder sb = new StringBuilder();
    - sb.append("struct(");
    - for (int i = 0; i < children.length; i++) {
    - if (i > 0) {
    - sb.append(',');
    - }
    - sb.append(children[i]);
    - }
    - sb.append(')');
    - return sb.toString();
    + return getStandardDisplayString("struct", children, ",");
        }
      }

    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFTranslate.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFTranslate.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFTranslate.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFTranslate.java Sat Feb 21 02:36:54 2015
    @@ -287,7 +287,7 @@ public class GenericUDFTranslate extends
        @Override
        public String getDisplayString(String[] children) {
          assert (children.length == 3);
    - return "translate(" + children[0] + ", " + children[1] + ", " + children[2] + ")";
    + return getStandardDisplayString("translate", children);
        }

      }

    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFUpper.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFUpper.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFUpper.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFUpper.java Sat Feb 21 02:36:54 2015
    @@ -107,17 +107,7 @@ public class GenericUDFUpper extends Gen

        @Override
        public String getDisplayString(String[] children) {
    - StringBuilder sb = new StringBuilder();
    - sb.append("upper(");
    - if (children.length > 0) {
    - sb.append(children[0]);
    - for (int i = 1; i < children.length; i++) {
    - sb.append(",");
    - sb.append(children[i]);
    - }
    - }
    - sb.append(")");
    - return sb.toString();
    + return getStandardDisplayString("upper", children);
        }

      }

    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFInline.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFInline.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFInline.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFInline.java Sat Feb 21 02:36:54 2015
    @@ -19,6 +19,7 @@
      package org.apache.hadoop.hive.ql.udf.generic;

      import java.util.ArrayList;
    +import java.util.List;

      import org.apache.hadoop.hive.ql.exec.Description;
      import org.apache.hadoop.hive.ql.exec.UDFArgumentException;
    @@ -57,8 +58,11 @@ public class GenericUDTFInline extends G

        @Override
        public void process(Object[] os) throws HiveException {
    - for (Object row : new ArrayList<Object>(li.getList(os[0]))) {
    - forward(row);
    + List<?> list = li.getList(os[0]);
    + if (list != null && !list.isEmpty()) {
    + for (Object row : list.toArray()) {
    + forward(row);
    + }
          }
        }


    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/NumDistinctValueEstimator.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/NumDistinctValueEstimator.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/NumDistinctValueEstimator.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/NumDistinctValueEstimator.java Sat Feb 21 02:36:54 2015
    @@ -35,18 +35,18 @@ public class NumDistinctValueEstimator {
         * independent. As a consequence, the hash values will not distribute uniformly from 0 to 2^p-1
         * thus introducing errors in the estimates.
         */
    - private static final int bitVectorSize = 31;
    - private int numBitVectors;
    + private static final int BIT_VECTOR_SIZE = 31;
    + private final int numBitVectors;

        // Refer to Flajolet-Martin'86 for the value of phi
    - private final double phi = 0.77351;
    + private static final double PHI = 0.77351;

    - private int[] a;
    - private int[] b;
    - private FastBitSet[] bitVector = new FastBitSet[numBitVectors];
    + private final int[] a;
    + private final int[] b;
    + private final FastBitSet[] bitVector;

    - private Random aValue;
    - private Random bValue;
    + private final Random aValue;
    + private final Random bValue;

        /* Create a new distinctValueEstimator
         */
    @@ -54,7 +54,7 @@ public class NumDistinctValueEstimator {
          this.numBitVectors = numBitVectors;
          bitVector = new FastBitSet[numBitVectors];
          for (int i=0; i< numBitVectors; i++) {
    - bitVector[i] = new FastBitSet(bitVectorSize);
    + bitVector[i] = new FastBitSet(BIT_VECTOR_SIZE);
          }

          a = new int[numBitVectors];
    @@ -98,23 +98,30 @@ public class NumDistinctValueEstimator {
            b[i] = randVal;

            if (a[i] < 0) {
    - a[i] = a[i] + (1 << bitVectorSize - 1);
    + a[i] = a[i] + (1 << BIT_VECTOR_SIZE - 1);
            }

            if (b[i] < 0) {
    - b[i] = b[i] + (1 << bitVectorSize - 1);
    + b[i] = b[i] + (1 << BIT_VECTOR_SIZE - 1);
            }
          }
        }

        public NumDistinctValueEstimator(String s, int numBitVectors) {
    - FastBitSet b[] = deserialize(s, numBitVectors);
    + this.numBitVectors = numBitVectors;
    + FastBitSet bitVectorDeser[] = deserialize(s, numBitVectors);
          bitVector = new FastBitSet[numBitVectors];
          for(int i=0; i <numBitVectors; i++) {
    - bitVector[i] = new FastBitSet(bitVectorSize);
    + bitVector[i] = new FastBitSet(BIT_VECTOR_SIZE);
             bitVector[i].clear();
    - bitVector[i].or(b[i]);
    + bitVector[i].or(bitVectorDeser[i]);
          }
    +
    + a = null;
    + b = null;
    +
    + aValue = null;
    + bValue = null;
        }

        /**
    @@ -135,7 +142,7 @@ public class NumDistinctValueEstimator {
        }

        public int getBitVectorSize() {
    - return bitVectorSize;
    + return BIT_VECTOR_SIZE;
        }

        public void printNumDistinctValueEstimator() {
    @@ -145,7 +152,7 @@ public class NumDistinctValueEstimator {
          LOG.debug("Number of Vectors:");
          LOG.debug(numBitVectors);
          LOG.debug("Vector Size: ");
    - LOG.debug(bitVectorSize);
    + LOG.debug(BIT_VECTOR_SIZE);

          for (int i=0; i < numBitVectors; i++) {
            t = t + bitVector[i].toString();
    @@ -173,7 +180,7 @@ public class NumDistinctValueEstimator {
        private FastBitSet[] deserialize(String s, int numBitVectors) {
          FastBitSet[] b = new FastBitSet[numBitVectors];
          for (int j=0; j < numBitVectors; j++) {
    - b[j] = new FastBitSet(bitVectorSize);
    + b[j] = new FastBitSet(BIT_VECTOR_SIZE);
            b[j].clear();
          }

    @@ -219,7 +226,7 @@ public class NumDistinctValueEstimator {
        }

        private int generateHash(long v, int hashNum) {
    - int mod = (1<<bitVectorSize) - 1;
    + int mod = (1<<BIT_VECTOR_SIZE) - 1;
          long tempHash = a[hashNum] * v + b[hashNum];
          tempHash %= mod;
          int hash = (int) tempHash;
    @@ -234,7 +241,7 @@ public class NumDistinctValueEstimator {
        }

        private int generateHashForPCSA(long v) {
    - int mod = 1 << (bitVectorSize - 1) - 1;
    + int mod = 1 << (BIT_VECTOR_SIZE - 1) - 1;
          long tempHash = a[0] * v + b[0];
          tempHash %= mod;
          int hash = (int) tempHash;
    @@ -259,8 +266,8 @@ public class NumDistinctValueEstimator {
            int index;

            // Find the index of the least significant bit that is 1
    - for (index=0; index<bitVectorSize; index++) {
    - if (hash % 2 == 1) {
    + for (index=0; index<BIT_VECTOR_SIZE; index++) {
    + if (hash % 2 != 0) {
                break;
              }
              hash = hash >> 1;
    @@ -277,8 +284,8 @@ public class NumDistinctValueEstimator {
          int index;

          // Find the index of the least significant bit that is 1
    - for (index=0; index<bitVectorSize; index++) {
    - if (rho % 2 == 1) {
    + for (index=0; index<BIT_VECTOR_SIZE; index++) {
    + if (rho % 2 != 0) {
              break;
            }
            rho = rho >> 1;
    @@ -321,13 +328,13 @@ public class NumDistinctValueEstimator {

          for (int i=0; i < numBitVectors; i++) {
            int index = 0;
    - while (bitVector[i].get(index) && index < bitVectorSize) {
    + while (bitVector[i].get(index) && index < BIT_VECTOR_SIZE) {
              index = index + 1;
            }
            S = S + index;
          }

    - numDistinctValues = ((numBitVectors/phi) * Math.pow(2.0, S/numBitVectors));
    + numDistinctValues = ((numBitVectors/PHI) * Math.pow(2.0, S/numBitVectors));
          return ((long)numDistinctValues);
        }

    @@ -345,7 +352,7 @@ public class NumDistinctValueEstimator {
          }

          avgLeastSigZero =
    - (double)(sumLeastSigZero/(numBitVectors * 1.0)) - (Math.log(phi)/Math.log(2.0));
    + (double)(sumLeastSigZero/(numBitVectors * 1.0)) - (Math.log(PHI)/Math.log(2.0));
          numDistinctValues = Math.pow(2.0, avgLeastSigZero);
          return ((long)(numDistinctValues));
        }

    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/MatchPath.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/MatchPath.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/MatchPath.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/MatchPath.java Sat Feb 21 02:36:54 2015
    @@ -28,6 +28,7 @@ import org.apache.hadoop.hive.ql.exec.Ex
      import org.apache.hadoop.hive.ql.exec.ExprNodeEvaluatorFactory;
      import org.apache.hadoop.hive.ql.exec.PTFPartition;
      import org.apache.hadoop.hive.ql.exec.PTFPartition.PTFPartitionIterator;
    +import org.apache.hadoop.hive.ql.exec.Utilities;
      import org.apache.hadoop.hive.ql.metadata.HiveException;
      import org.apache.hadoop.hive.ql.parse.ASTNode;
      import org.apache.hadoop.hive.ql.parse.PTFTranslator;
    @@ -195,6 +196,20 @@ public class MatchPath extends TableFunc

            setOutputOI(OI);
          }
    +
    + @Override
    + public List<String> getReferencedColumns() throws SemanticException {
    + MatchPath matchPath = (MatchPath) evaluator;
    + List<String> columns = new ArrayList<>();
    + for (ExprNodeDesc exprNode : matchPath.resultExprInfo.resultExprNodes) {
    + Utilities.mergeUniqElems(columns, exprNode.getCols());
    + }
    + for (ExprNodeDesc exprNode : matchPath.symInfo.symbolExprsDecs) {
    + Utilities.mergeUniqElems(columns, exprNode.getCols());
    + }
    + return columns;
    + }
    +
          /*
           * validate and setup patternStr
           */
    @@ -356,6 +371,7 @@ public class MatchPath extends TableFunc

        static class SymbolsInfo {
          int sz;
    + ArrayList<ExprNodeDesc> symbolExprsDecs;
          ArrayList<ExprNodeEvaluator> symbolExprsEvaluators;
          ArrayList<ObjectInspector> symbolExprsOIs;
          ArrayList<String> symbolExprsNames;
    @@ -366,6 +382,7 @@ public class MatchPath extends TableFunc
            symbolExprsEvaluators = new ArrayList<ExprNodeEvaluator>(sz);
            symbolExprsOIs = new ArrayList<ObjectInspector>(sz);
            symbolExprsNames = new ArrayList<String>(sz);
    + symbolExprsDecs = new ArrayList<>(sz);
          }

          void add(String name, PTFExpressionDef arg)
    @@ -373,6 +390,7 @@ public class MatchPath extends TableFunc
            symbolExprsNames.add(name);
            symbolExprsEvaluators.add(arg.getExprEvaluator());
            symbolExprsOIs.add(arg.getOI());
    + symbolExprsDecs.add(arg.getExprNode());
          }
        }

    @@ -749,8 +767,7 @@ public class MatchPath extends TableFunc
            /*
             * create SelectListOI
             */
    - selectListInputOI = (StructObjectInspector)
    - PTFTranslator.getStandardStructOI(selectListInputRowResolver);
    + selectListInputOI = PTFTranslator.getStandardStructOI(selectListInputRowResolver);
          }

          private void fixResultExprString()

    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/NoopWithMap.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/NoopWithMap.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/NoopWithMap.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/NoopWithMap.java Sat Feb 21 02:36:54 2015
    @@ -29,11 +29,6 @@ import org.apache.hadoop.hive.serde2.obj

      public class NoopWithMap extends Noop
      {
    - @Override
    - public PTFPartition execute(PTFPartition iPart) throws HiveException
    - {
    - return iPart;
    - }

        @Override
        protected PTFPartition _transformRawInput(PTFPartition iPart) throws HiveException

    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/TableFunctionEvaluator.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/TableFunctionEvaluator.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/TableFunctionEvaluator.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/TableFunctionEvaluator.java Sat Feb 21 02:36:54 2015
    @@ -30,7 +30,6 @@ import org.apache.hadoop.hive.ql.metadat
      import org.apache.hadoop.hive.ql.plan.PTFDesc;
      import org.apache.hadoop.hive.ql.plan.ptf.PartitionedTableFunctionDef;
      import org.apache.hadoop.hive.ql.udf.generic.GenericUDAFEvaluator;
    -import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
      import org.apache.hadoop.hive.serde2.objectinspector.StructObjectInspector;

      /*
    @@ -60,7 +59,7 @@ import org.apache.hadoop.hive.serde2.obj
       */

      /**
    - * Based on Hive {@link GenericUDAFEvaluator}. Break up the responsibility of the old AsbtractTableFunction
    + * Based on Hive {@link GenericUDAFEvaluator}. Break up the responsibility of the old AbstractTableFunction
       * class into a Resolver and Evaluator.
       * <p>
       * The Evaluator also holds onto the {@link TableFunctionDef}. This provides information
    @@ -79,7 +78,7 @@ import org.apache.hadoop.hive.serde2.obj
       */
      public abstract class TableFunctionEvaluator {
        /*
    - * how is this different from the OutpuShape set on the TableDef.
    + * how is this different from the OutputShape set on the TableDef.
         * This is the OI of the object coming out of the PTF.
         * It is put in an output Partition whose Serde is usually LazyBinarySerde.
         * So the next PTF (or Operator) in the chain gets a LazyBinaryStruct.

    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/TableFunctionResolver.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/TableFunctionResolver.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/TableFunctionResolver.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/TableFunctionResolver.java Sat Feb 21 02:36:54 2015
    @@ -60,7 +60,7 @@ public abstract class TableFunctionResol
        /*
         * - called during translation.
         * - invokes createEvaluator which must be implemented by a subclass
    - * - sets up the evaluator with references to the TableDef, PartitionClass, PartitonMemsize and
    + * - sets up the evaluator with references to the TableDef, PartitionClass, PartitionMemsize and
         * the transformsRawInput boolean.
         */
        public void initialize(HiveConf cfg, PTFDesc ptfDesc, PartitionedTableFunctionDef tDef)
    @@ -193,4 +193,14 @@ public abstract class TableFunctionResol
         * a subclass must provide the {@link TableFunctionEvaluator} instance.
         */
        protected abstract TableFunctionEvaluator createEvaluator(PTFDesc ptfDesc, PartitionedTableFunctionDef tDef);
    +
    + /**
    + * Provide referenced columns names to be used in partition function
    + *
    + * @return null for unknown (will get all columns from table including virtual columns)
    + * @throws SemanticException
    + */
    + public List<String> getReferencedColumns() throws SemanticException {
    + return null;
    + }
      }

    Modified: hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/WindowingTableFunction.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/WindowingTableFunction.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/WindowingTableFunction.java (original)
    +++ hive/branches/llap/ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/WindowingTableFunction.java Sat Feb 21 02:36:54 2015
    @@ -69,8 +69,7 @@ public class WindowingTableFunction exte
        public void execute(PTFPartitionIterator<Object> pItr, PTFPartition outP) throws HiveException {
          ArrayList<List<?>> oColumns = new ArrayList<List<?>>();
          PTFPartition iPart = pItr.getPartition();
    - StructObjectInspector inputOI;
    - inputOI = (StructObjectInspector) iPart.getOutputOI();
    + StructObjectInspector inputOI = iPart.getOutputOI();

          WindowTableFunctionDef wTFnDef = (WindowTableFunctionDef) getTableDef();
          Order order = wTFnDef.getOrder().getExpressions().get(0).getOrder();
    @@ -145,7 +144,8 @@ public class WindowingTableFunction exte
          return true;
        }

    - private boolean streamingPossible(Configuration cfg, WindowFunctionDef wFnDef) {
    + private boolean streamingPossible(Configuration cfg, WindowFunctionDef wFnDef)
    + throws HiveException {
          WindowFrameDef wdwFrame = wFnDef.getWindowFrame();
          WindowFunctionInfo wFnInfo = FunctionRegistry.getWindowFunctionInfo(wFnDef
              .getName());
    @@ -202,7 +202,7 @@ public class WindowingTableFunction exte
         * So no Unbounded Preceding or Following.
         */
        @SuppressWarnings("resource")
    - private int[] setCanAcceptInputAsStream(Configuration cfg) {
    + private int[] setCanAcceptInputAsStream(Configuration cfg) throws HiveException {

          canAcceptInputAsStream = false;

    @@ -514,7 +514,6 @@ public class WindowingTableFunction exte
            i++;
          }

    - i=0;
          for(i=0; i < iPart.getOutputOI().getAllStructFieldRefs().size(); i++) {
            output.add(null);
          }

    Modified: hive/branches/llap/ql/src/protobuf/org/apache/hadoop/hive/ql/io/orc/orc_proto.proto
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/protobuf/org/apache/hadoop/hive/ql/io/orc/orc_proto.proto?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/protobuf/org/apache/hadoop/hive/ql/io/orc/orc_proto.proto (original)
    +++ hive/branches/llap/ql/src/protobuf/org/apache/hadoop/hive/ql/io/orc/orc_proto.proto Sat Feb 21 02:36:54 2015
    @@ -16,7 +16,9 @@
       * limitations under the License.
       */

    -package org.apache.hadoop.hive.ql.io.orc;
    +package orc.proto;
    +
    +option java_package = "org.apache.hadoop.hive.ql.io.orc";

      message IntegerStatistics {
        optional sint64 minimum = 1;
    @@ -86,6 +88,15 @@ message RowIndex {
        repeated RowIndexEntry entry = 1;
      }

    +message BloomFilter {
    + optional uint32 numHashFunctions = 1;
    + repeated fixed64 bitset = 2;
    +}
    +
    +message BloomFilterIndex {
    + repeated BloomFilter bloomFilter = 1;
    +}
    +
      message Stream {
        // if you add new index stream kinds, you need to make sure to update
        // StreamName to ensure it is added to the stripe in the right area
    @@ -97,8 +108,9 @@ message Stream {
          DICTIONARY_COUNT = 4;
          SECONDARY = 5;
          ROW_INDEX = 6;
    + BLOOM_FILTER = 7;
        }
    - required Kind kind = 1;
    + optional Kind kind = 1;
        optional uint32 column = 2;
        optional uint64 length = 3;
      }
    @@ -110,7 +122,7 @@ message ColumnEncoding {
          DIRECT_V2 = 2;
          DICTIONARY_V2 = 3;
        }
    - required Kind kind = 1;
    + optional Kind kind = 1;
        optional uint32 dictionarySize = 2;
      }

    @@ -140,7 +152,7 @@ message Type {
          VARCHAR = 16;
          CHAR = 17;
        }
    - required Kind kind = 1;
    + optional Kind kind = 1;
        repeated uint32 subtypes = 2 [packed=true];
        repeated string fieldNames = 3;
        optional uint32 maximumLength = 4;
    @@ -157,8 +169,8 @@ message StripeInformation {
      }

      message UserMetadataItem {
    - required string name = 1;
    - required bytes value = 2;
    + optional string name = 1;
    + optional bytes value = 2;
      }

      message StripeStatistics {

    Modified: hive/branches/llap/ql/src/test/org/apache/hadoop/hive/metastore/TestMetastoreExpr.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/org/apache/hadoop/hive/metastore/TestMetastoreExpr.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/org/apache/hadoop/hive/metastore/TestMetastoreExpr.java (original)
    +++ hive/branches/llap/ql/src/test/org/apache/hadoop/hive/metastore/TestMetastoreExpr.java Sat Feb 21 02:36:54 2015
    @@ -26,12 +26,9 @@ import java.util.Stack;
      import junit.framework.TestCase;

      import org.apache.hadoop.hive.conf.HiveConf;
    -import org.apache.hadoop.hive.metastore.api.AlreadyExistsException;
      import org.apache.hadoop.hive.metastore.api.Database;
      import org.apache.hadoop.hive.metastore.api.FieldSchema;
    -import org.apache.hadoop.hive.metastore.api.InvalidObjectException;
      import org.apache.hadoop.hive.metastore.api.InvalidOperationException;
    -import org.apache.hadoop.hive.metastore.api.MetaException;
      import org.apache.hadoop.hive.metastore.api.NoSuchObjectException;
      import org.apache.hadoop.hive.metastore.api.Order;
      import org.apache.hadoop.hive.metastore.api.Partition;
    @@ -90,14 +87,14 @@ public class TestMetastoreExpr extends T
          }
        }

    - private static void silentDropDatabase(String dbName) throws MetaException, TException {
    + private static void silentDropDatabase(String dbName) throws TException {
          try {
            for (String tableName : client.getTables(dbName, "*")) {
              client.dropTable(dbName, tableName);
            }
            client.dropDatabase(dbName);
    - } catch (NoSuchObjectException e) {
    - } catch (InvalidOperationException e) {
    + } catch (NoSuchObjectException ignore) {
    + } catch (InvalidOperationException ignore) {
          }
        }

    @@ -153,16 +150,16 @@ public class TestMetastoreExpr extends T
            client.listPartitionsByExpr(dbName, tblName,
                new byte[] { 'f', 'o', 'o' }, null, (short)-1, new ArrayList<Partition>());
            fail("Should have thrown IncompatibleMetastoreException");
    - } catch (IMetaStoreClient.IncompatibleMetastoreException ex) {
    + } catch (IMetaStoreClient.IncompatibleMetastoreException ignore) {
          }

          // Invalid expression => throw some exception, but not incompatible metastore.
          try {
            checkExpr(-1, dbName, tblName, e.val(31).intCol("p3").pred(">", 2).build());
            fail("Should have thrown");
    - } catch (IMetaStoreClient.IncompatibleMetastoreException ex) {
    + } catch (IMetaStoreClient.IncompatibleMetastoreException ignore) {
            fail("Should not have thrown IncompatibleMetastoreException");
    - } catch (Exception ex) {
    + } catch (Exception ignore) {
          }
        }

    @@ -198,7 +195,7 @@ public class TestMetastoreExpr extends T
            for (int i = 0; i < args; ++i) {
              children.add(stack.pop());
            }
    - stack.push(new ExprNodeGenericFuncDesc(TypeInfoFactory.booleanTypeInfo,
    + stack.push(new ExprNodeGenericFuncDesc(ti,
                FunctionRegistry.getFunctionInfo(name).getGenericUDF(), children));
            return this;
          }
    @@ -249,8 +246,7 @@ public class TestMetastoreExpr extends T
        }

        private void addPartition(HiveMetaStoreClient client, Table table,
    - List<String> vals, String location) throws InvalidObjectException,
    - AlreadyExistsException, MetaException, TException {
    + List<String> vals, String location) throws TException {

          Partition part = new Partition();
          part.setDbName(table.getDbName());

    Modified: hive/branches/llap/ql/src/test/org/apache/hadoop/hive/ql/exec/spark/TestHiveKVResultCache.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/org/apache/hadoop/hive/ql/exec/spark/TestHiveKVResultCache.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/org/apache/hadoop/hive/ql/exec/spark/TestHiveKVResultCache.java (original)
    +++ hive/branches/llap/ql/src/test/org/apache/hadoop/hive/ql/exec/spark/TestHiveKVResultCache.java Sat Feb 21 02:36:54 2015
    @@ -27,8 +27,6 @@ import java.util.Iterator;
      import java.util.List;
      import java.util.concurrent.LinkedBlockingQueue;

    -import org.apache.hadoop.conf.Configuration;
    -import org.apache.hadoop.hive.conf.HiveConf;
      import org.apache.hadoop.hive.ql.io.HiveKey;
      import org.apache.hadoop.io.BytesWritable;
      import org.junit.Test;
    @@ -42,8 +40,7 @@ public class TestHiveKVResultCache {
        @Test
        public void testSimple() throws Exception {
          // Create KV result cache object, add one (k,v) pair and retrieve them.
    - HiveConf conf = new HiveConf();
    - HiveKVResultCache cache = new HiveKVResultCache(conf);
    + HiveKVResultCache cache = new HiveKVResultCache();

          HiveKey key = new HiveKey("key".getBytes(), "key".hashCode());
          BytesWritable value = new BytesWritable("value".getBytes());
    @@ -60,10 +57,9 @@ public class TestHiveKVResultCache {

        @Test
        public void testSpilling() throws Exception {
    - HiveConf conf = new HiveConf();
    - HiveKVResultCache cache = new HiveKVResultCache(conf);
    + HiveKVResultCache cache = new HiveKVResultCache();

    - final int recordCount = HiveKVResultCache.IN_MEMORY_CACHE_SIZE * 3;
    + final int recordCount = HiveKVResultCache.IN_MEMORY_NUM_ROWS * 3;

          // Test using the same cache where first n rows are inserted then cache is cleared.
          // Next reuse the same cache and insert another m rows and verify the cache stores correctly.
    @@ -104,10 +100,18 @@ public class TestHiveKVResultCache {
        @Test
        public void testResultList() throws Exception {
          scanAndVerify(10000, 0, 0, "a", "b");
    + scanAndVerify(10000, 511, 0, "a", "b");
    + scanAndVerify(10000, 511 * 2, 0, "a", "b");
    + scanAndVerify(10000, 511, 10, "a", "b");
    + scanAndVerify(10000, 511 * 2, 10, "a", "b");
          scanAndVerify(10000, 512, 0, "a", "b");
          scanAndVerify(10000, 512 * 2, 0, "a", "b");
    - scanAndVerify(10000, 512, 10, "a", "b");
    - scanAndVerify(10000, 512 * 2, 10, "a", "b");
    + scanAndVerify(10000, 512, 3, "a", "b");
    + scanAndVerify(10000, 512 * 6, 10, "a", "b");
    + scanAndVerify(10000, 512 * 7, 5, "a", "b");
    + scanAndVerify(10000, 512 * 9, 19, "a", "b");
    + scanAndVerify(10000, 1, 0, "a", "b");
    + scanAndVerify(10000, 1, 1, "a", "b");
        }

        private static void scanAndVerify(
    @@ -176,8 +180,8 @@ public class TestHiveKVResultCache {
          // A queue to notify separateRowGenerator to generate the next batch of rows.
          private LinkedBlockingQueue<Boolean> queue;

    - MyHiveFunctionResultList(Configuration conf, Iterator inputIterator) {
    - super(conf, inputIterator);
    + MyHiveFunctionResultList(Iterator inputIterator) {
    + super(inputIterator);
          }

          void init(long rows, int threshold, int separate, String p1, String p2) {
    @@ -258,8 +262,7 @@ public class TestHiveKVResultCache {
        private static long scanResultList(long rows, int threshold, int separate,
            List<Tuple2<HiveKey, BytesWritable>> output, String prefix1, String prefix2) {
          final long iteratorCount = threshold == 0 ? 1 : rows * (100 - separate) / 100 / threshold;
    - MyHiveFunctionResultList resultList = new MyHiveFunctionResultList(
    - new HiveConf(), new Iterator() {
    + MyHiveFunctionResultList resultList = new MyHiveFunctionResultList(new Iterator() {
            // Input record iterator, not used
            private int i = 0;
            @Override

    Modified: hive/branches/llap/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestFileDump.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestFileDump.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestFileDump.java (original)
    +++ hive/branches/llap/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestFileDump.java Sat Feb 21 02:36:54 2015
    @@ -21,19 +21,6 @@ package org.apache.hadoop.hive.ql.io.orc
      import static org.junit.Assert.assertEquals;
      import static org.junit.Assert.assertNull;

    -import org.apache.hadoop.conf.Configuration;
    -import org.apache.hadoop.fs.FileSystem;
    -import org.apache.hadoop.fs.Path;
    -import org.apache.hadoop.hive.common.type.HiveChar;
    -import org.apache.hadoop.hive.common.type.HiveDecimal;
    -import org.apache.hadoop.hive.common.type.HiveVarchar;
    -import org.apache.hadoop.hive.conf.HiveConf;
    -import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
    -import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorFactory;
    -import org.apache.hive.common.util.HiveTestUtils;
    -import org.junit.Before;
    -import org.junit.Test;
    -
      import java.io.BufferedReader;
      import java.io.ByteArrayOutputStream;
      import java.io.File;
    @@ -49,6 +36,19 @@ import java.util.List;
      import java.util.Map;
      import java.util.Random;

    +import org.apache.hadoop.conf.Configuration;
    +import org.apache.hadoop.fs.FileSystem;
    +import org.apache.hadoop.fs.Path;
    +import org.apache.hadoop.hive.common.type.HiveChar;
    +import org.apache.hadoop.hive.common.type.HiveDecimal;
    +import org.apache.hadoop.hive.common.type.HiveVarchar;
    +import org.apache.hadoop.hive.conf.HiveConf;
    +import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
    +import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorFactory;
    +import org.apache.hive.common.util.HiveTestUtils;
    +import org.junit.Before;
    +import org.junit.Test;
    +
      public class TestFileDump {

        Path workDir = new Path(System.getProperty("test.tmp.dir"));
    @@ -303,4 +303,101 @@ public class TestFileDump {

          checkOutput(outputFilename, workDir + File.separator + outputFilename);
        }
    +
    + @Test
    + public void testBloomFilter() throws Exception {
    + ObjectInspector inspector;
    + synchronized (TestOrcFile.class) {
    + inspector = ObjectInspectorFactory.getReflectionObjectInspector
    + (MyRecord.class, ObjectInspectorFactory.ObjectInspectorOptions.JAVA);
    + }
    + conf.set(HiveConf.ConfVars.HIVE_ORC_ENCODING_STRATEGY.varname, "COMPRESSION");
    + OrcFile.WriterOptions options = OrcFile.writerOptions(conf)
    + .fileSystem(fs)
    + .inspector(inspector)
    + .stripeSize(100000)
    + .compress(CompressionKind.ZLIB)
    + .bufferSize(10000)
    + .rowIndexStride(1000)
    + .bloomFilterColumns("s");
    + Writer writer = OrcFile.createWriter(testFilePath, options);
    + Random r1 = new Random(1);
    + String[] words = new String[]{"It", "was", "the", "best", "of", "times,",
    + "it", "was", "the", "worst", "of", "times,", "it", "was", "the", "age",
    + "of", "wisdom,", "it", "was", "the", "age", "of", "foolishness,", "it",
    + "was", "the", "epoch", "of", "belief,", "it", "was", "the", "epoch",
    + "of", "incredulity,", "it", "was", "the", "season", "of", "Light,",
    + "it", "was", "the", "season", "of", "Darkness,", "it", "was", "the",
    + "spring", "of", "hope,", "it", "was", "the", "winter", "of", "despair,",
    + "we", "had", "everything", "before", "us,", "we", "had", "nothing",
    + "before", "us,", "we", "were", "all", "going", "direct", "to",
    + "Heaven,", "we", "were", "all", "going", "direct", "the", "other",
    + "way"};
    + for(int i=0; i < 21000; ++i) {
    + writer.addRow(new MyRecord(r1.nextInt(), r1.nextLong(),
    + words[r1.nextInt(words.length)]));
    + }
    + writer.close();
    + PrintStream origOut = System.out;
    + String outputFilename = "orc-file-dump-bloomfilter.out";
    + FileOutputStream myOut = new FileOutputStream(workDir + File.separator + outputFilename);
    +
    + // replace stdout and run command
    + System.setOut(new PrintStream(myOut));
    + FileDump.main(new String[]{testFilePath.toString(), "--rowindex=3"});
    + System.out.flush();
    + System.setOut(origOut);
    +
    +
    + checkOutput(outputFilename, workDir + File.separator + outputFilename);
    + }
    +
    + @Test
    + public void testBloomFilter2() throws Exception {
    + ObjectInspector inspector;
    + synchronized (TestOrcFile.class) {
    + inspector = ObjectInspectorFactory.getReflectionObjectInspector
    + (MyRecord.class, ObjectInspectorFactory.ObjectInspectorOptions.JAVA);
    + }
    + conf.set(HiveConf.ConfVars.HIVE_ORC_ENCODING_STRATEGY.varname, "COMPRESSION");
    + OrcFile.WriterOptions options = OrcFile.writerOptions(conf)
    + .fileSystem(fs)
    + .inspector(inspector)
    + .stripeSize(100000)
    + .compress(CompressionKind.ZLIB)
    + .bufferSize(10000)
    + .rowIndexStride(1000)
    + .bloomFilterColumns("l")
    + .bloomFilterFpp(0.01);
    + Writer writer = OrcFile.createWriter(testFilePath, options);
    + Random r1 = new Random(1);
    + String[] words = new String[]{"It", "was", "the", "best", "of", "times,",
    + "it", "was", "the", "worst", "of", "times,", "it", "was", "the", "age",
    + "of", "wisdom,", "it", "was", "the", "age", "of", "foolishness,", "it",
    + "was", "the", "epoch", "of", "belief,", "it", "was", "the", "epoch",
    + "of", "incredulity,", "it", "was", "the", "season", "of", "Light,",
    + "it", "was", "the", "season", "of", "Darkness,", "it", "was", "the",
    + "spring", "of", "hope,", "it", "was", "the", "winter", "of", "despair,",
    + "we", "had", "everything", "before", "us,", "we", "had", "nothing",
    + "before", "us,", "we", "were", "all", "going", "direct", "to",
    + "Heaven,", "we", "were", "all", "going", "direct", "the", "other",
    + "way"};
    + for(int i=0; i < 21000; ++i) {
    + writer.addRow(new MyRecord(r1.nextInt(), r1.nextLong(),
    + words[r1.nextInt(words.length)]));
    + }
    + writer.close();
    + PrintStream origOut = System.out;
    + String outputFilename = "orc-file-dump-bloomfilter2.out";
    + FileOutputStream myOut = new FileOutputStream(workDir + File.separator + outputFilename);
    +
    + // replace stdout and run command
    + System.setOut(new PrintStream(myOut));
    + FileDump.main(new String[]{testFilePath.toString(), "--rowindex=2"});
    + System.out.flush();
    + System.setOut(origOut);
    +
    +
    + checkOutput(outputFilename, workDir + File.separator + outputFilename);
    + }
      }

    Modified: hive/branches/llap/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java (original)
    +++ hive/branches/llap/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java Sat Feb 21 02:36:54 2015
    @@ -515,6 +515,10 @@ public class TestOrcFile {
            Object row = rows.next(null);
            assertEquals(tslist.get(idx++).getNanos(), ((TimestampWritable) row).getNanos());
          }
    + assertEquals(1, OrcUtils.getFlattenedColumnsCount(inspector));
    + boolean[] expected = new boolean[] {false};
    + boolean[] included = OrcUtils.includeColumns("", "ts", inspector);
    + assertEquals(true, Arrays.equals(expected, included));
        }

        @Test
    @@ -538,6 +542,19 @@ public class TestOrcFile {
          Reader reader = OrcFile.createReader(testFilePath,
              OrcFile.readerOptions(conf).filesystem(fs));

    + assertEquals(3, OrcUtils.getFlattenedColumnsCount(inspector));
    + boolean[] expected = new boolean[] {false, false, true};
    + boolean[] included = OrcUtils.includeColumns("string1", "bytes1,string1", inspector);
    + assertEquals(true, Arrays.equals(expected, included));
    +
    + expected = new boolean[] {false, false, false};
    + included = OrcUtils.includeColumns("", "bytes1,string1", inspector);
    + assertEquals(true, Arrays.equals(expected, included));
    +
    + expected = new boolean[] {false, false, false};
    + included = OrcUtils.includeColumns(null, "bytes1,string1", inspector);
    + assertEquals(true, Arrays.equals(expected, included));
    +
          // check the stats
          ColumnStatistics[] stats = reader.getStatistics();
          assertEquals(4, stats[0].getNumberOfValues());
    @@ -634,6 +651,12 @@ public class TestOrcFile {
          writer.close();
          Reader reader = OrcFile.createReader(testFilePath,
              OrcFile.readerOptions(conf).filesystem(fs));
    +
    + assertEquals(3, OrcUtils.getFlattenedColumnsCount(inspector));
    + boolean[] expected = new boolean[] {false, true, false};
    + boolean[] included = OrcUtils.includeColumns("int1", "int1,string1", inspector);
    + assertEquals(true, Arrays.equals(expected, included));
    +
          Metadata metadata = reader.getMetadata();
          int numStripes = metadata.getStripeStatistics().size();
          assertEquals(3, numStripes);
    @@ -672,7 +695,7 @@ public class TestOrcFile {
          assertEquals(5000, ((StringColumnStatistics)ss3.getColumnStatistics()[2]).getSum());

          RecordReaderImpl recordReader = (RecordReaderImpl) reader.rows();
    - OrcProto.RowIndex[] index = recordReader.readRowIndex(0, null);
    + OrcProto.RowIndex[] index = recordReader.readRowIndex(0, null).getRowGroupIndex();
          assertEquals(3, index.length);
          List<OrcProto.RowIndexEntry> items = index[1].getEntryList();
          assertEquals(1, items.size());
    @@ -682,7 +705,7 @@ public class TestOrcFile {
          assertEquals(0, items.get(0).getPositions(2));
          assertEquals(1,
                       items.get(0).getStatistics().getIntStatistics().getMinimum());
    - index = recordReader.readRowIndex(1, null);
    + index = recordReader.readRowIndex(1, null).getRowGroupIndex();
          assertEquals(3, index.length);
          items = index[1].getEntryList();
          assertEquals(2,
    @@ -715,6 +738,44 @@ public class TestOrcFile {
          Reader reader = OrcFile.createReader(testFilePath,
              OrcFile.readerOptions(conf).filesystem(fs));

    + assertEquals(24, OrcUtils.getFlattenedColumnsCount(inspector));
    + boolean[] expected = new boolean[] {false, false, false, false, false,
    + false, false, false, false, false,
    + false, false, false, false, false,
    + false, false, false, false, false,
    + false, false, false, false};
    + boolean[] included = OrcUtils.includeColumns("",
    + "boolean1,byte1,short1,int1,long1,float1,double1,bytes1,string1,middle,list,map", inspector);
    + assertEquals(true, Arrays.equals(expected, included));
    +
    + expected = new boolean[] {false, true, false, false, false,
    + false, false, false, false, true,
    + true, true, true, true, true,
    + false, false, false, false, true,
    + true, true, true, true};
    + included = OrcUtils.includeColumns("boolean1,string1,middle,map",
    + "boolean1,byte1,short1,int1,long1,float1,double1,bytes1,string1,middle,list,map", inspector);
    + assertEquals(true, Arrays.equals(expected, included));
    +
    + expected = new boolean[] {false, true, false, false, false,
    + false, false, false, false, true,
    + true, true, true, true, true,
    + false, false, false, false, true,
    + true, true, true, true};
    + included = OrcUtils.includeColumns("boolean1,string1,middle,map",
    + "boolean1,byte1,short1,int1,long1,float1,double1,bytes1,string1,middle,list,map", inspector);
    + assertEquals(true, Arrays.equals(expected, included));
    +
    + expected = new boolean[] {false, true, true, true, true,
    + true, true, true, true, true,
    + true, true, true, true, true,
    + true, true, true, true, true,
    + true, true, true, true};
    + included = OrcUtils.includeColumns(
    + "boolean1,byte1,short1,int1,long1,float1,double1,bytes1,string1,middle,list,map",
    + "boolean1,byte1,short1,int1,long1,float1,double1,bytes1,string1,middle,list,map", inspector);
    + assertEquals(true, Arrays.equals(expected, included));
    +
          Metadata metadata = reader.getMetadata();

          // check the stats
    @@ -1183,6 +1244,20 @@ public class TestOrcFile {
          writer.close();
          Reader reader = OrcFile.createReader(testFilePath,
              OrcFile.readerOptions(conf).filesystem(fs));
    +
    + assertEquals(6, OrcUtils.getFlattenedColumnsCount(inspector));
    + boolean[] expected = new boolean[] {false, false, false, false, false, false};
    + boolean[] included = OrcUtils.includeColumns("", "time,union,decimal", inspector);
    + assertEquals(true, Arrays.equals(expected, included));
    +
    + expected = new boolean[] {false, true, false, false, false, true};
    + included = OrcUtils.includeColumns("time,decimal", "time,union,decimal", inspector);
    + assertEquals(true, Arrays.equals(expected, included));
    +
    + expected = new boolean[] {false, false, true, true, true, false};
    + included = OrcUtils.includeColumns("union", "time,union,decimal", inspector);
    + assertEquals(true, Arrays.equals(expected, included));
    +
          assertEquals(false, reader.getMetadataKeys().iterator().hasNext());
          assertEquals(5309, reader.getNumberOfRows());
          DecimalColumnStatistics stats =
  • Prasanthj at Feb 21, 2015 at 2:37 am
    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/load_dyn_part14.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/load_dyn_part14.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/load_dyn_part14.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/load_dyn_part14.q Sat Feb 21 02:36:54 2015
    @@ -14,20 +14,20 @@ set hive.exec.dynamic.partition.mode=non
      explain
      insert overwrite table nzhang_part14 partition(value)
      select key, value from (
    - select 'k1' as key, cast(null as string) as value from src limit 2
    + select * from (select 'k1' as key, cast(null as string) as value from src limit 2)a
        union all
    - select 'k2' as key, '' as value from src limit 2
    + select * from (select 'k2' as key, '' as value from src limit 2)b
        union all
    - select 'k3' as key, ' ' as value from src limit 2
    + select * from (select 'k3' as key, ' ' as value from src limit 2)c
      ) T;

      insert overwrite table nzhang_part14 partition(value)
      select key, value from (
    - select 'k1' as key, cast(null as string) as value from src limit 2
    + select * from (select 'k1' as key, cast(null as string) as value from src limit 2)a
        union all
    - select 'k2' as key, '' as value from src limit 2
    + select * from (select 'k2' as key, '' as value from src limit 2)b
        union all
    - select 'k3' as key, ' ' as value from src limit 2
    + select * from (select 'k3' as key, ' ' as value from src limit 2)c
      ) T;



    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/metadataOnlyOptimizer.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/metadataOnlyOptimizer.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/metadataOnlyOptimizer.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/metadataOnlyOptimizer.q Sat Feb 21 02:36:54 2015
    @@ -33,9 +33,9 @@ UNION all
      SELECT key as key from src
      ) tab group by key;

    -select k from (SELECT '1' as k from src limit 0 union all select key as k from src limit 1)tab;
    +select k from (select * from (SELECT '1' as k from src limit 0)a union all select key as k from src limit 1)tab;

    -select k from (SELECT '1' as k from src limit 1 union all select key as k from src limit 0)tab;
    +select k from (select * from (SELECT '1' as k from src limit 1)a union all select key as k from src limit 0)tab;

      select max(ds) from srcpart;


    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/orc_analyze.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/orc_analyze.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/orc_analyze.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/orc_analyze.q Sat Feb 21 02:36:54 2015
    @@ -30,8 +30,13 @@ STORED AS orc;
      INSERT OVERWRITE TABLE orc_create_people SELECT * FROM orc_create_people_staging ORDER BY id;

      set hive.stats.autogather = true;
    +analyze table orc_create_people compute statistics;
    +desc formatted orc_create_people;
    +
      analyze table orc_create_people compute statistics partialscan;
    +desc formatted orc_create_people;

    +analyze table orc_create_people compute statistics noscan;
      desc formatted orc_create_people;

      drop table orc_create_people;
    @@ -70,8 +75,15 @@ INSERT OVERWRITE TABLE orc_create_people
        SELECT * FROM orc_create_people_staging ORDER BY id;

      set hive.stats.autogather = true;
    +analyze table orc_create_people partition(state) compute statistics;
    +desc formatted orc_create_people partition(state="Ca");
    +desc formatted orc_create_people partition(state="Or");
    +
      analyze table orc_create_people partition(state) compute statistics partialscan;
    +desc formatted orc_create_people partition(state="Ca");
    +desc formatted orc_create_people partition(state="Or");

    +analyze table orc_create_people partition(state) compute statistics noscan;
      desc formatted orc_create_people partition(state="Ca");
      desc formatted orc_create_people partition(state="Or");

    @@ -116,8 +128,15 @@ INSERT OVERWRITE TABLE orc_create_people
        SELECT * FROM orc_create_people_staging ORDER BY id;

      set hive.stats.autogather = true;
    +analyze table orc_create_people partition(state) compute statistics;
    +desc formatted orc_create_people partition(state="Ca");
    +desc formatted orc_create_people partition(state="Or");
    +
      analyze table orc_create_people partition(state) compute statistics partialscan;
    +desc formatted orc_create_people partition(state="Ca");
    +desc formatted orc_create_people partition(state="Or");

    +analyze table orc_create_people partition(state) compute statistics noscan;
      desc formatted orc_create_people partition(state="Ca");
      desc formatted orc_create_people partition(state="Or");

    @@ -174,8 +193,15 @@ ALTER TABLE orc_create_people SET SERDE
      ALTER TABLE orc_create_people SET FILEFORMAT ORC;

      set hive.stats.autogather = true;
    -analyze table orc_create_people partition(state) compute statistics noscan;
    +analyze table orc_create_people partition(state) compute statistics;
    +desc formatted orc_create_people partition(state="Ca");
    +desc formatted orc_create_people partition(state="OH");

    +analyze table orc_create_people partition(state) compute statistics partialscan;
    +desc formatted orc_create_people partition(state="Ca");
    +desc formatted orc_create_people partition(state="OH");
    +
    +analyze table orc_create_people partition(state) compute statistics noscan;
      desc formatted orc_create_people partition(state="Ca");
      desc formatted orc_create_people partition(state="OH");


    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/parallel_join0.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/parallel_join0.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/parallel_join0.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/parallel_join0.q Sat Feb 21 02:36:54 2015
    @@ -1,6 +1,6 @@
      set mapreduce.job.reduces=4;

    --- SORT_BEFORE_DIFF
    +-- SORT_QUERY_RESULTS

      EXPLAIN
      SELECT src1.key as k1, src1.value as v1,

    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/parquet_types.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/parquet_types.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/parquet_types.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/parquet_types.q Sat Feb 21 02:36:54 2015
    @@ -55,8 +55,8 @@ SELECT ctinyint,
        MAX(cint),
        MIN(csmallint),
        COUNT(cstring1),
    - AVG(cfloat),
    - STDDEV_POP(cdouble)
    + ROUND(AVG(cfloat), 5),
    + ROUND(STDDEV_POP(cdouble),5)
      FROM parquet_types
      GROUP BY ctinyint
      ORDER BY ctinyint

    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/ptf.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/ptf.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/ptf.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/ptf.q Sat Feb 21 02:36:54 2015
    @@ -202,7 +202,7 @@ dense_rank() over (partition by p_mfgr o
      sum(p_retailprice) over (partition by p_mfgr order by p_name rows between unbounded preceding and current row) as s1
      from noop(on noopwithmap(on noop(on part
      partition by p_mfgr
    -order by p_mfgr, p_name
    +order by p_mfgr DESC, p_name
      )));

      select p_mfgr, p_name, p_size,
    @@ -211,7 +211,7 @@ dense_rank() over (partition by p_mfgr o
      sum(p_retailprice) over (partition by p_mfgr order by p_name rows between unbounded preceding and current row) as s1
      from noop(on noopwithmap(on noop(on part
      partition by p_mfgr
    -order by p_mfgr, p_name
    +order by p_mfgr DESC, p_name
      )));

      -- 13. testPTFAndWindowingInSubQ

    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/ptf_matchpath.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/ptf_matchpath.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/ptf_matchpath.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/ptf_matchpath.q Sat Feb 21 02:36:54 2015
    @@ -15,6 +15,17 @@ LOAD DATA LOCAL INPATH '../../data/files
      -- SORT_QUERY_RESULTS

      -- 1. basic Matchpath test
    +explain
    +select origin_city_name, fl_num, year, month, day_of_month, sz, tpath
    +from matchpath(on
    + flights_tiny
    + distribute by fl_num
    + sort by year, month, day_of_month
    + arg1('LATE.LATE+'),
    + arg2('LATE'), arg3(arr_delay > 15),
    + arg4('origin_city_name, fl_num, year, month, day_of_month, size(tpath) as sz, tpath[0].day_of_month as tpath')
    + );
    +
      select origin_city_name, fl_num, year, month, day_of_month, sz, tpath
      from matchpath(on
              flights_tiny
    @@ -26,6 +37,17 @@ from matchpath(on
         );

      -- 2. Matchpath on 1 partition
    +explain
    +select origin_city_name, fl_num, year, month, day_of_month, sz, tpath
    +from matchpath(on
    + flights_tiny
    + sort by fl_num, year, month, day_of_month
    + arg1('LATE.LATE+'),
    + arg2('LATE'), arg3(arr_delay > 15),
    + arg4('origin_city_name, fl_num, year, month, day_of_month, size(tpath) as sz, tpath[0].day_of_month as tpath')
    + )
    +where fl_num = 1142;
    +
      select origin_city_name, fl_num, year, month, day_of_month, sz, tpath
      from matchpath(on
              flights_tiny
    @@ -37,6 +59,17 @@ from matchpath(on
      where fl_num = 1142;

      -- 3. empty partition.
    +explain
    +select origin_city_name, fl_num, year, month, day_of_month, sz, tpath
    +from matchpath(on
    + (select * from flights_tiny where fl_num = -1142) flights_tiny
    + sort by fl_num, year, month, day_of_month
    + arg1('LATE.LATE+'),
    + arg2('LATE'), arg3(arr_delay > 15),
    + arg4('origin_city_name, fl_num, year, month, day_of_month, size(tpath) as sz, tpath[0].day_of_month as tpath')
    + );
    +
    +
      select origin_city_name, fl_num, year, month, day_of_month, sz, tpath
      from matchpath(on
              (select * from flights_tiny where fl_num = -1142) flights_tiny

    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/script_env_var1.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/script_env_var1.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/script_env_var1.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/script_env_var1.q Sat Feb 21 02:36:54 2015
    @@ -1,5 +1,5 @@
      -- Verifies that script operator ID environment variables have unique values
      -- in each instance of the script operator.
      SELECT count(1) FROM
    -( SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM src LIMIT 1 UNION ALL
    - SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM src LIMIT 1 ) a GROUP BY key;
    +( SELECT * FROM (SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM src LIMIT 1)x UNION ALL
    + SELECT * FROM (SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM src LIMIT 1)y ) a GROUP BY key;

    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/script_env_var2.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/script_env_var2.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/script_env_var2.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/script_env_var2.q Sat Feb 21 02:36:54 2015
    @@ -1,5 +1,5 @@
      set hive.script.operator.id.env.var = MY_ID;
      -- Same test as script_env_var1, but test setting the variable name
      SELECT count(1) FROM
    -( SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1 UNION ALL
    - SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1 ) a GROUP BY key;
    +( SELECT * FROM (SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1)a UNION ALL
    + SELECT * FROM (SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1)b ) a GROUP BY key;

    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/udaf_covar_pop.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/udaf_covar_pop.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/udaf_covar_pop.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/udaf_covar_pop.q Sat Feb 21 02:36:54 2015
    @@ -11,6 +11,6 @@ SELECT covar_pop(b, c) FROM covar_tab WH
      SELECT covar_pop(b, c) FROM covar_tab WHERE a < 3;
      SELECT covar_pop(b, c) FROM covar_tab WHERE a = 3;
      SELECT a, covar_pop(b, c) FROM covar_tab GROUP BY a ORDER BY a;
    -SELECT covar_pop(b, c) FROM covar_tab;
    +SELECT ROUND(covar_pop(b, c), 5) FROM covar_tab;

      DROP TABLE covar_tab;

    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/udaf_covar_samp.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/udaf_covar_samp.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/udaf_covar_samp.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/udaf_covar_samp.q Sat Feb 21 02:36:54 2015
    @@ -11,6 +11,6 @@ SELECT covar_samp(b, c) FROM covar_tab W
      SELECT covar_samp(b, c) FROM covar_tab WHERE a < 3;
      SELECT covar_samp(b, c) FROM covar_tab WHERE a = 3;
      SELECT a, covar_samp(b, c) FROM covar_tab GROUP BY a ORDER BY a;
    -SELECT covar_samp(b, c) FROM covar_tab;
    +SELECT ROUND(covar_samp(b, c), 5) FROM covar_tab;

      DROP TABLE covar_tab;

    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/udf_add_months.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/udf_add_months.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/udf_add_months.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/udf_add_months.q Sat Feb 21 02:36:54 2015
    @@ -1,2 +1,43 @@
      DESCRIBE FUNCTION add_months;
      DESCRIBE FUNCTION EXTENDED add_months;
    +
    +explain select add_months('2014-01-14', 1);
    +
    +select
    +add_months('2014-01-14', 1),
    +add_months('2014-01-31', 1),
    +add_months('2014-02-28', -1),
    +add_months('2014-02-28', 2),
    +add_months('2014-04-30', -2),
    +add_months('2015-02-28', 12),
    +add_months('2016-02-29', -12),
    +add_months('2016-01-29', 1),
    +add_months('2016-02-29', -1),
    +add_months('2014-01-32', 1),
    +add_months('01/14/2014', 1),
    +add_months(cast(null as string), 1),
    +add_months('2014-01-14', cast(null as int));
    +
    +select
    +add_months('2014-01-14 10:30:00', 1),
    +add_months('2014-01-31 10:30:00', 1),
    +add_months('2014-02-28 10:30:00', -1),
    +add_months('2014-02-28 16:30:00', 2),
    +add_months('2014-04-30 10:30:00', -2),
    +add_months('2015-02-28 10:30:00', 12),
    +add_months('2016-02-29 10:30:00', -12),
    +add_months('2016-01-29 10:30:00', 1),
    +add_months('2016-02-29 10:30:00', -1),
    +add_months('2014-01-32 10:30:00', 1);
    +
    +select
    +add_months(cast('2014-01-14 10:30:00' as timestamp), 1),
    +add_months(cast('2014-01-31 10:30:00' as timestamp), 1),
    +add_months(cast('2014-02-28 10:30:00' as timestamp), -1),
    +add_months(cast('2014-02-28 16:30:00' as timestamp), 2),
    +add_months(cast('2014-04-30 10:30:00' as timestamp), -2),
    +add_months(cast('2015-02-28 10:30:00' as timestamp), 12),
    +add_months(cast('2016-02-29 10:30:00' as timestamp), -12),
    +add_months(cast('2016-01-29 10:30:00' as timestamp), 1),
    +add_months(cast('2016-02-29 10:30:00' as timestamp), -1),
    +add_months(cast(null as timestamp), 1);
    \ No newline at end of file

    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/udf_last_day.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/udf_last_day.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/udf_last_day.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/udf_last_day.q Sat Feb 21 02:36:54 2015
    @@ -1,2 +1,52 @@
      DESCRIBE FUNCTION last_day;
      DESCRIBE FUNCTION EXTENDED last_day;
    +
    +explain select last_day('2015-02-05');
    +
    +select
    +last_day('2014-01-01'),
    +last_day('2014-01-14'),
    +last_day('2014-01-31'),
    +last_day('2014-02-02'),
    +last_day('2014-02-28'),
    +last_day('2016-02-03'),
    +last_day('2016-02-28'),
    +last_day('2016-02-29'),
    +last_day('2014-01-34'),
    +last_day(cast(null as string)),
    +last_day('01/29/2014');
    +
    +select
    +last_day('2014-01-01 10:30:45'),
    +last_day('2014-01-14 10:30:45'),
    +last_day('2014-01-31 10:30:45'),
    +last_day('2014-02-02 10:30:45'),
    +last_day('2014-02-28 10:30:45'),
    +last_day('2016-02-03 10:30:45'),
    +last_day('2016-02-28 10:30:45'),
    +last_day('2016-02-29 10:30:45'),
    +last_day('2014-01-34 10:30:45'),
    +last_day(cast(null as string)),
    +last_day('01/29/2014 10:30:45');
    +
    +select
    +last_day(cast('2014-01-01' as date)),
    +last_day(cast('2014-01-14' as date)),
    +last_day(cast('2014-01-31' as date)),
    +last_day(cast('2014-02-02' as date)),
    +last_day(cast('2014-02-28' as date)),
    +last_day(cast('2016-02-03' as date)),
    +last_day(cast('2016-02-28' as date)),
    +last_day(cast('2016-02-29' as date)),
    +last_day(cast(null as date));
    +
    +select
    +last_day(cast('2014-01-01 10:30:45' as timestamp)),
    +last_day(cast('2014-01-14 10:30:45' as timestamp)),
    +last_day(cast('2014-01-31 10:30:45' as timestamp)),
    +last_day(cast('2014-02-02 10:30:45' as timestamp)),
    +last_day(cast('2014-02-28 10:30:45' as timestamp)),
    +last_day(cast('2016-02-03 10:30:45' as timestamp)),
    +last_day(cast('2016-02-28 10:30:45' as timestamp)),
    +last_day(cast('2016-02-29 10:30:45' as timestamp)),
    +last_day(cast(null as timestamp));
    \ No newline at end of file

    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/union3.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/union3.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/union3.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/union3.q Sat Feb 21 02:36:54 2015
    @@ -1,21 +1,20 @@
    --- SORT_BEFORE_DIFF
    +-- SORT_QUERY_RESULTS

      explain
      SELECT *
      FROM (
        SELECT 1 AS id
        FROM (SELECT * FROM src LIMIT 1) s1
    - CLUSTER BY id
        UNION ALL
        SELECT 2 AS id
        FROM (SELECT * FROM src LIMIT 1) s1
    - CLUSTER BY id
        UNION ALL
        SELECT 3 AS id
        FROM (SELECT * FROM src LIMIT 1) s2
        UNION ALL
        SELECT 4 AS id
        FROM (SELECT * FROM src LIMIT 1) s2
    + CLUSTER BY id
      ) a;


    @@ -27,17 +26,16 @@ SELECT *
      FROM (
        SELECT 1 AS id
        FROM (SELECT * FROM src LIMIT 1) s1
    - CLUSTER BY id
        UNION ALL
        SELECT 2 AS id
        FROM (SELECT * FROM src LIMIT 1) s1
    - CLUSTER BY id
        UNION ALL
        SELECT 3 AS id
        FROM (SELECT * FROM src LIMIT 1) s2
        UNION ALL
        SELECT 4 AS id
        FROM (SELECT * FROM src LIMIT 1) s2
    + CLUSTER BY id
      ) a;

    -select * from union_out cluster by id;
    +select * from union_out;

    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/union4.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/union4.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/union4.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/union4.q Sat Feb 21 02:36:54 2015
    @@ -1,5 +1,7 @@
      set hive.map.aggr = true;

    +-- SORT_QUERY_RESULTS
    +
      -- union case: both subqueries are map-reduce jobs on same input, followed by filesink



    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/union_null.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/union_null.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/union_null.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/union_null.q Sat Feb 21 02:36:54 2015
    @@ -1,7 +1,7 @@
      -- SORT_BEFORE_DIFF

      -- HIVE-2901
    -select x from (select value as x from src order by x limit 5 union all select NULL as x from src limit 5) a;
    +select x from (select * from (select value as x from src order by x limit 5)a union all select * from (select NULL as x from src limit 5)b )a;

      -- HIVE-4837
    -select * from (select null as N from src1 group by key UNION ALL select null as N from src1 group by key ) a;
    +select * from (select * from (select null as N from src1 group by key)a UNION ALL select * from (select null as N from src1 group by key)b ) a;

    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/union_remove_25.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/union_remove_25.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/union_remove_25.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/union_remove_25.q Sat Feb 21 02:36:54 2015
    @@ -51,17 +51,17 @@ explain
      insert overwrite table outputTbl2 partition(ds)
      SELECT *
      FROM (
    - SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500
    + select * from (SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500)a
        UNION ALL
    - SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500
    + select * from (SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500)b
      ) a;

      insert overwrite table outputTbl2 partition(ds)
      SELECT *
      FROM (
    - SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500
    + select * from (SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500)a
        UNION ALL
    - SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500
    + select * from (SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500)b
      ) a;

      show partitions outputTbl2;
    @@ -70,17 +70,17 @@ desc formatted outputTbl2 partition(ds='
      explain insert overwrite table outputTbl3 partition(ds, hr)
      SELECT *
      FROM (
    - SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000
    + select * from (SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000)a
        UNION ALL
    - SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000
    + select * from (SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000)b
      ) a;

      insert overwrite table outputTbl3 partition(ds, hr)
      SELECT *
      FROM (
    - SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000
    + select * from (SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000)a
        UNION ALL
    - SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000
    + select * from (SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000)b
      ) a;

      show partitions outputTbl3;

    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/union_top_level.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/union_top_level.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/union_top_level.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/union_top_level.q Sat Feb 21 02:36:54 2015
    @@ -2,42 +2,42 @@

      -- top level
      explain
    -select key, 0 as value from src where key % 3 == 0 limit 3
    +select * from (select key, 0 as value from src where key % 3 == 0 limit 3)a
      union all
    -select key, 1 as value from src where key % 3 == 1 limit 3
    +select * from (select key, 1 as value from src where key % 3 == 1 limit 3)b
      union all
    -select key, 2 as value from src where key % 3 == 2 limit 3;
    +select * from (select key, 2 as value from src where key % 3 == 2 limit 3)c;

    -select key, 0 as value from src where key % 3 == 0 limit 3
    +select * from (select key, 0 as value from src where key % 3 == 0 limit 3)a
      union all
    -select key, 1 as value from src where key % 3 == 1 limit 3
    +select * from (select key, 1 as value from src where key % 3 == 1 limit 3)b
      union all
    -select key, 2 as value from src where key % 3 == 2 limit 3;
    +select * from (select key, 2 as value from src where key % 3 == 2 limit 3)c;

      explain
    -select s1.key as k, s2.value as v from src s1 join src s2 on (s1.key = s2.key) limit 10
    +select * from (select s1.key as k, s2.value as v from src s1 join src s2 on (s1.key = s2.key) limit 10)a
      union all
    -select s1.key as k, s2.value as v from src s1 join src s2 on (s1.key = s2.key) limit 10;
    +select * from (select s1.key as k, s2.value as v from src s1 join src s2 on (s1.key = s2.key) limit 10)b;

    -select s1.key as k, s2.value as v from src s1 join src s2 on (s1.key = s2.key) limit 10
    +select * from (select s1.key as k, s2.value as v from src s1 join src s2 on (s1.key = s2.key) limit 10)a
      union all
    -select s1.key as k, s2.value as v from src s1 join src s2 on (s1.key = s2.key) limit 10;
    +select * from (select s1.key as k, s2.value as v from src s1 join src s2 on (s1.key = s2.key) limit 10)b;

      -- ctas
      explain
      create table union_top as
    -select key, 0 as value from src where key % 3 == 0 limit 3
    +select * from (select key, 0 as value from src where key % 3 == 0 limit 3)a
      union all
    -select key, 1 as value from src where key % 3 == 1 limit 3
    +select * from (select key, 1 as value from src where key % 3 == 1 limit 3)b
      union all
    -select key, 2 as value from src where key % 3 == 2 limit 3;
    +select * from (select key, 2 as value from src where key % 3 == 2 limit 3)c;

      create table union_top as
    -select key, 0 as value from src where key % 3 == 0 limit 3
    +select * from (select key, 0 as value from src where key % 3 == 0 limit 3)a
      union all
    -select key, 1 as value from src where key % 3 == 1 limit 3
    +select * from (select key, 1 as value from src where key % 3 == 1 limit 3)b
      union all
    -select key, 2 as value from src where key % 3 == 2 limit 3;
    +select * from (select key, 2 as value from src where key % 3 == 2 limit 3)c;

      select * from union_top;

    @@ -46,53 +46,53 @@ truncate table union_top;
      -- insert into
      explain
      insert into table union_top
    -select key, 0 as value from src where key % 3 == 0 limit 3
    +select * from (select key, 0 as value from src where key % 3 == 0 limit 3)a
      union all
    -select key, 1 as value from src where key % 3 == 1 limit 3
    +select * from (select key, 1 as value from src where key % 3 == 1 limit 3)b
      union all
    -select key, 2 as value from src where key % 3 == 2 limit 3;
    +select * from (select key, 2 as value from src where key % 3 == 2 limit 3)c;

      insert into table union_top
    -select key, 0 as value from src where key % 3 == 0 limit 3
    +select * from (select key, 0 as value from src where key % 3 == 0 limit 3)a
      union all
    -select key, 1 as value from src where key % 3 == 1 limit 3
    +select * from (select key, 1 as value from src where key % 3 == 1 limit 3)b
      union all
    -select key, 2 as value from src where key % 3 == 2 limit 3;
    +select * from (select key, 2 as value from src where key % 3 == 2 limit 3)c;

      select * from union_top;

      explain
      insert overwrite table union_top
    -select key, 0 as value from src where key % 3 == 0 limit 3
    +select * from (select key, 0 as value from src where key % 3 == 0 limit 3)a
      union all
    -select key, 1 as value from src where key % 3 == 1 limit 3
    +select * from (select key, 1 as value from src where key % 3 == 1 limit 3)b
      union all
    -select key, 2 as value from src where key % 3 == 2 limit 3;
    +select * from (select key, 2 as value from src where key % 3 == 2 limit 3)c;

      insert overwrite table union_top
    -select key, 0 as value from src where key % 3 == 0 limit 3
    +select * from (select key, 0 as value from src where key % 3 == 0 limit 3)a
      union all
    -select key, 1 as value from src where key % 3 == 1 limit 3
    +select * from (select key, 1 as value from src where key % 3 == 1 limit 3)b
      union all
    -select key, 2 as value from src where key % 3 == 2 limit 3;
    +select * from (select key, 2 as value from src where key % 3 == 2 limit 3)c;

      select * from union_top;

      -- create view
      explain
      create view union_top_view as
    -select key, 0 as value from src where key % 3 == 0 limit 3
    +select * from (select key, 0 as value from src where key % 3 == 0 limit 3)a
      union all
    -select key, 1 as value from src where key % 3 == 1 limit 3
    +select * from (select key, 1 as value from src where key % 3 == 1 limit 3)b
      union all
    -select key, 2 as value from src where key % 3 == 2 limit 3;
    +select * from (select key, 2 as value from src where key % 3 == 2 limit 3)c;

      create view union_top_view as
    -select key, 0 as value from src where key % 3 == 0 limit 3
    +select * from (select key, 0 as value from src where key % 3 == 0 limit 3)a
      union all
    -select key, 1 as value from src where key % 3 == 1 limit 3
    +select * from (select key, 1 as value from src where key % 3 == 1 limit 3)b
      union all
    -select key, 2 as value from src where key % 3 == 2 limit 3;
    +select * from (select key, 2 as value from src where key % 3 == 2 limit 3)c;

      select * from union_top_view;


    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/vector_multi_insert.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/vector_multi_insert.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/vector_multi_insert.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/vector_multi_insert.q Sat Feb 21 02:36:54 2015
    @@ -8,11 +8,11 @@ create table orc1
          select rn
          from
          (
    - select cast(1 as int) as rn from src limit 1
    + select * from (select cast(1 as int) as rn from src limit 1)a
            union all
    - select cast(100 as int) as rn from src limit 1
    + select * from (select cast(100 as int) as rn from src limit 1)b
            union all
    - select cast(10000 as int) as rn from src limit 1
    + select * from (select cast(10000 as int) as rn from src limit 1)c
          ) t;

      create table orc_rn1 (rn int);

    Modified: hive/branches/llap/ql/src/test/queries/clientpositive/windowing_windowspec.q
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/queries/clientpositive/windowing_windowspec.q?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/queries/clientpositive/windowing_windowspec.q (original)
    +++ hive/branches/llap/ql/src/test/queries/clientpositive/windowing_windowspec.q Sat Feb 21 02:36:54 2015
    @@ -34,3 +34,7 @@ select f, sum(f) over (partition by ts o
      select s, i, round(avg(d) over (partition by s order by i) / 10.0 , 2) from over10k limit 7;

      select s, i, round((avg(d) over w1 + 10.0) - (avg(d) over w1 - 10.0),2) from over10k window w1 as (partition by s order by i) limit 7;
    +
    +set hive.cbo.enable=false;
    +-- HIVE-9228
    +select s, i from ( select s, i, round((avg(d) over w1 + 10.0) - (avg(d) over w1 - 10.0),2) from over10k window w1 as (partition by s order by i)) X limit 7;

    Modified: hive/branches/llap/ql/src/test/resources/orc-file-dump-dictionary-threshold.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/resources/orc-file-dump-dictionary-threshold.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/resources/orc-file-dump-dictionary-threshold.out (original)
    +++ hive/branches/llap/ql/src/test/resources/orc-file-dump-dictionary-threshold.out Sat Feb 21 02:36:54 2015
    @@ -52,24 +52,24 @@ Stripes:
          Encoding column 1: DIRECT_V2
          Encoding column 2: DIRECT_V2
          Encoding column 3: DIRECT_V2
    - Row group index column 1:
    - Entry 0:count: 1000 hasNull: false min: -2132329551 max: 2145911404 sum: 61941331718 positions: 0,0,0
    - Entry 1:count: 1000 hasNull: false min: -2138433136 max: 2145210552 sum: 14574030042 positions: 0,2050,488
    - Entry 2:count: 1000 hasNull: false min: -2147115959 max: 2137805337 sum: -2032493169 positions: 0,6150,464
    - Entry 3:count: 1000 hasNull: false min: -2137828953 max: 2145877119 sum: -3167202608 positions: 10003,250,440
    - Entry 4:count: 1000 hasNull: false min: -2146452517 max: 2142394906 sum: 88361503212 positions: 10003,4350,416
    - Row group index column 2:
    - Entry 0:count: 1000 hasNull: false min: -9206837518492372266 max: 9169230975203934579 positions: 0,0,0
    - Entry 1:count: 1000 hasNull: false min: -9188878639954124284 max: 9213664245516510068 positions: 0,4098,488
    - Entry 2:count: 1000 hasNull: false min: -9211329013123260308 max: 9217851628057711416 positions: 10003,2294,464
    - Entry 3:count: 1000 hasNull: false min: -9185745718227889962 max: 9181722705210917931 positions: 20006,490,440
    - Entry 4:count: 1000 hasNull: false min: -9216505819108477308 max: 9196474183833079923 positions: 20006,8686,416
    - Row group index column 3:
    - Entry 0:count: 1000 hasNull: false min: Darkness,-230 max: worst-54-290-346-648-908-996 sum: 18442 positions: 0,0,0,0,0
    - Entry 1:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966 sum: 46338 positions: 2777,8442,0,695,18
    - Entry 2:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660 sum: 75448 positions: 13595,4780,0,1554,14
    - Entry 3:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788 sum: 104868 positions: 31432,228,0,2372,90
    - Entry 4:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744 sum: 136158 positions: 54111,5096,0,3354,108
    + Row group indices for column 1:
    + Entry 0: count: 1000 hasNull: false min: -2132329551 max: 2145911404 sum: 61941331718 positions: 0,0,0
    + Entry 1: count: 1000 hasNull: false min: -2138433136 max: 2145210552 sum: 14574030042 positions: 0,2050,488
    + Entry 2: count: 1000 hasNull: false min: -2147115959 max: 2137805337 sum: -2032493169 positions: 0,6150,464
    + Entry 3: count: 1000 hasNull: false min: -2137828953 max: 2145877119 sum: -3167202608 positions: 10003,250,440
    + Entry 4: count: 1000 hasNull: false min: -2146452517 max: 2142394906 sum: 88361503212 positions: 10003,4350,416
    + Row group indices for column 2:
    + Entry 0: count: 1000 hasNull: false min: -9206837518492372266 max: 9169230975203934579 positions: 0,0,0
    + Entry 1: count: 1000 hasNull: false min: -9188878639954124284 max: 9213664245516510068 positions: 0,4098,488
    + Entry 2: count: 1000 hasNull: false min: -9211329013123260308 max: 9217851628057711416 positions: 10003,2294,464
    + Entry 3: count: 1000 hasNull: false min: -9185745718227889962 max: 9181722705210917931 positions: 20006,490,440
    + Entry 4: count: 1000 hasNull: false min: -9216505819108477308 max: 9196474183833079923 positions: 20006,8686,416
    + Row group indices for column 3:
    + Entry 0: count: 1000 hasNull: false min: Darkness,-230 max: worst-54-290-346-648-908-996 sum: 18442 positions: 0,0,0,0,0
    + Entry 1: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966 sum: 46338 positions: 2777,8442,0,695,18
    + Entry 2: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660 sum: 75448 positions: 13595,4780,0,1554,14
    + Entry 3: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788 sum: 104868 positions: 31432,228,0,2372,90
    + Entry 4: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744 sum: 136158 positions: 54111,5096,0,3354,108
        Stripe: offset: 151897 data: 336358 rows: 5000 tail: 69 index: 954
          Stream: column 0 section ROW_INDEX start: 151897 length 17
          Stream: column 1 section ROW_INDEX start: 151914 length 153
    @@ -83,24 +83,24 @@ Stripes:
          Encoding column 1: DIRECT_V2
          Encoding column 2: DIRECT_V2
          Encoding column 3: DIRECT_V2
    - Row group index column 1:
    - Entry 0:count: 1000 hasNull: false min: -2146021688 max: 2146838901 sum: -50979197646 positions: 0,0,0
    - Entry 1:count: 1000 hasNull: false min: -2143569489 max: 2141223179 sum: 22810066834 positions: 0,2050,488
    - Entry 2:count: 1000 hasNull: false min: -2140649392 max: 2146301701 sum: -31694882346 positions: 0,6150,464
    - Entry 3:count: 1000 hasNull: false min: -2147390285 max: 2146299933 sum: 79371934221 positions: 10003,250,440
    - Entry 4:count: 1000 hasNull: false min: -2145928262 max: 2147224606 sum: -34469378822 positions: 10003,4350,416
    - Row group index column 2:
    - Entry 0:count: 1000 hasNull: false min: -9222178666167296739 max: 9191250610515369723 positions: 0,0,0
    - Entry 1:count: 1000 hasNull: false min: -9220148577547102875 max: 9213945522531717278 positions: 0,4098,488
    - Entry 2:count: 1000 hasNull: false min: -9220818777591257749 max: 9221301751385928177 positions: 10003,2294,464
    - Entry 3:count: 1000 hasNull: false min: -9220031433030423388 max: 9207856144487414148 positions: 20006,490,440
    - Entry 4:count: 1000 hasNull: false min: -9201438531577205959 max: 9212462124593119846 positions: 20006,8686,416
    - Row group index column 3:
    - Entry 0:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726 sum: 166320 positions: 0,0,0,0,0
    - Entry 1:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994 sum: 193436 positions: 37112,6320,0,967,90
    - Entry 2:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988 sum: 224740 positions: 80822,9756,0,1945,222
    - Entry 3:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984 sum: 252094 positions: 137149,4496,0,3268,48
    - Entry 4:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938 sum: 281404 positions: 197972,6590,0,4064,342
    + Row group indices for column 1:
    + Entry 0: count: 1000 hasNull: false min: -2146021688 max: 2146838901 sum: -50979197646 positions: 0,0,0
    + Entry 1: count: 1000 hasNull: false min: -2143569489 max: 2141223179 sum: 22810066834 positions: 0,2050,488
    + Entry 2: count: 1000 hasNull: false min: -2140649392 max: 2146301701 sum: -31694882346 positions: 0,6150,464
    + Entry 3: count: 1000 hasNull: false min: -2147390285 max: 2146299933 sum: 79371934221 positions: 10003,250,440
    + Entry 4: count: 1000 hasNull: false min: -2145928262 max: 2147224606 sum: -34469378822 positions: 10003,4350,416
    + Row group indices for column 2:
    + Entry 0: count: 1000 hasNull: false min: -9222178666167296739 max: 9191250610515369723 positions: 0,0,0
    + Entry 1: count: 1000 hasNull: false min: -9220148577547102875 max: 9213945522531717278 positions: 0,4098,488
    + Entry 2: count: 1000 hasNull: false min: -9220818777591257749 max: 9221301751385928177 positions: 10003,2294,464
    + Entry 3: count: 1000 hasNull: false min: -9220031433030423388 max: 9207856144487414148 positions: 20006,490,440
    + Entry 4: count: 1000 hasNull: false min: -9201438531577205959 max: 9212462124593119846 positions: 20006,8686,416
    + Row group indices for column 3:
    + Entry 0: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726 sum: 166320 positions: 0,0,0,0,0
    + Entry 1: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994 sum: 193436 positions: 37112,6320,0,967,90
    + Entry 2: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988 sum: 224740 positions: 80822,9756,0,1945,222
    + Entry 3: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984 sum: 252094 positions: 137149,4496,0,3268,48
    + Entry 4: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938 sum: 281404 positions: 197972,6590,0,4064,342
        Stripe: offset: 489278 data: 558031 rows: 5000 tail: 69 index: 1181
          Stream: column 0 section ROW_INDEX start: 489278 length 17
          Stream: column 1 section ROW_INDEX start: 489295 length 166
    @@ -114,24 +114,24 @@ Stripes:
          Encoding column 1: DIRECT_V2
          Encoding column 2: DIRECT_V2
          Encoding column 3: DIRECT_V2
    - Row group index column 1:
    - Entry 0:count: 1000 hasNull: false min: -2138229212 max: 2144818981 sum: -22823642812 positions: 0,0,0
    - Entry 1:count: 1000 hasNull: false min: -2145842720 max: 2144179881 sum: -12562754334 positions: 0,2050,488
    - Entry 2:count: 1000 hasNull: false min: -2143045885 max: 2146718321 sum: 82993638644 positions: 0,6150,464
    - Entry 3:count: 1000 hasNull: false min: -2144745617 max: 2146570474 sum: 25138722367 positions: 10003,250,440
    - Entry 4:count: 1000 hasNull: false min: -2140127150 max: 2135081620 sum: 68346511655 positions: 10003,4350,416
    - Row group index column 2:
    - Entry 0:count: 1000 hasNull: false min: -9204340807292138409 max: 9208698732685326961 positions: 0,0,0
    - Entry 1:count: 1000 hasNull: false min: -9221963099397084326 max: 9222722740629726770 positions: 0,4098,488
    - Entry 2:count: 1000 hasNull: false min: -9210480084701091299 max: 9207767402467343058 positions: 10003,2294,464
    - Entry 3:count: 1000 hasNull: false min: -9195038026813631215 max: 9199201928563274421 positions: 20006,490,440
    - Entry 4:count: 1000 hasNull: false min: -9215483580266514322 max: 9220102792864959501 positions: 20006,8686,416
    - Row group index column 3:
    - Entry 0:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876 sum: 313880 positions: 0,0,0,0,0
    - Entry 1:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-11964 sum: 349542 positions: 80352,3880,0,1097,28
    - Entry 2:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-11964-12054-12096-12126-12136-12202-12246-12298-12616-12774-12782-12790-12802-12976 sum: 386538 positi
      ons: 170641,3422,0,2077,162
    - Entry 3:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022-12178-12418-12832-13304 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-11964-12054-12096-12126-12136-12202-12246-12298-12616-12774-12782-12790-12802-
      12976-13216-13246-13502-13766 sum: 421660 positions: 268420,9960,0,3369,16
    - Entry 4:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022-12178-12418-12832-13304-13448-13590-13618-13908-14188 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-11964-12054-12096-12126-12136-12202-12246-12298-
      12616-12774-12782-12790-12802-12976-13216-13246-13502-13766-14454-14974 sum: 453606 positions: 377916,1620,0,4041,470
    + Row group indices for column 1:
    + Entry 0: count: 1000 hasNull: false min: -2138229212 max: 2144818981 sum: -22823642812 positions: 0,0,0
    + Entry 1: count: 1000 hasNull: false min: -2145842720 max: 2144179881 sum: -12562754334 positions: 0,2050,488
    + Entry 2: count: 1000 hasNull: false min: -2143045885 max: 2146718321 sum: 82993638644 positions: 0,6150,464
    + Entry 3: count: 1000 hasNull: false min: -2144745617 max: 2146570474 sum: 25138722367 positions: 10003,250,440
    + Entry 4: count: 1000 hasNull: false min: -2140127150 max: 2135081620 sum: 68346511655 positions: 10003,4350,416
    + Row group indices for column 2:
    + Entry 0: count: 1000 hasNull: false min: -9204340807292138409 max: 9208698732685326961 positions: 0,0,0
    + Entry 1: count: 1000 hasNull: false min: -9221963099397084326 max: 9222722740629726770 positions: 0,4098,488
    + Entry 2: count: 1000 hasNull: false min: -9210480084701091299 max: 9207767402467343058 positions: 10003,2294,464
    + Entry 3: count: 1000 hasNull: false min: -9195038026813631215 max: 9199201928563274421 positions: 20006,490,440
    + Entry 4: count: 1000 hasNull: false min: -9215483580266514322 max: 9220102792864959501 positions: 20006,8686,416
    + Row group indices for column 3:
    + Entry 0: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876 sum: 313880 positions: 0,0,0,0,0
    + Entry 1: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-11964 sum: 349542 positions: 80352,3880,0,1097,28
    + Entry 2: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-11964-12054-12096-12126-12136-12202-12246-12298-12616-12774-12782-12790-12802-12976 sum: 386538 posit
      ions: 170641,3422,0,2077,162
    + Entry 3: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022-12178-12418-12832-13304 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-11964-12054-12096-12126-12136-12202-12246-12298-12616-12774-12782-12790-12802
      -12976-13216-13246-13502-13766 sum: 421660 positions: 268420,9960,0,3369,16
    + Entry 4: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022-12178-12418-12832-13304-13448-13590-13618-13908-14188 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-11964-12054-12096-12126-12136-12202-12246-12298
      -12616-12774-12782-12790-12802-12976-13216-13246-13502-13766-14454-14974 sum: 453606 positions: 377916,1620,0,4041,470
        Stripe: offset: 1048559 data: 792850 rows: 5000 tail: 69 index: 1369
          Stream: column 0 section ROW_INDEX start: 1048559 length 17
          Stream: column 1 section ROW_INDEX start: 1048576 length 153
    @@ -145,24 +145,24 @@ Stripes:
          Encoding column 1: DIRECT_V2
          Encoding column 2: DIRECT_V2
          Encoding column 3: DIRECT_V2
    - Row group index column 1:
    - Entry 0:count: 1000 hasNull: false min: -2145319330 max: 2146998132 sum: -50856753363 positions: 0,0,0
    - Entry 1:count: 1000 hasNull: false min: -2134288866 max: 2147453086 sum: -17911019023 positions: 0,2050,488
    - Entry 2:count: 1000 hasNull: false min: -2139010804 max: 2144727593 sum: -24993151857 positions: 0,6150,464
    - Entry 3:count: 1000 hasNull: false min: -2145378214 max: 2144098933 sum: -18055164052 positions: 10003,250,440
    - Entry 4:count: 1000 hasNull: false min: -2140494429 max: 2144595861 sum: -41863916235 positions: 10003,4350,416
    - Row group index column 2:
    - Entry 0:count: 1000 hasNull: false min: -9172774601303513941 max: 9212917101275642143 positions: 0,0,0
    - Entry 1:count: 1000 hasNull: false min: -9218164880949195469 max: 9222919052987871506 positions: 0,4098,488
    - Entry 2:count: 1000 hasNull: false min: -9222731174895935707 max: 9214167447015056056 positions: 10003,2294,464
    - Entry 3:count: 1000 hasNull: false min: -9196276654247395117 max: 9210639275226058005 positions: 20006,490,440
    - Entry 4:count: 1000 hasNull: false min: -9197393848859294562 max: 9208134757538374043 positions: 20006,8686,416
    - Row group index column 3:
    - Entry 0:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022-12178-12418-12832-13304-13448-13590-13618-13908-14188-14246-14340-14364-14394-14762-14850-14964-15048 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-
      11964-12054-12096-12126-12136-12202-12246-12298-12616-12774-12782-12790-12802-12976-13216-13246-13502-13766-14454-14974-15004-15124-15252-15294-15356-15530-15610 sum: 492916 positions: 0,0,0,0,0
    - Entry 1:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022-12178-12418-12832-13304-13448-13590-13618-13908-14188-14246-14340-14364-14394-14762-14850-14964-15048-15494-15674-15726-16006 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-
      11158-11410-11722-11836-11964-12054-12096-12126-12136-12202-12246-12298-12616-12774-12782-12790-12802-12976-13216-13246-13502-13766-14454-14974-15004-15124-15252-15294-15356-15530-15610-16316-16936 sum: 527290 positions: 126968,2916,0,1077,140
    - Entry 2:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022-12178-12418-12832-13304-13448-13590-13618-13908-14188-14246-14340-14364-14394-14762-14850-14964-15048-15494-15674-15726-16006-16056-16180-16304-16332-16452-16598-16730-16810-16994-17210 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-96
      50-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-11964-12054-12096-12126-12136-12202-12246-12298-12616-12774-12782-12790-12802-12976-13216-13246-13502-13766-14454-14974-15004-15124-15252-15294-15356-15530-15610-16316-16936-17024-17122-17214-17310-17528-17682-17742-17870-17878 sum: 568274 positions: 263111,206,0,1926,462
    - Entry 3:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022-12178-12418-12832-13304-13448-13590-13618-13908-14188-14246-14340-14364-14394-14762-14850-14964-15048-15494-15674-15726-16006-16056-16180-16304-16332-16452-16598-16730-16810-16994-17210-17268-17786-17962-18214 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-912
      8-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-11964-12054-12096-12126-12136-12202-12246-12298-12616-12774-12782-12790-12802-12976-13216-13246-13502-13766-14454-14974-15004-15124-15252-15294-15356-15530-15610-16316-16936-17024-17122-17214-17310-17528-17682-17742-17870-17878-18010-18410-18524-18788 sum: 594578 positions: 407371,8480,0,3444,250
    - Entry 4:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022-12178-12418-12832-13304-13448-13590-13618-13908-14188-14246-14340-14364-14394-14762-14850-14964-15048-15494-15674-15726-16006-16056-16180-16304-16332-16452-16598-16730-16810-16994-17210-17268-17786-17962-18214-18444-18446-18724-18912-18952-19164 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-83
      90-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-11964-12054-12096-12126-12136-12202-12246-12298-12616-12774-12782-12790-12802-12976-13216-13246-13502-13766-14454-14974-15004-15124-15252-15294-15356-15530-15610-16316-16936-17024-17122-17214-17310-17528-17682-17742-17870-17878-18010-18410-18524-18788-19204-19254-19518-19596-19786-19874-19904 sum: 631944 positions: 562094,3058,0,4643,292
    + Row group indices for column 1:
    + Entry 0: count: 1000 hasNull: false min: -2145319330 max: 2146998132 sum: -50856753363 positions: 0,0,0
    + Entry 1: count: 1000 hasNull: false min: -2134288866 max: 2147453086 sum: -17911019023 positions: 0,2050,488
    + Entry 2: count: 1000 hasNull: false min: -2139010804 max: 2144727593 sum: -24993151857 positions: 0,6150,464
    + Entry 3: count: 1000 hasNull: false min: -2145378214 max: 2144098933 sum: -18055164052 positions: 10003,250,440
    + Entry 4: count: 1000 hasNull: false min: -2140494429 max: 2144595861 sum: -41863916235 positions: 10003,4350,416
    + Row group indices for column 2:
    + Entry 0: count: 1000 hasNull: false min: -9172774601303513941 max: 9212917101275642143 positions: 0,0,0
    + Entry 1: count: 1000 hasNull: false min: -9218164880949195469 max: 9222919052987871506 positions: 0,4098,488
    + Entry 2: count: 1000 hasNull: false min: -9222731174895935707 max: 9214167447015056056 positions: 10003,2294,464
    + Entry 3: count: 1000 hasNull: false min: -9196276654247395117 max: 9210639275226058005 positions: 20006,490,440
    + Entry 4: count: 1000 hasNull: false min: -9197393848859294562 max: 9208134757538374043 positions: 20006,8686,416
    + Row group indices for column 3:
    + Entry 0: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022-12178-12418-12832-13304-13448-13590-13618-13908-14188-14246-14340-14364-14394-14762-14850-14964-15048 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836
      -11964-12054-12096-12126-12136-12202-12246-12298-12616-12774-12782-12790-12802-12976-13216-13246-13502-13766-14454-14974-15004-15124-15252-15294-15356-15530-15610 sum: 492916 positions: 0,0,0,0,0
    + Entry 1: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022-12178-12418-12832-13304-13448-13590-13618-13908-14188-14246-14340-14364-14394-14762-14850-14964-15048-15494-15674-15726-16006 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008
      -11158-11410-11722-11836-11964-12054-12096-12126-12136-12202-12246-12298-12616-12774-12782-12790-12802-12976-13216-13246-13502-13766-14454-14974-15004-15124-15252-15294-15356-15530-15610-16316-16936 sum: 527290 positions: 126968,2916,0,1077,140
    + Entry 2: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022-12178-12418-12832-13304-13448-13590-13618-13908-14188-14246-14340-14364-14394-14762-14850-14964-15048-15494-15674-15726-16006-16056-16180-16304-16332-16452-16598-16730-16810-16994-17210 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9
      650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-11964-12054-12096-12126-12136-12202-12246-12298-12616-12774-12782-12790-12802-12976-13216-13246-13502-13766-14454-14974-15004-15124-15252-15294-15356-15530-15610-16316-16936-17024-17122-17214-17310-17528-17682-17742-17870-17878 sum: 568274 positions: 263111,206,0,1926,462
    + Entry 3: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022-12178-12418-12832-13304-13448-13590-13618-13908-14188-14246-14340-14364-14394-14762-14850-14964-15048-15494-15674-15726-16006-16056-16180-16304-16332-16452-16598-16730-16810-16994-17210-17268-17786-17962-18214 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-91
      28-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-11964-12054-12096-12126-12136-12202-12246-12298-12616-12774-12782-12790-12802-12976-13216-13246-13502-13766-14454-14974-15004-15124-15252-15294-15356-15530-15610-16316-16936-17024-17122-17214-17310-17528-17682-17742-17870-17878-18010-18410-18524-18788 sum: 594578 positions: 407371,8480,0,3444,250
    + Entry 4: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022-12178-12418-12832-13304-13448-13590-13618-13908-14188-14246-14340-14364-14394-14762-14850-14964-15048-15494-15674-15726-16006-16056-16180-16304-16332-16452-16598-16730-16810-16994-17210-17268-17786-17962-18214-18444-18446-18724-18912-18952-19164 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7798-7920-7952-7960-7988-8232-8256-8
      390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-11964-12054-12096-12126-12136-12202-12246-12298-12616-12774-12782-12790-12802-12976-13216-13246-13502-13766-14454-14974-15004-15124-15252-15294-15356-15530-15610-16316-16936-17024-17122-17214-17310-17528-17682-17742-17870-17878-18010-18410-18524-18788-19204-19254-19518-19596-19786-19874-19904 sum: 631944 positions: 562094,3058,0,4643,292
        Stripe: offset: 1842847 data: 188033 rows: 1000 tail: 67 index: 841
          Stream: column 0 section ROW_INDEX start: 1842847 length 12
          Stream: column 1 section ROW_INDEX start: 1842859 length 38
    @@ -176,12 +176,12 @@ Stripes:
          Encoding column 1: DIRECT_V2
          Encoding column 2: DIRECT_V2
          Encoding column 3: DIRECT_V2
    - Row group index column 1:
    - Entry 0:count: 1000 hasNull: false min: -2143595397 max: 2136858458 sum: -22999664100 positions: 0,0,0
    - Row group index column 2:
    - Entry 0:count: 1000 hasNull: false min: -9212379634781416464 max: 9197412874152820822 positions: 0,0,0
    - Row group index column 3:
    - Entry 0:count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022-12178-12418-12832-13304-13448-13590-13618-13908-14188-14246-14340-14364-14394-14762-14850-14964-15048-15494-15674-15726-16006-16056-16180-16304-16332-16452-16598-16730-16810-16994-17210-17268-17786-17962-18214-18444-18446-18724-18912-18952-19164-19348-19400-19546-19776-19896-20084 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-7
      798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-11964-12054-12096-12126-12136-12202-12246-12298-12616-12774-12782-12790-12802-12976-13216-13246-13502-13766-14454-14974-15004-15124-15252-15294-15356-15530-15610-16316-16936-17024-17122-17214-17310-17528-17682-17742-17870-17878-18010-18410-18524-18788-19204-19254-19518-19596-19786-19874-19904-20390-20752-20936 sum: 670762 positions: 0,0,0,0,0
    + Row group indices for column 1:
    + Entry 0: count: 1000 hasNull: false min: -2143595397 max: 2136858458 sum: -22999664100 positions: 0,0,0
    + Row group indices for column 2:
    + Entry 0: count: 1000 hasNull: false min: -9212379634781416464 max: 9197412874152820822 positions: 0,0,0
    + Row group indices for column 3:
    + Entry 0: count: 1000 hasNull: false min: Darkness,-230-368-488-586-862-930-1686-2044-2636-2652-2872-3108-3162-3192-3404-3442-3508-3542-3550-3712-3980-4146-4204-4336-4390-4418-4424-4490-4512-4650-4768-4924-4950-5210-5524-5630-5678-5710-5758-5952-6238-6252-6300-6366-6668-6712-6926-6942-7100-7194-7802-8030-8452-8608-8640-8862-8868-9134-9234-9412-9602-9608-9642-9678-9740-9780-10426-10510-10514-10706-10814-10870-10942-11028-11244-11326-11462-11496-11656-11830-12022-12178-12418-12832-13304-13448-13590-13618-13908-14188-14246-14340-14364-14394-14762-14850-14964-15048-15494-15674-15726-16006-16056-16180-16304-16332-16452-16598-16730-16810-16994-17210-17268-17786-17962-18214-18444-18446-18724-18912-18952-19164-19348-19400-19546-19776-19896-20084 max: worst-54-290-346-648-908-996-1038-1080-1560-1584-1620-1744-1770-1798-1852-1966-2162-2244-2286-2296-2534-2660-3114-3676-3788-4068-4150-4706-4744-5350-5420-5582-5696-5726-6006-6020-6024-6098-6184-6568-6636-6802-6994-7004-7318-7498-7758-7780-
      7798-7920-7952-7960-7988-8232-8256-8390-8416-8478-8620-8840-8984-9038-9128-9236-9248-9344-9594-9650-9714-9928-9938-10178-10368-10414-10502-10732-10876-11008-11158-11410-11722-11836-11964-12054-12096-12126-12136-12202-12246-12298-12616-12774-12782-12790-12802-12976-13216-13246-13502-13766-14454-14974-15004-15124-15252-15294-15356-15530-15610-16316-16936-17024-17122-17214-17310-17528-17682-17742-17870-17878-18010-18410-18524-18788-19204-19254-19518-19596-19786-19874-19904-20390-20752-20936 sum: 670762 positions: 0,0,0,0,0

      File length: 2033643 bytes
      Padding length: 0 bytes
  • Prasanthj at Feb 21, 2015 at 2:37 am
    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/mapjoin_mapjoin.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/mapjoin_mapjoin.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/mapjoin_mapjoin.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/mapjoin_mapjoin.q.out Sat Feb 21 02:36:54 2015
    @@ -264,7 +264,7 @@ STAGE PLANS:
                                          columns _col0
                                          columns.types string
                                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                          serialization.format 1
                                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/optimize_nullscan.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/optimize_nullscan.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/optimize_nullscan.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/optimize_nullscan.q.out Sat Feb 21 02:36:54 2015
    @@ -134,7 +134,7 @@ STAGE PLANS:
                                columns _col0
                                columns.types bigint
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -535,7 +535,7 @@ STAGE PLANS:
                              columns _col0,_col1
                              columns.types string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -595,7 +595,7 @@ ABSTRACT SYNTAX TREE:
      TOK_QUERY
         TOK_FROM
            TOK_SUBQUERY
    - TOK_UNION
    + TOK_UNIONALL
                  TOK_QUERY
                     TOK_FROM
                        TOK_TABREF
    @@ -952,7 +952,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -979,7 +979,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1383,7 +1383,7 @@ STAGE PLANS:
                              columns _col0,_col1
                              columns.types string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1443,7 +1443,7 @@ ABSTRACT SYNTAX TREE:
      TOK_QUERY
         TOK_FROM
            TOK_SUBQUERY
    - TOK_UNION
    + TOK_UNIONALL
                  TOK_QUERY
                     TOK_FROM
                        TOK_TABREF
    @@ -1532,7 +1532,7 @@ STAGE PLANS:
                                    columns _col0
                                    columns.types string
                                    escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                    serialization.format 1
                                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1674,7 +1674,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1910,7 +1910,7 @@ STAGE PLANS:
                                columns _col0,_col1,_col2,_col3
                                columns.types string:string:string:string
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/outer_join_ppr.q.java1.7.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/outer_join_ppr.q.java1.7.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/outer_join_ppr.q.java1.7.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/outer_join_ppr.q.java1.7.out Sat Feb 21 02:36:54 2015
    @@ -424,7 +424,7 @@ STAGE PLANS:
                                  columns _col0,_col1,_col2,_col3
                                  columns.types string:string:string:string
                                  escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                  serialization.format 1
                                  serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -895,7 +895,7 @@ STAGE PLANS:
                                  columns _col0,_col1,_col2,_col3
                                  columns.types string:string:string:string
                                  escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                  serialization.format 1
                                  serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/outer_join_ppr.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/outer_join_ppr.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/outer_join_ppr.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/outer_join_ppr.q.out Sat Feb 21 02:36:54 2015
    @@ -422,7 +422,7 @@ STAGE PLANS:
                                  columns _col0,_col1,_col2,_col3
                                  columns.types string:string:string:string
                                  escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                  serialization.format 1
                                  serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -893,7 +893,7 @@ STAGE PLANS:
                                  columns _col0,_col1,_col2,_col3
                                  columns.types string:string:string:string
                                  escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                  serialization.format 1
                                  serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/parallel_join0.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/parallel_join0.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/parallel_join0.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/parallel_join0.q.out Sat Feb 21 02:36:54 2015
    @@ -1,5 +1,5 @@
      Warning: Shuffle Join JOIN[8][tables = [src1, src2]] in Work 'Reducer 2' is a cross product
    -PREHOOK: query: -- SORT_BEFORE_DIFF
    +PREHOOK: query: -- SORT_QUERY_RESULTS

      EXPLAIN
      SELECT src1.key as k1, src1.value as v1,
    @@ -9,7 +9,7 @@ SELECT src1.key as k1, src1.value as v1,
        (SELECT * FROM src WHERE src.key < 10) src2
        SORT BY k1, v1, k2, v2
      PREHOOK: type: QUERY
    -POSTHOOK: query: -- SORT_BEFORE_DIFF
    +POSTHOOK: query: -- SORT_QUERY_RESULTS

      EXPLAIN
      SELECT src1.key as k1, src1.value as v1,

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/pcr.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/pcr.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/pcr.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/pcr.q.out Sat Feb 21 02:36:54 2015
    @@ -241,7 +241,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2
                              columns.types int:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -510,7 +510,7 @@ STAGE PLANS:
                              columns _col0,_col1
                              columns.types int:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -781,7 +781,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2
                              columns.types int:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1022,7 +1022,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2
                              columns.types int:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1311,7 +1311,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2
                              columns.types int:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1610,7 +1610,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2
                              columns.types int:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1862,7 +1862,7 @@ STAGE PLANS:
                              columns _col0,_col1
                              columns.types int:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -2073,7 +2073,7 @@ STAGE PLANS:
                              columns _col0,_col1
                              columns.types int:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -2370,7 +2370,7 @@ STAGE PLANS:
                              columns _col0,_col1
                              columns.types int:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -2663,7 +2663,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2
                              columns.types int:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -2936,7 +2936,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2,_col3,_col4,_col5
                              columns.types int:string:string:int:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -3244,7 +3244,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2,_col3,_col4,_col5
                              columns.types int:string:string:int:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -3620,7 +3620,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2
                              columns.types int:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -3943,7 +3943,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2
                              columns.types int:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -4717,7 +4717,7 @@ STAGE PLANS:
                                columns _col0,_col1
                                columns.types string:string
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -4950,7 +4950,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2,_col3
                              columns.types string:string:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -5179,7 +5179,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2,_col3
                              columns.types string:string:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/ppd_join_filter.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/ppd_join_filter.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/ppd_join_filter.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/ppd_join_filter.q.out Sat Feb 21 02:36:54 2015
    @@ -324,7 +324,7 @@ STAGE PLANS:
                                columns _col0,_col1,_col2
                                columns.types string:double:double
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -698,7 +698,7 @@ STAGE PLANS:
                                columns _col0,_col1,_col2
                                columns.types string:double:double
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1072,7 +1072,7 @@ STAGE PLANS:
                                columns _col0,_col1,_col2
                                columns.types string:double:double
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1446,7 +1446,7 @@ STAGE PLANS:
                                columns _col0,_col1,_col2
                                columns.types string:double:double
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
  • Prasanthj at Feb 21, 2015 at 2:37 am
    Modified: hive/branches/llap/ql/src/test/results/clientpositive/tez/vector_left_outer_join.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/tez/vector_left_outer_join.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/tez/vector_left_outer_join.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/tez/vector_left_outer_join.q.out Sat Feb 21 02:36:54 2015
    @@ -32,11 +32,11 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: c
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: ctinyint (type: tinyint), cint (type: int)
                          outputColumnNames: _col0, _col1
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                          Map Join Operator
                            condition map:
                                 Left Outer Join0 to 1
    @@ -46,7 +46,7 @@ STAGE PLANS:
                            outputColumnNames: _col0
                            input vertices:
                              1 Map 3
    - Statistics: Num rows: 13516 Data size: 414960 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 13516 Data size: 2906160 Basic stats: COMPLETE Column stats: NONE
                            Map Join Operator
                              condition map:
                                   Left Outer Join0 to 1
    @@ -55,7 +55,7 @@ STAGE PLANS:
                                1 _col0 (type: tinyint)
                              input vertices:
                                1 Map 4
    - Statistics: Num rows: 14867 Data size: 456456 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 14867 Data size: 3196776 Basic stats: COMPLETE Column stats: NONE
                              Group By Operator
                                aggregations: count()
                                mode: hash
    @@ -70,31 +70,31 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: c
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: cint (type: int)
                          outputColumnNames: _col0
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                          Reduce Output Operator
                            key expressions: _col0 (type: int)
                            sort order: +
                            Map-reduce partition columns: _col0 (type: int)
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                  Execution mode: vectorized
              Map 4
                  Map Operator Tree:
                      TableScan
                        alias: c
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: ctinyint (type: tinyint)
                          outputColumnNames: _col0
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                          Reduce Output Operator
                            key expressions: _col0 (type: tinyint)
                            sort order: +
                            Map-reduce partition columns: _col0 (type: tinyint)
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                  Execution mode: vectorized
              Reducer 2
                  Reduce Operator Tree:

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/tez/vector_varchar_simple.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/tez/vector_varchar_simple.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/tez/vector_varchar_simple.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/tez/vector_varchar_simple.q.out Sat Feb 21 02:36:54 2015
    @@ -271,17 +271,17 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: cint (type: int)
                          outputColumnNames: _col0
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                          Limit
                            Number of rows: 10
    - Statistics: Num rows: 10 Data size: 300 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 10 Data size: 2150 Basic stats: COMPLETE Column stats: NONE
                            Reduce Output Operator
                              sort order:
    - Statistics: Num rows: 10 Data size: 300 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 10 Data size: 2150 Basic stats: COMPLETE Column stats: NONE
                              value expressions: _col0 (type: int)
                  Execution mode: vectorized
              Reducer 2
    @@ -289,17 +289,17 @@ STAGE PLANS:
                    Select Operator
                      expressions: VALUE._col0 (type: int)
                      outputColumnNames: _col0
    - Statistics: Num rows: 10 Data size: 300 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 10 Data size: 2150 Basic stats: COMPLETE Column stats: NONE
                      Limit
                        Number of rows: 10
    - Statistics: Num rows: 10 Data size: 300 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 10 Data size: 2150 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: CAST( _col0 AS varchar(25)) (type: varchar(25))
                          outputColumnNames: _col0
    - Statistics: Num rows: 10 Data size: 300 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 10 Data size: 2150 Basic stats: COMPLETE Column stats: NONE
                          File Output Operator
                            compressed: false
    - Statistics: Num rows: 10 Data size: 300 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 10 Data size: 2150 Basic stats: COMPLETE Column stats: NONE
                            table:
                                input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
                                output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_0.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_0.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_0.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_0.q.out Sat Feb 21 02:36:54 2015
    @@ -32,11 +32,11 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: ctinyint (type: tinyint)
                          outputColumnNames: ctinyint
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                          Group By Operator
                            aggregations: min(ctinyint), max(ctinyint), count(ctinyint), count()
                            mode: hash
    @@ -126,11 +126,11 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: ctinyint (type: tinyint)
                          outputColumnNames: _col0
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                          Group By Operator
                            aggregations: sum(_col0)
                            mode: hash
    @@ -229,11 +229,11 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: ctinyint (type: tinyint)
                          outputColumnNames: ctinyint
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                          Group By Operator
                            aggregations: avg(ctinyint), variance(ctinyint), var_pop(ctinyint), var_samp(ctinyint), std(ctinyint), stddev(ctinyint), stddev_pop(ctinyint), stddev_samp(ctinyint)
                            mode: hash
    @@ -338,11 +338,11 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: cbigint (type: bigint)
                          outputColumnNames: cbigint
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                          Group By Operator
                            aggregations: min(cbigint), max(cbigint), count(cbigint), count()
                            mode: hash
    @@ -432,11 +432,11 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: cbigint (type: bigint)
                          outputColumnNames: _col0
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                          Group By Operator
                            aggregations: sum(_col0)
                            mode: hash
    @@ -535,11 +535,11 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: cbigint (type: bigint)
                          outputColumnNames: cbigint
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                          Group By Operator
                            aggregations: avg(cbigint), variance(cbigint), var_pop(cbigint), var_samp(cbigint), std(cbigint), stddev(cbigint), stddev_pop(cbigint), stddev_samp(cbigint)
                            mode: hash
    @@ -644,11 +644,11 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: cfloat (type: float)
                          outputColumnNames: cfloat
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                          Group By Operator
                            aggregations: min(cfloat), max(cfloat), count(cfloat), count()
                            mode: hash
    @@ -738,11 +738,11 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: cfloat (type: float)
                          outputColumnNames: _col0
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                          Group By Operator
                            aggregations: sum(_col0)
                            mode: hash
    @@ -841,11 +841,11 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: cfloat (type: float)
                          outputColumnNames: cfloat
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                          Group By Operator
                            aggregations: avg(cfloat), variance(cfloat), var_pop(cfloat), var_samp(cfloat), std(cfloat), stddev(cfloat), stddev_pop(cfloat), stddev_samp(cfloat)
                            mode: hash
    @@ -988,14 +988,14 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Filter Operator
                          predicate: ((cstring2 like '%b%') or ((79.553 <> UDFToDouble(cint)) or (UDFToDouble(cbigint) < cdouble))) (type: boolean)
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                          Select Operator
                            expressions: cbigint (type: bigint), cfloat (type: float), ctinyint (type: tinyint)
                            outputColumnNames: _col0, _col1, _col2
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                            Group By Operator
                              aggregations: avg(_col0), stddev_pop(_col0), var_samp(_col0), count(), sum(_col1), min(_col2)
                              mode: hash

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_13.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_13.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_13.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_13.q.out Sat Feb 21 02:36:54 2015
    @@ -80,25 +80,25 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Filter Operator
                          predicate: (((cfloat < 3569.0) and ((10.175 >= cdouble) and (cboolean1 <> 1))) or ((UDFToDouble(ctimestamp1) > 11.0) and ((UDFToDouble(ctimestamp2) <> 12.0) and (UDFToDouble(ctinyint) < 9763215.5639)))) (type: boolean)
    - Statistics: Num rows: 2730 Data size: 83809 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 2730 Data size: 586959 Basic stats: COMPLETE Column stats: NONE
                          Select Operator
                            expressions: cboolean1 (type: boolean), ctinyint (type: tinyint), ctimestamp1 (type: timestamp), cfloat (type: float), cstring1 (type: string)
                            outputColumnNames: _col0, _col1, _col2, _col3, _col4
    - Statistics: Num rows: 2730 Data size: 83809 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 2730 Data size: 586959 Basic stats: COMPLETE Column stats: NONE
                            Group By Operator
                              aggregations: max(_col1), sum(_col3), stddev_pop(_col3), stddev_pop(_col1), max(_col3), min(_col1)
                              keys: _col0 (type: boolean), _col1 (type: tinyint), _col2 (type: timestamp), _col3 (type: float), _col4 (type: string)
                              mode: hash
                              outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10
    - Statistics: Num rows: 2730 Data size: 83809 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 2730 Data size: 586959 Basic stats: COMPLETE Column stats: NONE
                              Reduce Output Operator
                                key expressions: _col0 (type: boolean), _col1 (type: tinyint), _col2 (type: timestamp), _col3 (type: float), _col4 (type: string)
                                sort order: +++++
                                Map-reduce partition columns: _col0 (type: boolean), _col1 (type: tinyint), _col2 (type: timestamp), _col3 (type: float), _col4 (type: string)
    - Statistics: Num rows: 2730 Data size: 83809 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 2730 Data size: 586959 Basic stats: COMPLETE Column stats: NONE
                                value expressions: _col5 (type: tinyint), _col6 (type: double), _col7 (type: struct<count:bigint,sum:double,variance:double>), _col8 (type: struct<count:bigint,sum:double,variance:double>), _col9 (type: float), _col10 (type: tinyint)
                  Execution mode: vectorized
              Reducer 2
    @@ -108,28 +108,28 @@ STAGE PLANS:
                      keys: KEY._col0 (type: boolean), KEY._col1 (type: tinyint), KEY._col2 (type: timestamp), KEY._col3 (type: float), KEY._col4 (type: string)
                      mode: mergepartial
                      outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10
    - Statistics: Num rows: 1365 Data size: 41904 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 1365 Data size: 293479 Basic stats: COMPLETE Column stats: NONE
                      Select Operator
                        expressions: _col0 (type: boolean), _col1 (type: tinyint), (- _col6) (type: double), (79.553 * UDFToDouble(_col3)) (type: double), _col7 (type: double), _col8 (type: double), (UDFToDouble(((- _col1) + _col5)) - 10.175) (type: double), (- (- _col6)) (type: double), (-26.28 / (- (- _col6))) (type: double), _col9 (type: float), ((_col6 * UDFToDouble(((- _col1) + _col5))) / UDFToDouble(_col1)) (type: double), _col2 (type: timestamp), _col10 (type: tinyint), _col3 (type: float), _col4 (type: string), (- _col1) (type: tinyint), _col5 (type: tinyint), ((- _col1) + _col5) (type: tinyint), _col6 (type: double), (_col6 * UDFToDouble(((- _col1) + _col5))) (type: double)
                        outputColumnNames: _col0, _col1, _col10, _col11, _col12, _col14, _col15, _col16, _col17, _col18, _col19, _col2, _col20, _col3, _col4, _col5, _col6, _col7, _col8, _col9
    - Statistics: Num rows: 1365 Data size: 41904 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 1365 Data size: 293479 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col0 (type: boolean), _col1 (type: tinyint), _col2 (type: timestamp), _col3 (type: float), _col4 (type: string)
                          sort order: +++++
    - Statistics: Num rows: 1365 Data size: 41904 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 1365 Data size: 293479 Basic stats: COMPLETE Column stats: NONE
                          value expressions: _col5 (type: tinyint), _col6 (type: tinyint), _col7 (type: tinyint), _col8 (type: double), _col9 (type: double), _col10 (type: double), _col11 (type: double), _col12 (type: double), _col14 (type: double), _col15 (type: double), _col16 (type: double), _col17 (type: double), _col18 (type: float), _col19 (type: double), _col20 (type: tinyint)
              Reducer 3
                  Reduce Operator Tree:
                    Select Operator
                      expressions: KEY.reducesinkkey0 (type: boolean), KEY.reducesinkkey1 (type: tinyint), KEY.reducesinkkey2 (type: timestamp), KEY.reducesinkkey3 (type: float), KEY.reducesinkkey4 (type: string), VALUE._col0 (type: tinyint), VALUE._col1 (type: tinyint), VALUE._col2 (type: tinyint), VALUE._col3 (type: double), VALUE._col4 (type: double), VALUE._col5 (type: double), VALUE._col6 (type: double), VALUE._col7 (type: double), VALUE._col5 (type: double), VALUE._col8 (type: double), VALUE._col9 (type: double), VALUE._col10 (type: double), VALUE._col11 (type: double), VALUE._col12 (type: float), VALUE._col13 (type: double), VALUE._col14 (type: tinyint)
                      outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20
    - Statistics: Num rows: 1365 Data size: 41904 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 1365 Data size: 293479 Basic stats: COMPLETE Column stats: NONE
                      Limit
                        Number of rows: 40
    - Statistics: Num rows: 40 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 40 Data size: 8600 Basic stats: COMPLETE Column stats: NONE
                        File Output Operator
                          compressed: false
    - Statistics: Num rows: 40 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 40 Data size: 8600 Basic stats: COMPLETE Column stats: NONE
                          table:
                              input format: org.apache.hadoop.mapred.TextInputFormat
                              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
    @@ -334,25 +334,25 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Filter Operator
                          predicate: (((cfloat < 3569.0) and ((10.175 >= cdouble) and (cboolean1 <> 1))) or ((UDFToDouble(ctimestamp1) > -1.388) and ((UDFToDouble(ctimestamp2) <> -1.3359999999999999) and (UDFToDouble(ctinyint) < 9763215.5639)))) (type: boolean)
    - Statistics: Num rows: 2730 Data size: 83809 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 2730 Data size: 586959 Basic stats: COMPLETE Column stats: NONE
                          Select Operator
                            expressions: cboolean1 (type: boolean), ctinyint (type: tinyint), ctimestamp1 (type: timestamp), cfloat (type: float), cstring1 (type: string)
                            outputColumnNames: _col0, _col1, _col2, _col3, _col4
    - Statistics: Num rows: 2730 Data size: 83809 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 2730 Data size: 586959 Basic stats: COMPLETE Column stats: NONE
                            Group By Operator
                              aggregations: max(_col1), sum(_col3), stddev_pop(_col3), stddev_pop(_col1), max(_col3), min(_col1)
                              keys: _col0 (type: boolean), _col1 (type: tinyint), _col2 (type: timestamp), _col3 (type: float), _col4 (type: string)
                              mode: hash
                              outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10
    - Statistics: Num rows: 2730 Data size: 83809 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 2730 Data size: 586959 Basic stats: COMPLETE Column stats: NONE
                              Reduce Output Operator
                                key expressions: _col0 (type: boolean), _col1 (type: tinyint), _col2 (type: timestamp), _col3 (type: float), _col4 (type: string)
                                sort order: +++++
                                Map-reduce partition columns: _col0 (type: boolean), _col1 (type: tinyint), _col2 (type: timestamp), _col3 (type: float), _col4 (type: string)
    - Statistics: Num rows: 2730 Data size: 83809 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 2730 Data size: 586959 Basic stats: COMPLETE Column stats: NONE
                                value expressions: _col5 (type: tinyint), _col6 (type: double), _col7 (type: struct<count:bigint,sum:double,variance:double>), _col8 (type: struct<count:bigint,sum:double,variance:double>), _col9 (type: float), _col10 (type: tinyint)
                  Execution mode: vectorized
              Reducer 2
    @@ -362,28 +362,28 @@ STAGE PLANS:
                      keys: KEY._col0 (type: boolean), KEY._col1 (type: tinyint), KEY._col2 (type: timestamp), KEY._col3 (type: float), KEY._col4 (type: string)
                      mode: mergepartial
                      outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10
    - Statistics: Num rows: 1365 Data size: 41904 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 1365 Data size: 293479 Basic stats: COMPLETE Column stats: NONE
                      Select Operator
                        expressions: _col0 (type: boolean), _col1 (type: tinyint), (- _col6) (type: double), (79.553 * UDFToDouble(_col3)) (type: double), _col7 (type: double), _col8 (type: double), (UDFToDouble(((- _col1) + _col5)) - 10.175) (type: double), (- (- _col6)) (type: double), (-26.28 / (- (- _col6))) (type: double), _col9 (type: float), ((_col6 * UDFToDouble(((- _col1) + _col5))) / UDFToDouble(_col1)) (type: double), _col2 (type: timestamp), _col10 (type: tinyint), _col3 (type: float), _col4 (type: string), (- _col1) (type: tinyint), _col5 (type: tinyint), ((- _col1) + _col5) (type: tinyint), _col6 (type: double), (_col6 * UDFToDouble(((- _col1) + _col5))) (type: double)
                        outputColumnNames: _col0, _col1, _col10, _col11, _col12, _col14, _col15, _col16, _col17, _col18, _col19, _col2, _col20, _col3, _col4, _col5, _col6, _col7, _col8, _col9
    - Statistics: Num rows: 1365 Data size: 41904 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 1365 Data size: 293479 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col0 (type: boolean), _col1 (type: tinyint), _col2 (type: timestamp), _col3 (type: float), _col4 (type: string)
                          sort order: +++++
    - Statistics: Num rows: 1365 Data size: 41904 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 1365 Data size: 293479 Basic stats: COMPLETE Column stats: NONE
                          value expressions: _col5 (type: tinyint), _col6 (type: tinyint), _col7 (type: tinyint), _col8 (type: double), _col9 (type: double), _col10 (type: double), _col11 (type: double), _col12 (type: double), _col14 (type: double), _col15 (type: double), _col16 (type: double), _col17 (type: double), _col18 (type: float), _col19 (type: double), _col20 (type: tinyint)
              Reducer 3
                  Reduce Operator Tree:
                    Select Operator
                      expressions: KEY.reducesinkkey0 (type: boolean), KEY.reducesinkkey1 (type: tinyint), KEY.reducesinkkey2 (type: timestamp), KEY.reducesinkkey3 (type: float), KEY.reducesinkkey4 (type: string), VALUE._col0 (type: tinyint), VALUE._col1 (type: tinyint), VALUE._col2 (type: tinyint), VALUE._col3 (type: double), VALUE._col4 (type: double), VALUE._col5 (type: double), VALUE._col6 (type: double), VALUE._col7 (type: double), VALUE._col5 (type: double), VALUE._col8 (type: double), VALUE._col9 (type: double), VALUE._col10 (type: double), VALUE._col11 (type: double), VALUE._col12 (type: float), VALUE._col13 (type: double), VALUE._col14 (type: tinyint)
                      outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20
    - Statistics: Num rows: 1365 Data size: 41904 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 1365 Data size: 293479 Basic stats: COMPLETE Column stats: NONE
                      Limit
                        Number of rows: 40
    - Statistics: Num rows: 40 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 40 Data size: 8600 Basic stats: COMPLETE Column stats: NONE
                        File Output Operator
                          compressed: false
    - Statistics: Num rows: 40 Data size: 1200 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 40 Data size: 8600 Basic stats: COMPLETE Column stats: NONE
                          table:
                              input format: org.apache.hadoop.mapred.TextInputFormat
                              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_14.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_14.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_14.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_14.q.out Sat Feb 21 02:36:54 2015
    @@ -80,25 +80,25 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Filter Operator
                          predicate: ((((UDFToLong(ctinyint) <= cbigint) and ((UDFToDouble(cint) <= cdouble) or (ctimestamp2 < ctimestamp1))) and (cdouble < UDFToDouble(ctinyint))) and ((cbigint > -257) or (cfloat < UDFToFloat(cint)))) (type: boolean)
    - Statistics: Num rows: 606 Data size: 18603 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 606 Data size: 130292 Basic stats: COMPLETE Column stats: NONE
                          Select Operator
                            expressions: ctimestamp1 (type: timestamp), cfloat (type: float), cstring1 (type: string), cboolean1 (type: boolean), cdouble (type: double), (- (-26.28 + cdouble)) (type: double)
                            outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
    - Statistics: Num rows: 606 Data size: 18603 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 606 Data size: 130292 Basic stats: COMPLETE Column stats: NONE
                            Group By Operator
                              aggregations: stddev_samp(_col5), max(_col1), stddev_pop(_col1), count(_col1), var_pop(_col1), var_samp(_col1)
                              keys: _col0 (type: timestamp), _col1 (type: float), _col2 (type: string), _col3 (type: boolean), _col4 (type: double)
                              mode: hash
                              outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10
    - Statistics: Num rows: 606 Data size: 18603 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 606 Data size: 130292 Basic stats: COMPLETE Column stats: NONE
                              Reduce Output Operator
                                key expressions: _col0 (type: timestamp), _col1 (type: float), _col2 (type: string), _col3 (type: boolean), _col4 (type: double)
                                sort order: +++++
                                Map-reduce partition columns: _col0 (type: timestamp), _col1 (type: float), _col2 (type: string), _col3 (type: boolean), _col4 (type: double)
    - Statistics: Num rows: 606 Data size: 18603 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 606 Data size: 130292 Basic stats: COMPLETE Column stats: NONE
                                value expressions: _col5 (type: struct<count:bigint,sum:double,variance:double>), _col6 (type: float), _col7 (type: struct<count:bigint,sum:double,variance:double>), _col8 (type: bigint), _col9 (type: struct<count:bigint,sum:double,variance:double>), _col10 (type: struct<count:bigint,sum:double,variance:double>)
                  Execution mode: vectorized
              Reducer 2
    @@ -108,25 +108,25 @@ STAGE PLANS:
                      keys: KEY._col0 (type: timestamp), KEY._col1 (type: float), KEY._col2 (type: string), KEY._col3 (type: boolean), KEY._col4 (type: double)
                      mode: mergepartial
                      outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10
    - Statistics: Num rows: 303 Data size: 9301 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 303 Data size: 65146 Basic stats: COMPLETE Column stats: NONE
                      Select Operator
                        expressions: _col0 (type: timestamp), _col1 (type: float), _col2 (type: string), _col3 (type: boolean), _col4 (type: double), (-26.28 + _col4) (type: double), (- (-26.28 + _col4)) (type: double), _col5 (type: double), (UDFToDouble(_col1) * -26.28) (type: double), _col6 (type: float), (- _col1) (type: float), (- _col6) (type: float), ((- (-26.28 + _col4)) / 10.175) (type: double), _col7 (type: double), _col8 (type: bigint), (- ((- (-26.28 + _col4)) / 10.175)) (type: double), (-1.389 % _col5) (type: double), (UDFToDouble(_col1) - _col4) (type: double), _col9 (type: double), (_col9 % 10.175) (type: double), _col10 (type: double), (- (UDFToDouble(_col1) - _col4)) (type: double)
                        outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21
    - Statistics: Num rows: 303 Data size: 9301 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 303 Data size: 65146 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col2 (type: string), _col1 (type: float), _col4 (type: double), _col0 (type: timestamp)
                          sort order: ++++
    - Statistics: Num rows: 303 Data size: 9301 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 303 Data size: 65146 Basic stats: COMPLETE Column stats: NONE
                          value expressions: _col3 (type: boolean), _col5 (type: double), _col6 (type: double), _col7 (type: double), _col8 (type: double), _col9 (type: float), _col10 (type: float), _col11 (type: float), _col12 (type: double), _col13 (type: double), _col14 (type: bigint), _col15 (type: double), _col16 (type: double), _col17 (type: double), _col18 (type: double), _col19 (type: double), _col20 (type: double), _col21 (type: double)
              Reducer 3
                  Reduce Operator Tree:
                    Select Operator
                      expressions: KEY.reducesinkkey3 (type: timestamp), KEY.reducesinkkey1 (type: float), KEY.reducesinkkey0 (type: string), VALUE._col0 (type: boolean), KEY.reducesinkkey2 (type: double), VALUE._col1 (type: double), VALUE._col2 (type: double), VALUE._col3 (type: double), VALUE._col4 (type: double), VALUE._col5 (type: float), VALUE._col6 (type: float), VALUE._col7 (type: float), VALUE._col8 (type: double), VALUE._col9 (type: double), VALUE._col10 (type: bigint), VALUE._col11 (type: double), VALUE._col12 (type: double), VALUE._col13 (type: double), VALUE._col14 (type: double), VALUE._col15 (type: double), VALUE._col16 (type: double), VALUE._col17 (type: double)
                      outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21
    - Statistics: Num rows: 303 Data size: 9301 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 303 Data size: 65146 Basic stats: COMPLETE Column stats: NONE
                      File Output Operator
                        compressed: false
    - Statistics: Num rows: 303 Data size: 9301 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 303 Data size: 65146 Basic stats: COMPLETE Column stats: NONE
                        table:
                            input format: org.apache.hadoop.mapred.TextInputFormat
                            output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_15.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_15.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_15.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_15.q.out Sat Feb 21 02:36:54 2015
    @@ -76,25 +76,25 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Filter Operator
                          predicate: ((cstring2 like '%ss%') or ((cstring1 like '10%') or ((cint >= -75) and ((UDFToShort(ctinyint) = csmallint) and (cdouble >= -3728.0))))) (type: boolean)
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                          Select Operator
                            expressions: cfloat (type: float), cboolean1 (type: boolean), cdouble (type: double), cstring1 (type: string), ctinyint (type: tinyint), cint (type: int), ctimestamp1 (type: timestamp)
                            outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                            Group By Operator
                              aggregations: stddev_samp(_col0), min(_col2), stddev_samp(_col4), var_pop(_col4), var_samp(_col5), stddev_pop(_col5)
                              keys: _col0 (type: float), _col1 (type: boolean), _col2 (type: double), _col3 (type: string), _col4 (type: tinyint), _col5 (type: int), _col6 (type: timestamp)
                              mode: hash
                              outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                              Reduce Output Operator
                                key expressions: _col0 (type: float), _col1 (type: boolean), _col2 (type: double), _col3 (type: string), _col4 (type: tinyint), _col5 (type: int), _col6 (type: timestamp)
                                sort order: +++++++
                                Map-reduce partition columns: _col0 (type: float), _col1 (type: boolean), _col2 (type: double), _col3 (type: string), _col4 (type: tinyint), _col5 (type: int), _col6 (type: timestamp)
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                                value expressions: _col7 (type: struct<count:bigint,sum:double,variance:double>), _col8 (type: double), _col9 (type: struct<count:bigint,sum:double,variance:double>), _col10 (type: struct<count:bigint,sum:double,variance:double>), _col11 (type: struct<count:bigint,sum:double,variance:double>), _col12 (type: struct<count:bigint,sum:double,variance:double>)
                  Execution mode: vectorized
              Reducer 2
    @@ -104,25 +104,25 @@ STAGE PLANS:
                      keys: KEY._col0 (type: float), KEY._col1 (type: boolean), KEY._col2 (type: double), KEY._col3 (type: string), KEY._col4 (type: tinyint), KEY._col5 (type: int), KEY._col6 (type: timestamp)
                      mode: mergepartial
                      outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12
    - Statistics: Num rows: 6144 Data size: 188618 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 6144 Data size: 1320982 Basic stats: COMPLETE Column stats: NONE
                      Select Operator
                        expressions: _col0 (type: float), _col1 (type: boolean), _col2 (type: double), _col3 (type: string), _col4 (type: tinyint), _col5 (type: int), _col6 (type: timestamp), _col7 (type: double), (-26.28 - UDFToDouble(_col5)) (type: double), _col8 (type: double), (_col2 * 79.553) (type: double), (33.0 % _col0) (type: float), _col9 (type: double), _col10 (type: double), (-23.0 % _col2) (type: double), (- _col4) (type: tinyint), _col11 (type: double), (UDFToFloat(_col5) - _col0) (type: float), (-23 % UDFToInteger(_col4)) (type: int), (- (-26.28 - UDFToDouble(_col5))) (type: double), _col12 (type: double)
                        outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20
    - Statistics: Num rows: 6144 Data size: 188618 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 6144 Data size: 1320982 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col0 (type: float), _col1 (type: boolean), _col2 (type: double), _col3 (type: string), _col4 (type: tinyint), _col5 (type: int), _col6 (type: timestamp)
                          sort order: +++++++
    - Statistics: Num rows: 6144 Data size: 188618 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 6144 Data size: 1320982 Basic stats: COMPLETE Column stats: NONE
                          value expressions: _col7 (type: double), _col8 (type: double), _col9 (type: double), _col10 (type: double), _col11 (type: float), _col12 (type: double), _col13 (type: double), _col14 (type: double), _col15 (type: tinyint), _col16 (type: double), _col17 (type: float), _col18 (type: int), _col19 (type: double), _col20 (type: double)
              Reducer 3
                  Reduce Operator Tree:
                    Select Operator
                      expressions: KEY.reducesinkkey0 (type: float), KEY.reducesinkkey1 (type: boolean), KEY.reducesinkkey2 (type: double), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: tinyint), KEY.reducesinkkey5 (type: int), KEY.reducesinkkey6 (type: timestamp), VALUE._col0 (type: double), VALUE._col1 (type: double), VALUE._col2 (type: double), VALUE._col3 (type: double), VALUE._col4 (type: float), VALUE._col5 (type: double), VALUE._col6 (type: double), VALUE._col7 (type: double), VALUE._col8 (type: tinyint), VALUE._col9 (type: double), VALUE._col10 (type: float), VALUE._col11 (type: int), VALUE._col12 (type: double), VALUE._col13 (type: double)
                      outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20
    - Statistics: Num rows: 6144 Data size: 188618 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 6144 Data size: 1320982 Basic stats: COMPLETE Column stats: NONE
                      File Output Operator
                        compressed: false
    - Statistics: Num rows: 6144 Data size: 188618 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 6144 Data size: 1320982 Basic stats: COMPLETE Column stats: NONE
                        table:
                            input format: org.apache.hadoop.mapred.TextInputFormat
                            output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_16.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_16.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_16.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_16.q.out Sat Feb 21 02:36:54 2015
    @@ -57,25 +57,25 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Filter Operator
                          predicate: ((cstring2 like '%b%') and ((cdouble >= -1.389) or (cstring1 < 'a'))) (type: boolean)
    - Statistics: Num rows: 4096 Data size: 125745 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 4096 Data size: 880654 Basic stats: COMPLETE Column stats: NONE
                          Select Operator
                            expressions: cstring1 (type: string), cdouble (type: double), ctimestamp1 (type: timestamp)
                            outputColumnNames: _col0, _col1, _col2
    - Statistics: Num rows: 4096 Data size: 125745 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 4096 Data size: 880654 Basic stats: COMPLETE Column stats: NONE
                            Group By Operator
                              aggregations: count(_col1), stddev_samp(_col1), min(_col1)
                              keys: _col0 (type: string), _col1 (type: double), _col2 (type: timestamp)
                              mode: hash
                              outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
    - Statistics: Num rows: 4096 Data size: 125745 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 4096 Data size: 880654 Basic stats: COMPLETE Column stats: NONE
                              Reduce Output Operator
                                key expressions: _col0 (type: string), _col1 (type: double), _col2 (type: timestamp)
                                sort order: +++
                                Map-reduce partition columns: _col0 (type: string), _col1 (type: double), _col2 (type: timestamp)
    - Statistics: Num rows: 4096 Data size: 125745 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 4096 Data size: 880654 Basic stats: COMPLETE Column stats: NONE
                                value expressions: _col3 (type: bigint), _col4 (type: struct<count:bigint,sum:double,variance:double>), _col5 (type: double)
                  Execution mode: vectorized
              Reducer 2
    @@ -85,14 +85,14 @@ STAGE PLANS:
                      keys: KEY._col0 (type: string), KEY._col1 (type: double), KEY._col2 (type: timestamp)
                      mode: mergepartial
                      outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
    - Statistics: Num rows: 2048 Data size: 62872 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 2048 Data size: 440327 Basic stats: COMPLETE Column stats: NONE
                      Select Operator
                        expressions: _col0 (type: string), _col1 (type: double), _col2 (type: timestamp), (_col1 - 9763215.5639) (type: double), (- (_col1 - 9763215.5639)) (type: double), _col3 (type: bigint), _col4 (type: double), (- _col4) (type: double), (_col4 * UDFToDouble(_col3)) (type: double), _col5 (type: double), (9763215.5639 / _col1) (type: double), (UDFToDouble(_col3) / -1.389) (type: double), _col4 (type: double)
                        outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12
    - Statistics: Num rows: 2048 Data size: 62872 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 2048 Data size: 440327 Basic stats: COMPLETE Column stats: NONE
                        File Output Operator
                          compressed: false
    - Statistics: Num rows: 2048 Data size: 62872 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 2048 Data size: 440327 Basic stats: COMPLETE Column stats: NONE
                          table:
                              input format: org.apache.hadoop.mapred.TextInputFormat
                              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_7.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_7.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_7.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_7.q.out Sat Feb 21 02:36:54 2015
    @@ -63,20 +63,20 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Filter Operator
                          predicate: (((ctinyint <> 0) and ((UDFToDouble(ctimestamp1) <= 0.0) or ((UDFToInteger(ctinyint) = cint) or (cstring2 like 'ss')))) and ((988888.0 < cdouble) or ((UDFToDouble(ctimestamp2) > -15.0) and (3569.0 >= cdouble)))) (type: boolean)
    - Statistics: Num rows: 7281 Data size: 223523 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 7281 Data size: 1565441 Basic stats: COMPLETE Column stats: NONE
                          Select Operator
                            expressions: cboolean1 (type: boolean), cbigint (type: bigint), csmallint (type: smallint), ctinyint (type: tinyint), ctimestamp1 (type: timestamp), cstring1 (type: string), (cbigint + cbigint) (type: bigint), (UDFToInteger(csmallint) % -257) (type: int), (- csmallint) (type: smallint), (- ctinyint) (type: tinyint), (UDFToInteger((- ctinyint)) + 17) (type: int), (cbigint * UDFToLong((- csmallint))) (type: bigint), (cint % UDFToInteger(csmallint)) (type: int), (- ctinyint) (type: tinyint), ((- ctinyint) % ctinyint) (type: tinyint)
                            outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14
    - Statistics: Num rows: 7281 Data size: 223523 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 7281 Data size: 1565441 Basic stats: COMPLETE Column stats: NONE
                            Limit
                              Number of rows: 25
    - Statistics: Num rows: 25 Data size: 750 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 25 Data size: 5375 Basic stats: COMPLETE Column stats: NONE
                              File Output Operator
                                compressed: false
    - Statistics: Num rows: 25 Data size: 750 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 25 Data size: 5375 Basic stats: COMPLETE Column stats: NONE
                                table:
                                    input format: org.apache.hadoop.mapred.TextInputFormat
                                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
    @@ -235,20 +235,20 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Filter Operator
                          predicate: (((ctinyint <> 0) and ((UDFToDouble(ctimestamp1) <= 0.0) or ((UDFToInteger(ctinyint) = cint) or (cstring2 like 'ss')))) and ((988888.0 < cdouble) or ((UDFToDouble(ctimestamp2) > 7.6850000000000005) and (3569.0 >= cdouble)))) (type: boolean)
    - Statistics: Num rows: 7281 Data size: 223523 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 7281 Data size: 1565441 Basic stats: COMPLETE Column stats: NONE
                          Select Operator
                            expressions: cboolean1 (type: boolean), cbigint (type: bigint), csmallint (type: smallint), ctinyint (type: tinyint), ctimestamp1 (type: timestamp), cstring1 (type: string), (cbigint + cbigint) (type: bigint), (UDFToInteger(csmallint) % -257) (type: int), (- csmallint) (type: smallint), (- ctinyint) (type: tinyint), (UDFToInteger((- ctinyint)) + 17) (type: int), (cbigint * UDFToLong((- csmallint))) (type: bigint), (cint % UDFToInteger(csmallint)) (type: int), (- ctinyint) (type: tinyint), ((- ctinyint) % ctinyint) (type: tinyint)
                            outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14
    - Statistics: Num rows: 7281 Data size: 223523 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 7281 Data size: 1565441 Basic stats: COMPLETE Column stats: NONE
                            Limit
                              Number of rows: 25
    - Statistics: Num rows: 25 Data size: 750 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 25 Data size: 5375 Basic stats: COMPLETE Column stats: NONE
                              File Output Operator
                                compressed: false
    - Statistics: Num rows: 25 Data size: 750 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 25 Data size: 5375 Basic stats: COMPLETE Column stats: NONE
                                table:
                                    input format: org.apache.hadoop.mapred.TextInputFormat
                                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_8.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_8.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_8.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_8.q.out Sat Feb 21 02:36:54 2015
    @@ -59,20 +59,20 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Filter Operator
                          predicate: ((cstring2 is not null and ((UDFToDouble(ctimestamp1) <= 10.0) and (UDFToDouble(ctimestamp2) <> 16.0))) or ((cfloat < -6432.0) or (cboolean1 is not null and (cdouble = 988888.0)))) (type: boolean)
    - Statistics: Num rows: 9216 Data size: 282927 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 9216 Data size: 1981473 Basic stats: COMPLETE Column stats: NONE
                          Select Operator
                            expressions: ctimestamp1 (type: timestamp), cdouble (type: double), cboolean1 (type: boolean), cstring1 (type: string), cfloat (type: float), (- cdouble) (type: double), (-5638.15 - cdouble) (type: double), (cdouble * -257.0) (type: double), (UDFToFloat(cint) + cfloat) (type: float), ((- cdouble) + UDFToDouble(cbigint)) (type: double), (- cdouble) (type: double), (-1.389 - UDFToDouble(cfloat)) (type: double), (- cfloat) (type: float), ((-5638.15 - cdouble) + UDFToDouble((UDFToFloat(cint) + cfloat))) (type: double)
                            outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13
    - Statistics: Num rows: 9216 Data size: 282927 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 9216 Data size: 1981473 Basic stats: COMPLETE Column stats: NONE
                            Limit
                              Number of rows: 20
    - Statistics: Num rows: 20 Data size: 600 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 20 Data size: 4300 Basic stats: COMPLETE Column stats: NONE
                              File Output Operator
                                compressed: false
    - Statistics: Num rows: 20 Data size: 600 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 20 Data size: 4300 Basic stats: COMPLETE Column stats: NONE
                                table:
                                    input format: org.apache.hadoop.mapred.TextInputFormat
                                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
    @@ -218,20 +218,20 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Filter Operator
                          predicate: ((cstring2 is not null and ((UDFToDouble(ctimestamp1) <= 12.503) and (UDFToDouble(ctimestamp2) <> 11.998))) or ((cfloat < -6432.0) or (cboolean1 is not null and (cdouble = 988888.0)))) (type: boolean)
    - Statistics: Num rows: 9216 Data size: 282927 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 9216 Data size: 1981473 Basic stats: COMPLETE Column stats: NONE
                          Select Operator
                            expressions: ctimestamp1 (type: timestamp), cdouble (type: double), cboolean1 (type: boolean), cstring1 (type: string), cfloat (type: float), (- cdouble) (type: double), (-5638.15 - cdouble) (type: double), (cdouble * -257.0) (type: double), (UDFToFloat(cint) + cfloat) (type: float), ((- cdouble) + UDFToDouble(cbigint)) (type: double), (- cdouble) (type: double), (-1.389 - UDFToDouble(cfloat)) (type: double), (- cfloat) (type: float), ((-5638.15 - cdouble) + UDFToDouble((UDFToFloat(cint) + cfloat))) (type: double)
                            outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13
    - Statistics: Num rows: 9216 Data size: 282927 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 9216 Data size: 1981473 Basic stats: COMPLETE Column stats: NONE
                            Limit
                              Number of rows: 20
    - Statistics: Num rows: 20 Data size: 600 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 20 Data size: 4300 Basic stats: COMPLETE Column stats: NONE
                              File Output Operator
                                compressed: false
    - Statistics: Num rows: 20 Data size: 600 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 20 Data size: 4300 Basic stats: COMPLETE Column stats: NONE
                                table:
                                    input format: org.apache.hadoop.mapred.TextInputFormat
                                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_9.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_9.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_9.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_9.q.out Sat Feb 21 02:36:54 2015
    @@ -53,25 +53,25 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Filter Operator
                          predicate: ((cstring2 like '%b%') and ((cdouble >= -1.389) or (cstring1 < 'a'))) (type: boolean)
    - Statistics: Num rows: 4096 Data size: 125745 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 4096 Data size: 880654 Basic stats: COMPLETE Column stats: NONE
                          Select Operator
                            expressions: cstring1 (type: string), cdouble (type: double), ctimestamp1 (type: timestamp)
                            outputColumnNames: _col0, _col1, _col2
    - Statistics: Num rows: 4096 Data size: 125745 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 4096 Data size: 880654 Basic stats: COMPLETE Column stats: NONE
                            Group By Operator
                              aggregations: count(_col1), stddev_samp(_col1), min(_col1)
                              keys: _col0 (type: string), _col1 (type: double), _col2 (type: timestamp)
                              mode: hash
                              outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
    - Statistics: Num rows: 4096 Data size: 125745 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 4096 Data size: 880654 Basic stats: COMPLETE Column stats: NONE
                              Reduce Output Operator
                                key expressions: _col0 (type: string), _col1 (type: double), _col2 (type: timestamp)
                                sort order: +++
                                Map-reduce partition columns: _col0 (type: string), _col1 (type: double), _col2 (type: timestamp)
    - Statistics: Num rows: 4096 Data size: 125745 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 4096 Data size: 880654 Basic stats: COMPLETE Column stats: NONE
                                value expressions: _col3 (type: bigint), _col4 (type: struct<count:bigint,sum:double,variance:double>), _col5 (type: double)
                  Execution mode: vectorized
              Reducer 2
    @@ -81,14 +81,14 @@ STAGE PLANS:
                      keys: KEY._col0 (type: string), KEY._col1 (type: double), KEY._col2 (type: timestamp)
                      mode: mergepartial
                      outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
    - Statistics: Num rows: 2048 Data size: 62872 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 2048 Data size: 440327 Basic stats: COMPLETE Column stats: NONE
                      Select Operator
                        expressions: _col0 (type: string), _col1 (type: double), _col2 (type: timestamp), (_col1 - 9763215.5639) (type: double), (- (_col1 - 9763215.5639)) (type: double), _col3 (type: bigint), _col4 (type: double), (- _col4) (type: double), (_col4 * UDFToDouble(_col3)) (type: double), _col5 (type: double), (9763215.5639 / _col1) (type: double), (UDFToDouble(_col3) / -1.389) (type: double), _col4 (type: double)
                        outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12
    - Statistics: Num rows: 2048 Data size: 62872 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 2048 Data size: 440327 Basic stats: COMPLETE Column stats: NONE
                        File Output Operator
                          compressed: false
    - Statistics: Num rows: 2048 Data size: 62872 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 2048 Data size: 440327 Basic stats: COMPLETE Column stats: NONE
                          table:
                              input format: org.apache.hadoop.mapred.TextInputFormat
                              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_div0.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_div0.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_div0.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/tez/vectorization_div0.q.out Sat Feb 21 02:36:54 2015
    @@ -160,18 +160,18 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Filter Operator
                          predicate: ((cbigint > 0) and (cbigint < 100000000)) (type: boolean)
    - Statistics: Num rows: 1365 Data size: 41904 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 1365 Data size: 293479 Basic stats: COMPLETE Column stats: NONE
                          Select Operator
                            expressions: (cbigint - 988888) (type: bigint), (cdouble / UDFToDouble((cbigint - 988888))) (type: double), (1.2 / UDFToDouble((cbigint - 988888))) (type: double)
                            outputColumnNames: _col0, _col1, _col2
    - Statistics: Num rows: 1365 Data size: 41904 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 1365 Data size: 293479 Basic stats: COMPLETE Column stats: NONE
                            Reduce Output Operator
                              key expressions: _col0 (type: bigint), _col1 (type: double)
                              sort order: ++
    - Statistics: Num rows: 1365 Data size: 41904 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 1365 Data size: 293479 Basic stats: COMPLETE Column stats: NONE
                              value expressions: _col2 (type: double)
                  Execution mode: vectorized
              Reducer 2
    @@ -179,13 +179,13 @@ STAGE PLANS:
                    Select Operator
                      expressions: KEY.reducesinkkey0 (type: bigint), KEY.reducesinkkey1 (type: double), VALUE._col0 (type: double)
                      outputColumnNames: _col0, _col1, _col2
    - Statistics: Num rows: 1365 Data size: 41904 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 1365 Data size: 293479 Basic stats: COMPLETE Column stats: NONE
                      Limit
                        Number of rows: 100
    - Statistics: Num rows: 100 Data size: 3000 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 100 Data size: 21500 Basic stats: COMPLETE Column stats: NONE
                        File Output Operator
                          compressed: false
    - Statistics: Num rows: 100 Data size: 3000 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 100 Data size: 21500 Basic stats: COMPLETE Column stats: NONE
                          table:
                              input format: org.apache.hadoop.mapred.TextInputFormat
                              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
    @@ -335,18 +335,18 @@ STAGE PLANS:
                  Map Operator Tree:
                      TableScan
                        alias: alltypesorc
    - Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 12288 Data size: 2641964 Basic stats: COMPLETE Column stats: NONE
                        Filter Operator
                          predicate: ((cdouble >= -500.0) and (cdouble < -199.0)) (type: boolean)
    - Statistics: Num rows: 1365 Data size: 41904 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 1365 Data size: 293479 Basic stats: COMPLETE Column stats: NONE
                          Select Operator
                            expressions: (cdouble + 200.0) (type: double), (UDFToDouble(cbigint) / (cdouble + 200.0)) (type: double), ((cdouble + 200.0) / (cdouble + 200.0)) (type: double), (3.0 / (cdouble + 200.0)) (type: double), (1.2 / (cdouble + 200.0)) (type: double)
                            outputColumnNames: _col0, _col1, _col2, _col4, _col5
    - Statistics: Num rows: 1365 Data size: 41904 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 1365 Data size: 293479 Basic stats: COMPLETE Column stats: NONE
                            Reduce Output Operator
                              key expressions: _col0 (type: double), _col1 (type: double)
                              sort order: ++
    - Statistics: Num rows: 1365 Data size: 41904 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 1365 Data size: 293479 Basic stats: COMPLETE Column stats: NONE
                              value expressions: _col2 (type: double), _col4 (type: double), _col5 (type: double)
                  Execution mode: vectorized
              Reducer 2
    @@ -354,13 +354,13 @@ STAGE PLANS:
                    Select Operator
                      expressions: KEY.reducesinkkey0 (type: double), KEY.reducesinkkey1 (type: double), VALUE._col0 (type: double), KEY.reducesinkkey1 (type: double), VALUE._col1 (type: double), VALUE._col2 (type: double)
                      outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
    - Statistics: Num rows: 1365 Data size: 41904 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 1365 Data size: 293479 Basic stats: COMPLETE Column stats: NONE
                      Limit
                        Number of rows: 100
    - Statistics: Num rows: 100 Data size: 3000 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 100 Data size: 21500 Basic stats: COMPLETE Column stats: NONE
                        File Output Operator
                          compressed: false
    - Statistics: Num rows: 100 Data size: 3000 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 100 Data size: 21500 Basic stats: COMPLETE Column stats: NONE
                          table:
                              input format: org.apache.hadoop.mapred.TextInputFormat
                              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
  • Prasanthj at Feb 21, 2015 at 2:37 am
    Modified: hive/branches/llap/ql/src/test/results/clientpositive/tez/ptf.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/tez/ptf.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/tez/ptf.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/tez/ptf.q.out Sat Feb 21 02:36:54 2015
    @@ -54,6 +54,18 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5, _col7
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col1
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col2 (type: string), _col1 (type: string)
    @@ -68,6 +80,38 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5, _col7
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: _col2
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col1
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol1
    + arguments: _col1
    + name: dense_rank
    + window function: GenericUDAFDenseRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol2
    + arguments: _col7
    + name: sum
    + window function: GenericUDAFSumDouble
    + window frame: PRECEDING(MAX)~
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int), _wcol0 (type: int), _wcol1 (type: int), _wcol2 (type: double)
    @@ -214,6 +258,18 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: j
    + output shape: _col1: string, _col2: string, _col5: int
    + type: SUBQUERY
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col1
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col2 (type: string), _col1 (type: string)
    @@ -228,6 +284,25 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: _col2
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col5, 1, _col5
    + name: lag
    + window function: GenericUDAFLagEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
                        Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int), (_col5 - _wcol0) (type: int)
    @@ -330,10 +405,22 @@ STAGE PLANS:
              Reducer 2
                  Reduce Operator Tree:
                    Select Operator
    - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), VALUE._col3 (type: int)
    - outputColumnNames: _col2, _col1, _col5
    + expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int)
    + outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col1: string, _col2: string, _col5: int
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col1
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int)
    @@ -445,6 +532,18 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5, _col7
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: TABLE
    + Partition table definition
    + input alias: abc
    + name: noop
    + order by: _col1
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col2 (type: string), _col1 (type: string)
    @@ -459,6 +558,38 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5, _col7
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: _col2
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col1
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol1
    + arguments: _col1
    + name: dense_rank
    + window function: GenericUDAFDenseRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol2
    + arguments: _col7
    + name: sum
    + window function: GenericUDAFSumDouble
    + window frame: PRECEDING(MAX)~
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int), _wcol0 (type: int), _wcol1 (type: int), _wcol2 (type: double)
    @@ -578,6 +709,18 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col1: string, _col2: string, _col5: int
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col1
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col2 (type: string), _col1 (type: string)
    @@ -592,6 +735,39 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: _col2
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col1
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol1
    + arguments: _col1
    + name: dense_rank
    + window function: GenericUDAFDenseRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol2
    + arguments: _col5, 1, _col5
    + name: lag
    + window function: GenericUDAFLagEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int), _wcol0 (type: int), _wcol1 (type: int), _col5 (type: int), (_col5 - _wcol2) (type: int)
    @@ -714,6 +890,18 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col1: string, _col2: string, _col5: int
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col1
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int)
    @@ -749,6 +937,39 @@ STAGE PLANS:
                      outputColumnNames: _col0, _col1, _col2
                      Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col0: string, _col1: string, _col2: int
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: _col0
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col1
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol1
    + arguments: _col1
    + name: dense_rank
    + window function: GenericUDAFDenseRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol2
    + arguments: _col2, 1, _col2
    + name: lag
    + window function: GenericUDAFLagEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
                        Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _wcol0 (type: int), _wcol1 (type: int), _col2 (type: int), (_col2 - _wcol2) (type: int)
    @@ -876,6 +1097,18 @@ STAGE PLANS:
                      outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col0: int, _col1: string, _col2: string, _col3: string, _col4: string, _col5: int, _col6: string, _col7: double, _col8: string
    + type: TABLE
    + Partition table definition
    + input alias: abc
    + name: noop
    + order by: _col1
    + output shape: _col0: int, _col1: string, _col2: string, _col3: string, _col4: string, _col5: int, _col6: string, _col7: double, _col8: string
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Filter Operator
                          predicate: _col0 is not null (type: boolean)
    @@ -1034,6 +1267,18 @@ STAGE PLANS:
                      outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col0: int, _col1: string, _col2: string, _col3: string, _col4: string, _col5: int, _col6: string, _col7: double, _col8: string
    + type: TABLE
    + Partition table definition
    + input alias: abc
    + name: noop
    + order by: _col1
    + output shape: _col0: int, _col1: string, _col2: string, _col3: string, _col4: string, _col5: int, _col6: string, _col7: double, _col8: string
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Filter Operator
                          predicate: _col0 is not null (type: boolean)
    @@ -1129,7 +1374,21 @@ STAGE PLANS:
                        alias: part
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: p_name: string, p_mfgr: string, p_size: int
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noopwithmap
    + order by: p_name, p_size(DESC)
    + output shape: p_name: string, p_mfgr: string, p_size: int
    + partition by: p_mfgr
    + raw input shape:
    + transforms raw input: true
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
    + Map-side function: true
                          Reduce Output Operator
                            key expressions: p_mfgr (type: string), p_name (type: string), p_size (type: int)
                            sort order: ++-
    @@ -1142,6 +1401,19 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col1: string, _col2: string, _col5: int
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noopwithmap
    + order by: _col1, _col5(DESC)
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2
    + raw input shape:
    + transforms raw input: true
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int)
    @@ -1155,6 +1427,25 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1, _col5(DESC)
    + partition by: _col2
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col1, _col5
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int), _wcol0 (type: int)
    @@ -1254,7 +1545,21 @@ STAGE PLANS:
                        alias: part
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: p_name: string, p_mfgr: string, p_size: int, p_retailprice: double
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noopwithmap
    + order by: p_name
    + output shape: p_name: string, p_mfgr: string, p_size: int, p_retailprice: double
    + partition by: p_mfgr
    + raw input shape:
    + transforms raw input: true
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
    + Map-side function: true
                          Reduce Output Operator
                            key expressions: p_mfgr (type: string), p_name (type: string)
                            sort order: ++
    @@ -1268,6 +1573,19 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5, _col7
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noopwithmap
    + order by: _col1
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + partition by: _col2
    + raw input shape:
    + transforms raw input: true
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col2 (type: string), _col1 (type: string)
    @@ -1282,6 +1600,38 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5, _col7
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: _col2
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col1
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol1
    + arguments: _col1
    + name: dense_rank
    + window function: GenericUDAFDenseRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol2
    + arguments: _col7
    + name: sum
    + window function: GenericUDAFSumDouble
    + window frame: PRECEDING(MAX)~
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int), _wcol0 (type: int), _wcol1 (type: int), _wcol2 (type: double)
    @@ -1397,6 +1747,18 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5, _col7
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col1
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col2 (type: string), _col1 (type: string)
    @@ -1411,6 +1773,38 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5, _col7
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: _col2
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col1
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol1
    + arguments: _col1
    + name: dense_rank
    + window function: GenericUDAFDenseRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol2
    + arguments: _col7
    + name: sum
    + window function: GenericUDAFSumDouble
    + window frame: PRECEDING(MAX)~
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int), _wcol0 (type: int), _wcol1 (type: int), _wcol2 (type: double)
    @@ -1484,7 +1878,7 @@ dense_rank() over (partition by p_mfgr o
      sum(p_retailprice) over (partition by p_mfgr order by p_name rows between unbounded preceding and current row) as s1
      from noop(on noopwithmap(on noop(on part
      partition by p_mfgr
    -order by p_mfgr, p_name
    +order by p_mfgr DESC, p_name
      )))
      PREHOOK: type: QUERY
      POSTHOOK: query: -- 12. testFunctionChain
    @@ -1495,7 +1889,7 @@ dense_rank() over (partition by p_mfgr o
      sum(p_retailprice) over (partition by p_mfgr order by p_name rows between unbounded preceding and current row) as s1
      from noop(on noopwithmap(on noop(on part
      partition by p_mfgr
    -order by p_mfgr, p_name
    +order by p_mfgr DESC, p_name
      )))
      POSTHOOK: type: QUERY
      STAGE DEPENDENCIES:
    @@ -1517,34 +1911,87 @@ STAGE PLANS:
                        alias: part
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
    - key expressions: p_mfgr (type: string), p_mfgr (type: string), p_name (type: string)
    - sort order: +++
    + key expressions: p_mfgr (type: string), p_name (type: string)
    + sort order: -+
                          Map-reduce partition columns: p_mfgr (type: string)
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                          value expressions: p_size (type: int), p_retailprice (type: double)
              Reducer 2
                  Reduce Operator Tree:
                    Select Operator
    - expressions: KEY.reducesinkkey2 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int), VALUE._col5 (type: double)
    + expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int), VALUE._col5 (type: double)
                      outputColumnNames: _col1, _col2, _col5, _col7
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col2(DESC), _col1
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: PTFCOMPONENT
    + Partition table definition
    + input alias: ptf_1
    + name: noopwithmap
    + order by: _col2(DESC), _col1
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + partition by: _col2
    + raw input shape:
    + transforms raw input: true
    + Partition table definition
    + input alias: ptf_2
    + name: noop
    + order by: _col2(DESC), _col1
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + partition by: _col2
    + raw input shape:
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
    + Map-side function: true
                          Reduce Output Operator
    - key expressions: _col2 (type: string), _col2 (type: string), _col1 (type: string)
    - sort order: +++
    + key expressions: _col2 (type: string), _col1 (type: string)
    + sort order: -+
                            Map-reduce partition columns: _col2 (type: string)
                            Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                            value expressions: _col5 (type: int), _col7 (type: double)
              Reducer 3
                  Reduce Operator Tree:
                    Select Operator
    - expressions: KEY.reducesinkkey2 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int), VALUE._col5 (type: double)
    + expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col3 (type: int), VALUE._col5 (type: double)
                      outputColumnNames: _col1, _col2, _col5, _col7
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: PTFCOMPONENT
    + Partition table definition
    + input alias: ptf_1
    + name: noopwithmap
    + order by: _col2(DESC), _col1
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + partition by: _col2
    + raw input shape:
    + transforms raw input: true
    + Partition table definition
    + input alias: ptf_2
    + name: noop
    + order by: _col2(DESC), _col1
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col2 (type: string), _col1 (type: string)
    @@ -1559,6 +2006,38 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5, _col7
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: _col2
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col1
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol1
    + arguments: _col1
    + name: dense_rank
    + window function: GenericUDAFDenseRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol2
    + arguments: _col7
    + name: sum
    + window function: GenericUDAFSumDouble
    + window frame: PRECEDING(MAX)~
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int), _wcol0 (type: int), _wcol1 (type: int), _wcol2 (type: double)
    @@ -1584,7 +2063,7 @@ dense_rank() over (partition by p_mfgr o
      sum(p_retailprice) over (partition by p_mfgr order by p_name rows between unbounded preceding and current row) as s1
      from noop(on noopwithmap(on noop(on part
      partition by p_mfgr
    -order by p_mfgr, p_name
    +order by p_mfgr DESC, p_name
      )))
      PREHOOK: type: QUERY
      PREHOOK: Input: default@part
    @@ -1595,7 +2074,7 @@ dense_rank() over (partition by p_mfgr o
      sum(p_retailprice) over (partition by p_mfgr order by p_name rows between unbounded preceding and current row) as s1
      from noop(on noopwithmap(on noop(on part
      partition by p_mfgr
    -order by p_mfgr, p_name
    +order by p_mfgr DESC, p_name
      )))
      POSTHOOK: type: QUERY
      POSTHOOK: Input: default@part
    @@ -1684,6 +2163,18 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5, _col7
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col1
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col2 (type: string), _col1 (type: string)
    @@ -1698,6 +2189,30 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5, _col7
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: _col2
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col5
    + name: count
    + window function: GenericUDAFCountEvaluator
    + window frame: PRECEDING(MAX)~
    + window function definition
    + alias: _wcol1
    + arguments: _col7
    + name: sum
    + window function: GenericUDAFSumDouble
    + window frame: PRECEDING(2)~FOLLOWING(2)
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _wcol0 (type: bigint), _wcol1 (type: double)
    @@ -1841,6 +2356,18 @@ STAGE PLANS:
                      outputColumnNames: _col0, _col1, _col2, _col5, _col7
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col0: int, _col1: string, _col2: string, _col5: int, _col7: double
    + type: TABLE
    + Partition table definition
    + input alias: abc
    + name: noop
    + order by: _col1
    + output shape: _col0: int, _col1: string, _col2: string, _col5: int, _col7: double
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Filter Operator
                          predicate: _col0 is not null (type: boolean)
    @@ -1874,6 +2401,51 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5, _col7
                      Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: _col2
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col1
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol1
    + arguments: _col1
    + name: dense_rank
    + window function: GenericUDAFDenseRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol2
    + arguments: _col1
    + name: count
    + window function: GenericUDAFCountEvaluator
    + window frame: PRECEDING(MAX)~
    + window function definition
    + alias: _wcol3
    + arguments: _col7
    + name: sum
    + window function: GenericUDAFSumDouble
    + window frame: PRECEDING(MAX)~
    + window function definition
    + alias: _wcol4
    + arguments: _col5, 1, _col5
    + name: lag
    + window function: GenericUDAFLagEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
                        Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _wcol0 (type: int), _wcol1 (type: int), _wcol2 (type: bigint), _col7 (type: double), _wcol3 (type: double), _col5 (type: int), (_col5 - _wcol4) (type: int)
    @@ -1991,6 +2563,18 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col1: string, _col2: string, _col5: int
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col1
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int)
    @@ -2112,7 +2696,6 @@ STAGE PLANS:
            Edges:
              Reducer 2 <- Map 1 (SIMPLE_EDGE)
              Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
    - Reducer 4 <- Reducer 3 (SIMPLE_EDGE)
      #### A masked pattern was here ####
            Vertices:
              Map 1
    @@ -2133,7 +2716,7 @@ STAGE PLANS:
                            Reduce Output Operator
                              key expressions: _col0 (type: string), _col1 (type: string)
                              sort order: ++
    - Map-reduce partition columns: _col0 (type: string), _col1 (type: string)
    + Map-reduce partition columns: _col0 (type: string)
                              Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                              value expressions: _col2 (type: double)
              Reducer 2
    @@ -2144,19 +2727,19 @@ STAGE PLANS:
                      mode: mergepartial
                      outputColumnNames: _col0, _col1, _col2
                      Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE
    - Reduce Output Operator
    - key expressions: _col0 (type: string), _col0 (type: string)
    - sort order: ++
    - Map-reduce partition columns: _col0 (type: string)
    - Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE
    - value expressions: _col1 (type: string), _col2 (type: double)
    - Reducer 3
    - Reduce Operator Tree:
    - Select Operator
    - expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: string), VALUE._col1 (type: double)
    - outputColumnNames: _col0, _col1, _col2
    - Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: mfgr_price_view
    + output shape: _col0: string, _col1: string, _col2: double
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col0
    + output shape: _col0: string, _col1: string, _col2: double
    + partition by: _col0
    + raw input shape:
                        Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col0 (type: string), _col1 (type: string)
    @@ -2164,13 +2747,31 @@ STAGE PLANS:
                          Map-reduce partition columns: _col0 (type: string)
                          Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE
                          value expressions: _col2 (type: double)
    - Reducer 4
    + Reducer 3
                  Reduce Operator Tree:
                    Select Operator
                      expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), VALUE._col0 (type: double)
                      outputColumnNames: _col0, _col1, _col2
                      Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col0: string, _col1: string, _col2: double
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: _col0
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col2
    + name: sum
    + window function: GenericUDAFSumDouble
    + window frame: PRECEDING(2)~
                        Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double), _wcol0 (type: double)
    @@ -2340,6 +2941,18 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5, _col7
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col1
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col2 (type: string), _col1 (type: string)
    @@ -2360,6 +2973,38 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5, _col7
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int, _col7: double
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: _col2
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col1
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol1
    + arguments: _col1
    + name: dense_rank
    + window function: GenericUDAFDenseRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol2
    + arguments: _col7
    + name: sum
    + window function: GenericUDAFSumDouble
    + window frame: PRECEDING(MAX)~
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int), _wcol0 (type: int), _wcol1 (type: int), _wcol2 (type: double)
    @@ -2380,20 +3025,81 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col5
    + partition by: _col2
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col5
    + name: sum
    + window function: GenericUDAFSumLong
    + window frame: PRECEDING(5)~
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
    - Reduce Output Operator
    - key expressions: _col2 (type: string), _col2 (type: string), _col1 (type: string)
    - sort order: +++
    - Map-reduce partition columns: _col2 (type: string)
    + Select Operator
    + expressions: _col1 (type: string), _col2 (type: string), _col5 (type: int), _wcol0 (type: bigint)
    + outputColumnNames: _col1, _col2, _col5, _wcol0
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
    - value expressions: _wcol0 (type: bigint), _col5 (type: int)
    + Reduce Output Operator
    + key expressions: _col2 (type: string), _col1 (type: string)
    + sort order: ++
    + Map-reduce partition columns: _col2 (type: string)
    + Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
    + value expressions: _wcol0 (type: bigint), _col5 (type: int)
              Reducer 5
                  Reduce Operator Tree:
                    Select Operator
    - expressions: VALUE._col0 (type: bigint), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col4 (type: int)
    + expressions: VALUE._col0 (type: bigint), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), VALUE._col4 (type: int)
                      outputColumnNames: _col0, _col2, _col3, _col6
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col0: bigint, _col2: string, _col3: string, _col6: int
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col3, _col2
    + partition by: _col3
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol1
    + arguments: _col3, _col2
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol2
    + arguments: _col3, _col2
    + name: dense_rank
    + window function: GenericUDAFDenseRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol3
    + arguments: _col3, _col2
    + name: cume_dist
    + window function: GenericUDAFCumeDistEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol4
    + arguments: _col6, true
    + name: first_value
    + window function: GenericUDAFFirstValueEvaluator
    + window frame: PRECEDING(2)~FOLLOWING(2)
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col3 (type: string), _col2 (type: string), _col6 (type: int), UDFToInteger(round(_col0, 1)) (type: int), _wcol1 (type: int), _wcol2 (type: int), _wcol3 (type: double), _wcol4 (type: int)
    @@ -2610,8 +3316,8 @@ STAGE PLANS:
                        alias: part
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
    - key expressions: p_mfgr (type: string), p_mfgr (type: string)
    - sort order: ++
    + key expressions: p_mfgr (type: string)
    + sort order: +
                          Map-reduce partition columns: p_mfgr (type: string)
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                          value expressions: p_name (type: string), p_size (type: int)
    @@ -2622,12 +3328,52 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col1: string, _col2: string, _col5: int
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col2
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2
    + raw input shape:
    + Partition table definition
    + input alias: ptf_2
    + name: noop
    + order by: _col2
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int
    + type: PTFCOMPONENT
    + Partition table definition
    + input alias: ptf_1
    + name: noopwithmap
    + order by: _col2, _col1
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2, _col1
    + raw input shape:
    + transforms raw input: true
    + Partition table definition
    + input alias: ptf_2
    + name: noop
    + order by: _col2, _col1
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2, _col1
    + raw input shape:
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
    + Map-side function: true
                          Reduce Output Operator
    - key expressions: _col2 (type: string), _col1 (type: string), _col2 (type: string), _col1 (type: string)
    - sort order: ++++
    + key expressions: _col2 (type: string), _col1 (type: string)
    + sort order: ++
                            Map-reduce partition columns: _col2 (type: string), _col1 (type: string)
                            Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                            value expressions: _col5 (type: int)
    @@ -2638,10 +3384,30 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int
    + type: PTFCOMPONENT
    + Partition table definition
    + input alias: ptf_1
    + name: noopwithmap
    + order by: _col2, _col1
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2, _col1
    + raw input shape:
    + transforms raw input: true
    + Partition table definition
    + input alias: ptf_2
    + name: noop
    + order by: _col2, _col1
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2, _col1
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
    - key expressions: _col2 (type: string), _col1 (type: string), _col2 (type: string), _col1 (type: string)
    - sort order: ++++
    + key expressions: _col2 (type: string), _col1 (type: string)
    + sort order: ++
                          Map-reduce partition columns: _col2 (type: string), _col1 (type: string)
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                          value expressions: _col5 (type: int)
    @@ -2652,6 +3418,38 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col2, _col1
    + partition by: _col2, _col1
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col2, _col1
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol1
    + arguments: _col2, _col1
    + name: dense_rank
    + window function: GenericUDAFDenseRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol2
    + arguments: _col5
    + name: sum
    + window function: GenericUDAFSumLong
    + window frame: PRECEDING(MAX)~
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _wcol0 (type: int), _wcol1 (type: int), _col5 (type: int), _wcol2 (type: bigint)
    @@ -2789,8 +3587,8 @@ STAGE PLANS:
                        alias: part
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
    - key expressions: p_mfgr (type: string), p_mfgr (type: string)
    - sort order: ++
    + key expressions: p_mfgr (type: string)
    + sort order: +
                          Map-reduce partition columns: p_mfgr (type: string)
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                          value expressions: p_name (type: string), p_size (type: int)
    @@ -2801,10 +3599,29 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col1: string, _col2: string, _col5: int
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col2
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2
    + raw input shape:
    + Partition table definition
    + input alias: ptf_2
    + name: noop
    + order by: _col2
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
    - key expressions: _col2 (type: string), _col1 (type: string), _col2 (type: string), _col1 (type: string)
    - sort order: ++++
    + key expressions: _col2 (type: string), _col1 (type: string)
    + sort order: ++
                          Map-reduce partition columns: _col2 (type: string), _col1 (type: string)
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                          value expressions: _col5 (type: int)
    @@ -2815,10 +3632,22 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int
    + type: PTFCOMPONENT
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col2, _col1
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2, _col1
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
    - key expressions: _col2 (type: string), _col2 (type: string)
    - sort order: ++
    + key expressions: _col2 (type: string)
    + sort order: +
                          Map-reduce partition columns: _col2 (type: string)
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                          value expressions: _col1 (type: string), _col5 (type: int)
    @@ -2829,6 +3658,18 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int
    + type: PTFCOMPONENT
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col2
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col2 (type: string), _col1 (type: string)
    @@ -2843,6 +3684,38 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: _col2
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col1
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol1
    + arguments: _col1
    + name: dense_rank
    + window function: GenericUDAFDenseRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol2
    + arguments: _col5
    + name: sum
    + window function: GenericUDAFSumLong
    + window frame: PRECEDING(MAX)~
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _wcol0 (type: int), _wcol1 (type: int), _col5 (type: int), _wcol2 (type: bigint)
    @@ -2975,8 +3848,8 @@ STAGE PLANS:
                        alias: part
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
    - key expressions: p_mfgr (type: string), p_name (type: string), p_mfgr (type: string), p_name (type: string)
    - sort order: ++++
    + key expressions: p_mfgr (type: string), p_name (type: string)
    + sort order: ++
                          Map-reduce partition columns: p_mfgr (type: string), p_name (type: string)
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                          value expressions: p_size (type: int)
    @@ -2987,10 +3860,29 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col1: string, _col2: string, _col5: int
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col2, _col1
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2, _col1
    + raw input shape:
    + Partition table definition
    + input alias: ptf_2
    + name: noop
    + order by: _col2, _col1
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2, _col1
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
    - key expressions: _col2 (type: string), _col2 (type: string)
    - sort order: ++
    + key expressions: _col2 (type: string)
    + sort order: +
                          Map-reduce partition columns: _col2 (type: string)
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                          value expressions: _col1 (type: string), _col5 (type: int)
    @@ -3001,6 +3893,25 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int
    + type: PTFCOMPONENT
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col2
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2
    + raw input shape:
    + Partition table definition
    + input alias: ptf_2
    + name: noop
    + order by: _col2
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
                          key expressions: _col2 (type: string), _col1 (type: string)
    @@ -3015,6 +3926,38 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int
    + type: WINDOWING
    + Windowing table definition
    + input alias: ptf_1
    + name: windowingtablefunction
    + order by: _col1
    + partition by: _col2
    + raw input shape:
    + window functions:
    + window function definition
    + alias: _wcol0
    + arguments: _col1
    + name: rank
    + window function: GenericUDAFRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol1
    + arguments: _col1
    + name: dense_rank
    + window function: GenericUDAFDenseRankEvaluator
    + window frame: PRECEDING(MAX)~FOLLOWING(MAX)
    + isPivotResult: true
    + window function definition
    + alias: _wcol2
    + arguments: _col5
    + name: sum
    + window function: GenericUDAFSumLong
    + window frame: PRECEDING(MAX)~
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Select Operator
                          expressions: _col2 (type: string), _col1 (type: string), _wcol0 (type: int), _wcol1 (type: int), _col5 (type: int), _wcol2 (type: bigint)
    @@ -3148,8 +4091,8 @@ STAGE PLANS:
                        alias: part
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
    - key expressions: p_mfgr (type: string), p_name (type: string), p_mfgr (type: string), p_name (type: string)
    - sort order: ++++
    + key expressions: p_mfgr (type: string), p_name (type: string)
    + sort order: ++
                          Map-reduce partition columns: p_mfgr (type: string), p_name (type: string)
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                          value expressions: p_size (type: int)
    @@ -3160,10 +4103,29 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: part
    + output shape: _col1: string, _col2: string, _col5: int
    + type: TABLE
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col2, _col1
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2, _col1
    + raw input shape:
    + Partition table definition
    + input alias: ptf_2
    + name: noop
    + order by: _col2, _col1
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2, _col1
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
    - key expressions: _col2 (type: string), _col2 (type: string)
    - sort order: ++
    + key expressions: _col2 (type: string)
    + sort order: +
                          Map-reduce partition columns: _col2 (type: string)
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                          value expressions: _col1 (type: string), _col5 (type: int)
    @@ -3174,12 +4136,38 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int
    + type: PTFCOMPONENT
    + Partition table definition
    + input alias: ptf_1
    + name: noop
    + order by: _col2
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2
    + raw input shape:
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int
    + type: PTFCOMPONENT
    + Partition table definition
    + input alias: ptf_1
    + name: noopwithmap
    + order by: _col2, _col1
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2, _col1
    + raw input shape:
    + transforms raw input: true
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
    + Map-side function: true
                          Reduce Output Operator
    - key expressions: _col2 (type: string), _col1 (type: string), _col2 (type: string), _col1 (type: string)
    - sort order: ++++
    + key expressions: _col2 (type: string), _col1 (type: string)
    + sort order: ++
                            Map-reduce partition columns: _col2 (type: string), _col1 (type: string)
                            Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                            value expressions: _col5 (type: int)
    @@ -3190,10 +4178,23 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0
    + output shape: _col1: string, _col2: string, _col5: int
    + type: PTFCOMPONENT
    + Partition table definition
    + input alias: ptf_1
    + name: noopwithmap
    + order by: _col2, _col1
    + output shape: _col1: string, _col2: string, _col5: int
    + partition by: _col2, _col1
    + raw input shape:
    + transforms raw input: true
                        Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                        Reduce Output Operator
    - key expressions: _col2 (type: string), _col1 (type: string), _col2 (type: string), _col1 (type: string)
    - sort order: ++++
    + key expressions: _col2 (type: string), _col1 (type: string)
    + sort order: ++
                          Map-reduce partition columns: _col2 (type: string), _col1 (type: string)
                          Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                          value expressions: _col5 (type: int)
    @@ -3204,6 +4205,38 @@ STAGE PLANS:
                      outputColumnNames: _col1, _col2, _col5
                      Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
                      PTF Operator
    + Function definitions:
    + Input definition
    + input alias: ptf_0

    [... 282 lines stripped ...]
  • Prasanthj at Feb 21, 2015 at 2:37 am
    Modified: hive/branches/llap/ql/src/test/results/clientpositive/orc_analyze.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/orc_analyze.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/orc_analyze.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/orc_analyze.q.out Sat Feb 21 02:36:54 2015
    @@ -71,6 +71,55 @@ POSTHOOK: Lineage: orc_create_people.las
      POSTHOOK: Lineage: orc_create_people.salary SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:salary, type:decimal(10,0), comment:null), ]
      POSTHOOK: Lineage: orc_create_people.start_date SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:start_date, type:timestamp, comment:null), ]
      POSTHOOK: Lineage: orc_create_people.state SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:state, type:string, comment:null), ]
    +PREHOOK: query: analyze table orc_create_people compute statistics
    +PREHOOK: type: QUERY
    +PREHOOK: Input: default@orc_create_people
    +PREHOOK: Output: default@orc_create_people
    +POSTHOOK: query: analyze table orc_create_people compute statistics
    +POSTHOOK: type: QUERY
    +POSTHOOK: Input: default@orc_create_people
    +POSTHOOK: Output: default@orc_create_people
    +PREHOOK: query: desc formatted orc_create_people
    +PREHOOK: type: DESCTABLE
    +PREHOOK: Input: default@orc_create_people
    +POSTHOOK: query: desc formatted orc_create_people
    +POSTHOOK: type: DESCTABLE
    +POSTHOOK: Input: default@orc_create_people
    +# col_name data_type comment
    +
    +id int
    +first_name string
    +last_name string
    +address string
    +salary decimal(10,0)
    +start_date timestamp
    +state string
    +
    +# Detailed Table Information
    +Database: default
    +#### A masked pattern was here ####
    +Protect Mode: None
    +Retention: 0
    +#### A masked pattern was here ####
    +Table Type: MANAGED_TABLE
    +Table Parameters:
    + COLUMN_STATS_ACCURATE true
    + numFiles 1
    + numRows 100
    + rawDataSize 52600
    + totalSize 3158
    +#### A masked pattern was here ####
    +
    +# Storage Information
    +SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde
    +InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
    +OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
    +Compressed: No
    +Num Buckets: -1
    +Bucket Columns: []
    +Sort Columns: []
    +Storage Desc Params:
    + serialization.format 1
      PREHOOK: query: analyze table orc_create_people compute statistics partialscan
      PREHOOK: type: QUERY
      PREHOOK: Input: default@orc_create_people
    @@ -120,6 +169,55 @@ Bucket Columns: []
      Sort Columns: []
      Storage Desc Params:
       serialization.format 1
    +PREHOOK: query: analyze table orc_create_people compute statistics noscan
    +PREHOOK: type: QUERY
    +PREHOOK: Input: default@orc_create_people
    +PREHOOK: Output: default@orc_create_people
    +POSTHOOK: query: analyze table orc_create_people compute statistics noscan
    +POSTHOOK: type: QUERY
    +POSTHOOK: Input: default@orc_create_people
    +POSTHOOK: Output: default@orc_create_people
    +PREHOOK: query: desc formatted orc_create_people
    +PREHOOK: type: DESCTABLE
    +PREHOOK: Input: default@orc_create_people
    +POSTHOOK: query: desc formatted orc_create_people
    +POSTHOOK: type: DESCTABLE
    +POSTHOOK: Input: default@orc_create_people
    +# col_name data_type comment
    +
    +id int
    +first_name string
    +last_name string
    +address string
    +salary decimal(10,0)
    +start_date timestamp
    +state string
    +
    +# Detailed Table Information
    +Database: default
    +#### A masked pattern was here ####
    +Protect Mode: None
    +Retention: 0
    +#### A masked pattern was here ####
    +Table Type: MANAGED_TABLE
    +Table Parameters:
    + COLUMN_STATS_ACCURATE true
    + numFiles 1
    + numRows 100
    + rawDataSize 52600
    + totalSize 3158
    +#### A masked pattern was here ####
    +
    +# Storage Information
    +SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde
    +InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
    +OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
    +Compressed: No
    +Num Buckets: -1
    +Bucket Columns: []
    +Sort Columns: []
    +Storage Desc Params:
    + serialization.format 1
      PREHOOK: query: drop table orc_create_people
      PREHOOK: type: DROPTABLE
      PREHOOK: Input: default@orc_create_people
    @@ -269,13 +367,13 @@ POSTHOOK: Lineage: orc_create_people PAR
      POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).last_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:last_name, type:string, comment:null), ]
      POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).salary SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:salary, type:decimal(10,0), comment:null), ]
      POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).start_date SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:start_date, type:timestamp, comment:null), ]
    -PREHOOK: query: analyze table orc_create_people partition(state) compute statistics partialscan
    +PREHOOK: query: analyze table orc_create_people partition(state) compute statistics
      PREHOOK: type: QUERY
      PREHOOK: Input: default@orc_create_people
      PREHOOK: Output: default@orc_create_people
      PREHOOK: Output: default@orc_create_people@state=Ca
      PREHOOK: Output: default@orc_create_people@state=Or
    -POSTHOOK: query: analyze table orc_create_people partition(state) compute statistics partialscan
    +POSTHOOK: query: analyze table orc_create_people partition(state) compute statistics
      POSTHOOK: type: QUERY
      POSTHOOK: Input: default@orc_create_people
      POSTHOOK: Output: default@orc_create_people
    @@ -371,63 +469,18 @@ Bucket Columns: []
      Sort Columns: []
      Storage Desc Params:
       serialization.format 1
    -PREHOOK: query: drop table orc_create_people
    -PREHOOK: type: DROPTABLE
    +PREHOOK: query: analyze table orc_create_people partition(state) compute statistics partialscan
    +PREHOOK: type: QUERY
      PREHOOK: Input: default@orc_create_people
      PREHOOK: Output: default@orc_create_people
    -POSTHOOK: query: drop table orc_create_people
    -POSTHOOK: type: DROPTABLE
    +PREHOOK: Output: default@orc_create_people@state=Ca
    +PREHOOK: Output: default@orc_create_people@state=Or
    +POSTHOOK: query: analyze table orc_create_people partition(state) compute statistics partialscan
    +POSTHOOK: type: QUERY
      POSTHOOK: Input: default@orc_create_people
      POSTHOOK: Output: default@orc_create_people
    -PREHOOK: query: -- auto stats gather
    -CREATE TABLE orc_create_people (
    - id int,
    - first_name string,
    - last_name string,
    - address string,
    - salary decimal,
    - start_date timestamp)
    -PARTITIONED BY (state string)
    -STORED AS orc
    -PREHOOK: type: CREATETABLE
    -PREHOOK: Output: database:default
    -PREHOOK: Output: default@orc_create_people
    -POSTHOOK: query: -- auto stats gather
    -CREATE TABLE orc_create_people (
    - id int,
    - first_name string,
    - last_name string,
    - address string,
    - salary decimal,
    - start_date timestamp)
    -PARTITIONED BY (state string)
    -STORED AS orc
    -POSTHOOK: type: CREATETABLE
    -POSTHOOK: Output: database:default
    -POSTHOOK: Output: default@orc_create_people
    -PREHOOK: query: INSERT OVERWRITE TABLE orc_create_people PARTITION (state)
    - SELECT * FROM orc_create_people_staging ORDER BY id
    -PREHOOK: type: QUERY
    -PREHOOK: Input: default@orc_create_people_staging
    -PREHOOK: Output: default@orc_create_people
    -POSTHOOK: query: INSERT OVERWRITE TABLE orc_create_people PARTITION (state)
    - SELECT * FROM orc_create_people_staging ORDER BY id
    -POSTHOOK: type: QUERY
    -POSTHOOK: Input: default@orc_create_people_staging
      POSTHOOK: Output: default@orc_create_people@state=Ca
      POSTHOOK: Output: default@orc_create_people@state=Or
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).address SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:address, type:string, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).first_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:first_name, type:string, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).id SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:id, type:int, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).last_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:last_name, type:string, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).salary SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:salary, type:decimal(10,0), comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).start_date SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:start_date, type:timestamp, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).address SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:address, type:string, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).first_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:first_name, type:string, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).id SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:id, type:int, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).last_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:last_name, type:string, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).salary SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:salary, type:decimal(10,0), comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).start_date SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:start_date, type:timestamp, comment:null), ]
      PREHOOK: query: desc formatted orc_create_people partition(state="Ca")
      PREHOOK: type: DESCTABLE
      PREHOOK: Input: default@orc_create_people
    @@ -518,78 +571,13 @@ Bucket Columns: []
      Sort Columns: []
      Storage Desc Params:
       serialization.format 1
    -PREHOOK: query: drop table orc_create_people
    -PREHOOK: type: DROPTABLE
    -PREHOOK: Input: default@orc_create_people
    -PREHOOK: Output: default@orc_create_people
    -POSTHOOK: query: drop table orc_create_people
    -POSTHOOK: type: DROPTABLE
    -POSTHOOK: Input: default@orc_create_people
    -POSTHOOK: Output: default@orc_create_people
    -PREHOOK: query: -- partitioned and bucketed table
    --- partial scan gather
    -CREATE TABLE orc_create_people (
    - id int,
    - first_name string,
    - last_name string,
    - address string,
    - salary decimal,
    - start_date timestamp)
    -PARTITIONED BY (state string)
    -clustered by (first_name)
    -sorted by (last_name)
    -into 4 buckets
    -STORED AS orc
    -PREHOOK: type: CREATETABLE
    -PREHOOK: Output: database:default
    -PREHOOK: Output: default@orc_create_people
    -POSTHOOK: query: -- partitioned and bucketed table
    --- partial scan gather
    -CREATE TABLE orc_create_people (
    - id int,
    - first_name string,
    - last_name string,
    - address string,
    - salary decimal,
    - start_date timestamp)
    -PARTITIONED BY (state string)
    -clustered by (first_name)
    -sorted by (last_name)
    -into 4 buckets
    -STORED AS orc
    -POSTHOOK: type: CREATETABLE
    -POSTHOOK: Output: database:default
    -POSTHOOK: Output: default@orc_create_people
    -PREHOOK: query: INSERT OVERWRITE TABLE orc_create_people PARTITION (state)
    - SELECT * FROM orc_create_people_staging ORDER BY id
    -PREHOOK: type: QUERY
    -PREHOOK: Input: default@orc_create_people_staging
    -PREHOOK: Output: default@orc_create_people
    -POSTHOOK: query: INSERT OVERWRITE TABLE orc_create_people PARTITION (state)
    - SELECT * FROM orc_create_people_staging ORDER BY id
    -POSTHOOK: type: QUERY
    -POSTHOOK: Input: default@orc_create_people_staging
    -POSTHOOK: Output: default@orc_create_people@state=Ca
    -POSTHOOK: Output: default@orc_create_people@state=Or
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).address SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:address, type:string, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).first_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:first_name, type:string, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).id SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:id, type:int, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).last_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:last_name, type:string, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).salary SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:salary, type:decimal(10,0), comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).start_date SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:start_date, type:timestamp, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).address SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:address, type:string, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).first_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:first_name, type:string, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).id SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:id, type:int, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).last_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:last_name, type:string, comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).salary SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:salary, type:decimal(10,0), comment:null), ]
    -POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).start_date SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:start_date, type:timestamp, comment:null), ]
    -PREHOOK: query: analyze table orc_create_people partition(state) compute statistics partialscan
    +PREHOOK: query: analyze table orc_create_people partition(state) compute statistics noscan
      PREHOOK: type: QUERY
      PREHOOK: Input: default@orc_create_people
      PREHOOK: Output: default@orc_create_people
      PREHOOK: Output: default@orc_create_people@state=Ca
      PREHOOK: Output: default@orc_create_people@state=Or
    -POSTHOOK: query: analyze table orc_create_people partition(state) compute statistics partialscan
    +POSTHOOK: query: analyze table orc_create_people partition(state) compute statistics noscan
      POSTHOOK: type: QUERY
      POSTHOOK: Input: default@orc_create_people
      POSTHOOK: Output: default@orc_create_people
    @@ -635,9 +623,9 @@ SerDe Library: org.apache.hadoop.h
      InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
      OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
      Compressed: No
    -Num Buckets: 4
    -Bucket Columns: [first_name]
    -Sort Columns: [Order(col:last_name, order:1)]
    +Num Buckets: -1
    +Bucket Columns: []
    +Sort Columns: []
      Storage Desc Params:
       serialization.format 1
      PREHOOK: query: desc formatted orc_create_people partition(state="Or")
    @@ -680,9 +668,9 @@ SerDe Library: org.apache.hadoop.h
      InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
      OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
      Compressed: No
    -Num Buckets: 4
    -Bucket Columns: [first_name]
    -Sort Columns: [Order(col:last_name, order:1)]
    +Num Buckets: -1
    +Bucket Columns: []
    +Sort Columns: []
      Storage Desc Params:
       serialization.format 1
      PREHOOK: query: drop table orc_create_people
    @@ -702,9 +690,6 @@ CREATE TABLE orc_create_people (
        salary decimal,
        start_date timestamp)
      PARTITIONED BY (state string)
    -clustered by (first_name)
    -sorted by (last_name)
    -into 4 buckets
      STORED AS orc
      PREHOOK: type: CREATETABLE
      PREHOOK: Output: database:default
    @@ -718,9 +703,6 @@ CREATE TABLE orc_create_people (
        salary decimal,
        start_date timestamp)
      PARTITIONED BY (state string)
    -clustered by (first_name)
    -sorted by (last_name)
    -into 4 buckets
      STORED AS orc
      POSTHOOK: type: CREATETABLE
      POSTHOOK: Output: database:default
    @@ -788,9 +770,9 @@ SerDe Library: org.apache.hadoop.h
      InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
      OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
      Compressed: No
    -Num Buckets: 4
    -Bucket Columns: [first_name]
    -Sort Columns: [Order(col:last_name, order:1)]
    +Num Buckets: -1
    +Bucket Columns: []
    +Sort Columns: []
      Storage Desc Params:
       serialization.format 1
      PREHOOK: query: desc formatted orc_create_people partition(state="Or")
    @@ -833,9 +815,9 @@ SerDe Library: org.apache.hadoop.h
      InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
      OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
      Compressed: No
    -Num Buckets: 4
    -Bucket Columns: [first_name]
    -Sort Columns: [Order(col:last_name, order:1)]
    +Num Buckets: -1
    +Bucket Columns: []
    +Sort Columns: []
      Storage Desc Params:
       serialization.format 1
      PREHOOK: query: drop table orc_create_people
    @@ -846,9 +828,533 @@ POSTHOOK: query: drop table orc_create_p
      POSTHOOK: type: DROPTABLE
      POSTHOOK: Input: default@orc_create_people
      POSTHOOK: Output: default@orc_create_people
    -PREHOOK: query: -- create table with partitions containing text and ORC files.
    --- ORC files implements StatsProvidingRecordReader but text files does not.
    --- So the partition containing text file should not have statistics.
    +PREHOOK: query: -- partitioned and bucketed table
    +-- partial scan gather
    +CREATE TABLE orc_create_people (
    + id int,
    + first_name string,
    + last_name string,
    + address string,
    + salary decimal,
    + start_date timestamp)
    +PARTITIONED BY (state string)
    +clustered by (first_name)
    +sorted by (last_name)
    +into 4 buckets
    +STORED AS orc
    +PREHOOK: type: CREATETABLE
    +PREHOOK: Output: database:default
    +PREHOOK: Output: default@orc_create_people
    +POSTHOOK: query: -- partitioned and bucketed table
    +-- partial scan gather
    +CREATE TABLE orc_create_people (
    + id int,
    + first_name string,
    + last_name string,
    + address string,
    + salary decimal,
    + start_date timestamp)
    +PARTITIONED BY (state string)
    +clustered by (first_name)
    +sorted by (last_name)
    +into 4 buckets
    +STORED AS orc
    +POSTHOOK: type: CREATETABLE
    +POSTHOOK: Output: database:default
    +POSTHOOK: Output: default@orc_create_people
    +PREHOOK: query: INSERT OVERWRITE TABLE orc_create_people PARTITION (state)
    + SELECT * FROM orc_create_people_staging ORDER BY id
    +PREHOOK: type: QUERY
    +PREHOOK: Input: default@orc_create_people_staging
    +PREHOOK: Output: default@orc_create_people
    +POSTHOOK: query: INSERT OVERWRITE TABLE orc_create_people PARTITION (state)
    + SELECT * FROM orc_create_people_staging ORDER BY id
    +POSTHOOK: type: QUERY
    +POSTHOOK: Input: default@orc_create_people_staging
    +POSTHOOK: Output: default@orc_create_people@state=Ca
    +POSTHOOK: Output: default@orc_create_people@state=Or
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).address SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:address, type:string, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).first_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:first_name, type:string, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).id SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:id, type:int, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).last_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:last_name, type:string, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).salary SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:salary, type:decimal(10,0), comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).start_date SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:start_date, type:timestamp, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).address SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:address, type:string, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).first_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:first_name, type:string, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).id SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:id, type:int, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).last_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:last_name, type:string, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).salary SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:salary, type:decimal(10,0), comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).start_date SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:start_date, type:timestamp, comment:null), ]
    +PREHOOK: query: analyze table orc_create_people partition(state) compute statistics
    +PREHOOK: type: QUERY
    +PREHOOK: Input: default@orc_create_people
    +PREHOOK: Output: default@orc_create_people
    +PREHOOK: Output: default@orc_create_people@state=Ca
    +PREHOOK: Output: default@orc_create_people@state=Or
    +POSTHOOK: query: analyze table orc_create_people partition(state) compute statistics
    +POSTHOOK: type: QUERY
    +POSTHOOK: Input: default@orc_create_people
    +POSTHOOK: Output: default@orc_create_people
    +POSTHOOK: Output: default@orc_create_people@state=Ca
    +POSTHOOK: Output: default@orc_create_people@state=Or
    +PREHOOK: query: desc formatted orc_create_people partition(state="Ca")
    +PREHOOK: type: DESCTABLE
    +PREHOOK: Input: default@orc_create_people
    +POSTHOOK: query: desc formatted orc_create_people partition(state="Ca")
    +POSTHOOK: type: DESCTABLE
    +POSTHOOK: Input: default@orc_create_people
    +# col_name data_type comment
    +
    +id int
    +first_name string
    +last_name string
    +address string
    +salary decimal(10,0)
    +start_date timestamp
    +
    +# Partition Information
    +# col_name data_type comment
    +
    +state string
    +
    +# Detailed Partition Information
    +Partition Value: [Ca]
    +Database: default
    +Table: orc_create_people
    +#### A masked pattern was here ####
    +Protect Mode: None
    +#### A masked pattern was here ####
    +Partition Parameters:
    + COLUMN_STATS_ACCURATE true
    + numFiles 1
    + numRows 50
    + rawDataSize 21950
    + totalSize 2055
    +#### A masked pattern was here ####
    +
    +# Storage Information
    +SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde
    +InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
    +OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
    +Compressed: No
    +Num Buckets: 4
    +Bucket Columns: [first_name]
    +Sort Columns: [Order(col:last_name, order:1)]
    +Storage Desc Params:
    + serialization.format 1
    +PREHOOK: query: desc formatted orc_create_people partition(state="Or")
    +PREHOOK: type: DESCTABLE
    +PREHOOK: Input: default@orc_create_people
    +POSTHOOK: query: desc formatted orc_create_people partition(state="Or")
    +POSTHOOK: type: DESCTABLE
    +POSTHOOK: Input: default@orc_create_people
    +# col_name data_type comment
    +
    +id int
    +first_name string
    +last_name string
    +address string
    +salary decimal(10,0)
    +start_date timestamp
    +
    +# Partition Information
    +# col_name data_type comment
    +
    +state string
    +
    +# Detailed Partition Information
    +Partition Value: [Or]
    +Database: default
    +Table: orc_create_people
    +#### A masked pattern was here ####
    +Protect Mode: None
    +#### A masked pattern was here ####
    +Partition Parameters:
    + COLUMN_STATS_ACCURATE true
    + numFiles 1
    + numRows 50
    + rawDataSize 22050
    + totalSize 2071
    +#### A masked pattern was here ####
    +
    +# Storage Information
    +SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde
    +InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
    +OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
    +Compressed: No
    +Num Buckets: 4
    +Bucket Columns: [first_name]
    +Sort Columns: [Order(col:last_name, order:1)]
    +Storage Desc Params:
    + serialization.format 1
    +PREHOOK: query: analyze table orc_create_people partition(state) compute statistics partialscan
    +PREHOOK: type: QUERY
    +PREHOOK: Input: default@orc_create_people
    +PREHOOK: Output: default@orc_create_people
    +PREHOOK: Output: default@orc_create_people@state=Ca
    +PREHOOK: Output: default@orc_create_people@state=Or
    +POSTHOOK: query: analyze table orc_create_people partition(state) compute statistics partialscan
    +POSTHOOK: type: QUERY
    +POSTHOOK: Input: default@orc_create_people
    +POSTHOOK: Output: default@orc_create_people
    +POSTHOOK: Output: default@orc_create_people@state=Ca
    +POSTHOOK: Output: default@orc_create_people@state=Or
    +PREHOOK: query: desc formatted orc_create_people partition(state="Ca")
    +PREHOOK: type: DESCTABLE
    +PREHOOK: Input: default@orc_create_people
    +POSTHOOK: query: desc formatted orc_create_people partition(state="Ca")
    +POSTHOOK: type: DESCTABLE
    +POSTHOOK: Input: default@orc_create_people
    +# col_name data_type comment
    +
    +id int
    +first_name string
    +last_name string
    +address string
    +salary decimal(10,0)
    +start_date timestamp
    +
    +# Partition Information
    +# col_name data_type comment
    +
    +state string
    +
    +# Detailed Partition Information
    +Partition Value: [Ca]
    +Database: default
    +Table: orc_create_people
    +#### A masked pattern was here ####
    +Protect Mode: None
    +#### A masked pattern was here ####
    +Partition Parameters:
    + COLUMN_STATS_ACCURATE true
    + numFiles 1
    + numRows 50
    + rawDataSize 21950
    + totalSize 2055
    +#### A masked pattern was here ####
    +
    +# Storage Information
    +SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde
    +InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
    +OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
    +Compressed: No
    +Num Buckets: 4
    +Bucket Columns: [first_name]
    +Sort Columns: [Order(col:last_name, order:1)]
    +Storage Desc Params:
    + serialization.format 1
    +PREHOOK: query: desc formatted orc_create_people partition(state="Or")
    +PREHOOK: type: DESCTABLE
    +PREHOOK: Input: default@orc_create_people
    +POSTHOOK: query: desc formatted orc_create_people partition(state="Or")
    +POSTHOOK: type: DESCTABLE
    +POSTHOOK: Input: default@orc_create_people
    +# col_name data_type comment
    +
    +id int
    +first_name string
    +last_name string
    +address string
    +salary decimal(10,0)
    +start_date timestamp
    +
    +# Partition Information
    +# col_name data_type comment
    +
    +state string
    +
    +# Detailed Partition Information
    +Partition Value: [Or]
    +Database: default
    +Table: orc_create_people
    +#### A masked pattern was here ####
    +Protect Mode: None
    +#### A masked pattern was here ####
    +Partition Parameters:
    + COLUMN_STATS_ACCURATE true
    + numFiles 1
    + numRows 50
    + rawDataSize 22050
    + totalSize 2071
    +#### A masked pattern was here ####
    +
    +# Storage Information
    +SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde
    +InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
    +OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
    +Compressed: No
    +Num Buckets: 4
    +Bucket Columns: [first_name]
    +Sort Columns: [Order(col:last_name, order:1)]
    +Storage Desc Params:
    + serialization.format 1
    +PREHOOK: query: analyze table orc_create_people partition(state) compute statistics noscan
    +PREHOOK: type: QUERY
    +PREHOOK: Input: default@orc_create_people
    +PREHOOK: Output: default@orc_create_people
    +PREHOOK: Output: default@orc_create_people@state=Ca
    +PREHOOK: Output: default@orc_create_people@state=Or
    +POSTHOOK: query: analyze table orc_create_people partition(state) compute statistics noscan
    +POSTHOOK: type: QUERY
    +POSTHOOK: Input: default@orc_create_people
    +POSTHOOK: Output: default@orc_create_people
    +POSTHOOK: Output: default@orc_create_people@state=Ca
    +POSTHOOK: Output: default@orc_create_people@state=Or
    +PREHOOK: query: desc formatted orc_create_people partition(state="Ca")
    +PREHOOK: type: DESCTABLE
    +PREHOOK: Input: default@orc_create_people
    +POSTHOOK: query: desc formatted orc_create_people partition(state="Ca")
    +POSTHOOK: type: DESCTABLE
    +POSTHOOK: Input: default@orc_create_people
    +# col_name data_type comment
    +
    +id int
    +first_name string
    +last_name string
    +address string
    +salary decimal(10,0)
    +start_date timestamp
    +
    +# Partition Information
    +# col_name data_type comment
    +
    +state string
    +
    +# Detailed Partition Information
    +Partition Value: [Ca]
    +Database: default
    +Table: orc_create_people
    +#### A masked pattern was here ####
    +Protect Mode: None
    +#### A masked pattern was here ####
    +Partition Parameters:
    + COLUMN_STATS_ACCURATE true
    + numFiles 1
    + numRows 50
    + rawDataSize 21950
    + totalSize 2055
    +#### A masked pattern was here ####
    +
    +# Storage Information
    +SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde
    +InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
    +OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
    +Compressed: No
    +Num Buckets: 4
    +Bucket Columns: [first_name]
    +Sort Columns: [Order(col:last_name, order:1)]
    +Storage Desc Params:
    + serialization.format 1
    +PREHOOK: query: desc formatted orc_create_people partition(state="Or")
    +PREHOOK: type: DESCTABLE
    +PREHOOK: Input: default@orc_create_people
    +POSTHOOK: query: desc formatted orc_create_people partition(state="Or")
    +POSTHOOK: type: DESCTABLE
    +POSTHOOK: Input: default@orc_create_people
    +# col_name data_type comment
    +
    +id int
    +first_name string
    +last_name string
    +address string
    +salary decimal(10,0)
    +start_date timestamp
    +
    +# Partition Information
    +# col_name data_type comment
    +
    +state string
    +
    +# Detailed Partition Information
    +Partition Value: [Or]
    +Database: default
    +Table: orc_create_people
    +#### A masked pattern was here ####
    +Protect Mode: None
    +#### A masked pattern was here ####
    +Partition Parameters:
    + COLUMN_STATS_ACCURATE true
    + numFiles 1
    + numRows 50
    + rawDataSize 22050
    + totalSize 2071
    +#### A masked pattern was here ####
    +
    +# Storage Information
    +SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde
    +InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
    +OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
    +Compressed: No
    +Num Buckets: 4
    +Bucket Columns: [first_name]
    +Sort Columns: [Order(col:last_name, order:1)]
    +Storage Desc Params:
    + serialization.format 1
    +PREHOOK: query: drop table orc_create_people
    +PREHOOK: type: DROPTABLE
    +PREHOOK: Input: default@orc_create_people
    +PREHOOK: Output: default@orc_create_people
    +POSTHOOK: query: drop table orc_create_people
    +POSTHOOK: type: DROPTABLE
    +POSTHOOK: Input: default@orc_create_people
    +POSTHOOK: Output: default@orc_create_people
    +PREHOOK: query: -- auto stats gather
    +CREATE TABLE orc_create_people (
    + id int,
    + first_name string,
    + last_name string,
    + address string,
    + salary decimal,
    + start_date timestamp)
    +PARTITIONED BY (state string)
    +clustered by (first_name)
    +sorted by (last_name)
    +into 4 buckets
    +STORED AS orc
    +PREHOOK: type: CREATETABLE
    +PREHOOK: Output: database:default
    +PREHOOK: Output: default@orc_create_people
    +POSTHOOK: query: -- auto stats gather
    +CREATE TABLE orc_create_people (
    + id int,
    + first_name string,
    + last_name string,
    + address string,
    + salary decimal,
    + start_date timestamp)
    +PARTITIONED BY (state string)
    +clustered by (first_name)
    +sorted by (last_name)
    +into 4 buckets
    +STORED AS orc
    +POSTHOOK: type: CREATETABLE
    +POSTHOOK: Output: database:default
    +POSTHOOK: Output: default@orc_create_people
    +PREHOOK: query: INSERT OVERWRITE TABLE orc_create_people PARTITION (state)
    + SELECT * FROM orc_create_people_staging ORDER BY id
    +PREHOOK: type: QUERY
    +PREHOOK: Input: default@orc_create_people_staging
    +PREHOOK: Output: default@orc_create_people
    +POSTHOOK: query: INSERT OVERWRITE TABLE orc_create_people PARTITION (state)
    + SELECT * FROM orc_create_people_staging ORDER BY id
    +POSTHOOK: type: QUERY
    +POSTHOOK: Input: default@orc_create_people_staging
    +POSTHOOK: Output: default@orc_create_people@state=Ca
    +POSTHOOK: Output: default@orc_create_people@state=Or
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).address SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:address, type:string, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).first_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:first_name, type:string, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).id SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:id, type:int, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).last_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:last_name, type:string, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).salary SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:salary, type:decimal(10,0), comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Ca).start_date SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:start_date, type:timestamp, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).address SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:address, type:string, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).first_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:first_name, type:string, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).id SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:id, type:int, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).last_name SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:last_name, type:string, comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).salary SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:salary, type:decimal(10,0), comment:null), ]
    +POSTHOOK: Lineage: orc_create_people PARTITION(state=Or).start_date SIMPLE [(orc_create_people_staging)orc_create_people_staging.FieldSchema(name:start_date, type:timestamp, comment:null), ]
    +PREHOOK: query: desc formatted orc_create_people partition(state="Ca")
    +PREHOOK: type: DESCTABLE
    +PREHOOK: Input: default@orc_create_people
    +POSTHOOK: query: desc formatted orc_create_people partition(state="Ca")
    +POSTHOOK: type: DESCTABLE
    +POSTHOOK: Input: default@orc_create_people
    +# col_name data_type comment
    +
    +id int
    +first_name string
    +last_name string
    +address string
    +salary decimal(10,0)
    +start_date timestamp
    +
    +# Partition Information
    +# col_name data_type comment
    +
    +state string
    +
    +# Detailed Partition Information
    +Partition Value: [Ca]
    +Database: default
    +Table: orc_create_people
    +#### A masked pattern was here ####
    +Protect Mode: None
    +#### A masked pattern was here ####
    +Partition Parameters:
    + COLUMN_STATS_ACCURATE true
    + numFiles 1
    + numRows 50
    + rawDataSize 21950
    + totalSize 2055
    +#### A masked pattern was here ####
    +
    +# Storage Information
    +SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde
    +InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
    +OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
    +Compressed: No
    +Num Buckets: 4
    +Bucket Columns: [first_name]
    +Sort Columns: [Order(col:last_name, order:1)]
    +Storage Desc Params:
    + serialization.format 1
    +PREHOOK: query: desc formatted orc_create_people partition(state="Or")
    +PREHOOK: type: DESCTABLE
    +PREHOOK: Input: default@orc_create_people
    +POSTHOOK: query: desc formatted orc_create_people partition(state="Or")
    +POSTHOOK: type: DESCTABLE
    +POSTHOOK: Input: default@orc_create_people
    +# col_name data_type comment
    +
    +id int
    +first_name string
    +last_name string
    +address string
    +salary decimal(10,0)
    +start_date timestamp
    +
    +# Partition Information
    +# col_name data_type comment
    +
    +state string
    +
    +# Detailed Partition Information
    +Partition Value: [Or]
    +Database: default
    +Table: orc_create_people
    +#### A masked pattern was here ####
    +Protect Mode: None
    +#### A masked pattern was here ####
    +Partition Parameters:
    + COLUMN_STATS_ACCURATE true
    + numFiles 1
    + numRows 50
    + rawDataSize 22050
    + totalSize 2071
    +#### A masked pattern was here ####
    +
    +# Storage Information
    +SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde
    +InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
    +OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
    +Compressed: No
    +Num Buckets: 4
    +Bucket Columns: [first_name]
    +Sort Columns: [Order(col:last_name, order:1)]
    +Storage Desc Params:
    + serialization.format 1
    +PREHOOK: query: drop table orc_create_people
    +PREHOOK: type: DROPTABLE
    +PREHOOK: Input: default@orc_create_people
    +PREHOOK: Output: default@orc_create_people
    +POSTHOOK: query: drop table orc_create_people
    +POSTHOOK: type: DROPTABLE
    +POSTHOOK: Input: default@orc_create_people
    +POSTHOOK: Output: default@orc_create_people
    +PREHOOK: query: -- create table with partitions containing text and ORC files.
    +-- ORC files implements StatsProvidingRecordReader but text files does not.
    +-- So the partition containing text file should not have statistics.
      CREATE TABLE orc_create_people (
        id int,
        first_name string,
    @@ -946,6 +1452,214 @@ POSTHOOK: query: ALTER TABLE orc_create_
      POSTHOOK: type: ALTERTABLE_FILEFORMAT
      POSTHOOK: Input: default@orc_create_people
      POSTHOOK: Output: default@orc_create_people
    +PREHOOK: query: analyze table orc_create_people partition(state) compute statistics
    +PREHOOK: type: QUERY
    +PREHOOK: Input: default@orc_create_people
    +PREHOOK: Output: default@orc_create_people
    +PREHOOK: Output: default@orc_create_people@state=Ca
    +PREHOOK: Output: default@orc_create_people@state=OH
    +PREHOOK: Output: default@orc_create_people@state=Or
    +POSTHOOK: query: analyze table orc_create_people partition(state) compute statistics
    +POSTHOOK: type: QUERY
    +POSTHOOK: Input: default@orc_create_people
    +POSTHOOK: Output: default@orc_create_people
    +POSTHOOK: Output: default@orc_create_people@state=Ca
    +POSTHOOK: Output: default@orc_create_people@state=OH
    +POSTHOOK: Output: default@orc_create_people@state=Or
    +PREHOOK: query: desc formatted orc_create_people partition(state="Ca")
    +PREHOOK: type: DESCTABLE
    +PREHOOK: Input: default@orc_create_people
    +POSTHOOK: query: desc formatted orc_create_people partition(state="Ca")
    +POSTHOOK: type: DESCTABLE
    +POSTHOOK: Input: default@orc_create_people
    +# col_name data_type comment
    +
    +id int
    +first_name string
    +last_name string
    +address string
    +salary decimal(10,0)
    +start_date timestamp
    +
    +# Partition Information
    +# col_name data_type comment
    +
    +state string
    +
    +# Detailed Partition Information
    +Partition Value: [Ca]
    +Database: default
    +Table: orc_create_people
    +#### A masked pattern was here ####
    +Protect Mode: None
    +#### A masked pattern was here ####
    +Partition Parameters:
    + COLUMN_STATS_ACCURATE true
    + numFiles 1
    + numRows 50
    + rawDataSize 21950
    + totalSize 2055
    +#### A masked pattern was here ####
    +
    +# Storage Information
    +SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde
    +InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
    +OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
    +Compressed: No
    +Num Buckets: -1
    +Bucket Columns: []
    +Sort Columns: []
    +Storage Desc Params:
    + serialization.format 1
    +PREHOOK: query: desc formatted orc_create_people partition(state="OH")
    +PREHOOK: type: DESCTABLE
    +PREHOOK: Input: default@orc_create_people
    +POSTHOOK: query: desc formatted orc_create_people partition(state="OH")
    +POSTHOOK: type: DESCTABLE
    +POSTHOOK: Input: default@orc_create_people
    +# col_name data_type comment
    +
    +id int
    +first_name string
    +last_name string
    +address string
    +salary decimal(10,0)
    +start_date timestamp
    +
    +# Partition Information
    +# col_name data_type comment
    +
    +state string
    +
    +# Detailed Partition Information
    +Partition Value: [OH]
    +Database: default
    +Table: orc_create_people
    +#### A masked pattern was here ####
    +Protect Mode: None
    +#### A masked pattern was here ####
    +Partition Parameters:
    + COLUMN_STATS_ACCURATE false
    + numFiles 1
    + numRows -1
    + rawDataSize -1
    + totalSize 5812
    +#### A masked pattern was here ####
    +
    +# Storage Information
    +SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    +InputFormat: org.apache.hadoop.mapred.TextInputFormat
    +OutputFormat: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
    +Compressed: No
    +Num Buckets: -1
    +Bucket Columns: []
    +Sort Columns: []
    +Storage Desc Params:
    + serialization.format 1
    +PREHOOK: query: analyze table orc_create_people partition(state) compute statistics partialscan
    +PREHOOK: type: QUERY
    +PREHOOK: Input: default@orc_create_people
    +PREHOOK: Output: default@orc_create_people
    +PREHOOK: Output: default@orc_create_people@state=Ca
    +PREHOOK: Output: default@orc_create_people@state=OH
    +PREHOOK: Output: default@orc_create_people@state=Or
    +POSTHOOK: query: analyze table orc_create_people partition(state) compute statistics partialscan
    +POSTHOOK: type: QUERY
    +POSTHOOK: Input: default@orc_create_people
    +POSTHOOK: Output: default@orc_create_people
    +POSTHOOK: Output: default@orc_create_people@state=Ca
    +POSTHOOK: Output: default@orc_create_people@state=OH
    +POSTHOOK: Output: default@orc_create_people@state=Or
    +PREHOOK: query: desc formatted orc_create_people partition(state="Ca")
    +PREHOOK: type: DESCTABLE
    +PREHOOK: Input: default@orc_create_people
    +POSTHOOK: query: desc formatted orc_create_people partition(state="Ca")
    +POSTHOOK: type: DESCTABLE
    +POSTHOOK: Input: default@orc_create_people
    +# col_name data_type comment
    +
    +id int
    +first_name string
    +last_name string
    +address string
    +salary decimal(10,0)
    +start_date timestamp
    +
    +# Partition Information
    +# col_name data_type comment
    +
    +state string
    +
    +# Detailed Partition Information
    +Partition Value: [Ca]
    +Database: default
    +Table: orc_create_people
    +#### A masked pattern was here ####
    +Protect Mode: None
    +#### A masked pattern was here ####
    +Partition Parameters:
    + COLUMN_STATS_ACCURATE true
    + numFiles 1
    + numRows 50
    + rawDataSize 21950
    + totalSize 2055
    +#### A masked pattern was here ####
    +
    +# Storage Information
    +SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde
    +InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
    +OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
    +Compressed: No
    +Num Buckets: -1
    +Bucket Columns: []
    +Sort Columns: []
    +Storage Desc Params:
    + serialization.format 1
    +PREHOOK: query: desc formatted orc_create_people partition(state="OH")
    +PREHOOK: type: DESCTABLE
    +PREHOOK: Input: default@orc_create_people
    +POSTHOOK: query: desc formatted orc_create_people partition(state="OH")
    +POSTHOOK: type: DESCTABLE
    +POSTHOOK: Input: default@orc_create_people
    +# col_name data_type comment
    +
    +id int
    +first_name string
    +last_name string
    +address string
    +salary decimal(10,0)
    +start_date timestamp
    +
    +# Partition Information
    +# col_name data_type comment
    +
    +state string
    +
    +# Detailed Partition Information
    +Partition Value: [OH]
    +Database: default
    +Table: orc_create_people
    +#### A masked pattern was here ####
    +Protect Mode: None
    +#### A masked pattern was here ####
    +Partition Parameters:
    + COLUMN_STATS_ACCURATE false
    + numFiles 1
    + numRows -1
    + rawDataSize -1
    + totalSize 5812
    +#### A masked pattern was here ####
    +
    +# Storage Information
    +SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    +InputFormat: org.apache.hadoop.mapred.TextInputFormat
    +OutputFormat: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
    +Compressed: No
    +Num Buckets: -1
    +Bucket Columns: []
    +Sort Columns: []
    +Storage Desc Params:
    + serialization.format 1
      PREHOOK: query: analyze table orc_create_people partition(state) compute statistics noscan
      PREHOOK: type: QUERY
      PREHOOK: Input: default@orc_create_people

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/outer_join_ppr.q.java1.7.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/outer_join_ppr.q.java1.7.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/outer_join_ppr.q.java1.7.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/outer_join_ppr.q.java1.7.out Sat Feb 21 02:36:54 2015
    @@ -420,7 +420,7 @@ STAGE PLANS:
                            columns _col0,_col1,_col2,_col3
                            columns.types string:string:string:string
                            escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                            serialization.format 1
                            serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -801,7 +801,7 @@ STAGE PLANS:
                            columns _col0,_col1,_col2,_col3
                            columns.types string:string:string:string
                            escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                            serialization.format 1
                            serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/parallel_join0.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/parallel_join0.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/parallel_join0.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/parallel_join0.q.out Sat Feb 21 02:36:54 2015
    @@ -1,5 +1,5 @@
      Warning: Shuffle Join JOIN[8][tables = [src1, src2]] in Stage 'Stage-1:MAPRED' is a cross product
    -PREHOOK: query: -- SORT_BEFORE_DIFF
    +PREHOOK: query: -- SORT_QUERY_RESULTS

      EXPLAIN
      SELECT src1.key as k1, src1.value as v1,
    @@ -9,7 +9,7 @@ SELECT src1.key as k1, src1.value as v1,
        (SELECT * FROM src WHERE src.key < 10) src2
        SORT BY k1, v1, k2, v2
      PREHOOK: type: QUERY
    -POSTHOOK: query: -- SORT_BEFORE_DIFF
    +POSTHOOK: query: -- SORT_QUERY_RESULTS

      EXPLAIN
      SELECT src1.key as k1, src1.value as v1,
    @@ -138,101 +138,101 @@ POSTHOOK: Input: default@src
      #### A masked pattern was here ####
      0 val_0 0 val_0
      0 val_0 0 val_0
    -0 val_0 2 val_2
    -0 val_0 4 val_4
    -0 val_0 5 val_5
    -0 val_0 5 val_5
    -0 val_0 8 val_8
    -2 val_2 2 val_2
    -2 val_2 5 val_5
    -4 val_4 0 val_0
    -4 val_4 4 val_4
    -4 val_4 5 val_5
    -4 val_4 9 val_9
    -5 val_5 0 val_0
    -5 val_5 0 val_0
    -5 val_5 4 val_4
    -5 val_5 5 val_5
    -5 val_5 8 val_8
    -5 val_5 8 val_8
    -5 val_5 9 val_9
    -5 val_5 9 val_9
    -8 val_8 0 val_0
    -8 val_8 5 val_5
    -8 val_8 5 val_5
    -8 val_8 5 val_5
    -9 val_9 0 val_0
    -9 val_9 0 val_0
    -9 val_9 0 val_0
    -9 val_9 2 val_2
    -9 val_9 5 val_5
    -9 val_9 5 val_5
      0 val_0 0 val_0
      0 val_0 0 val_0
      0 val_0 0 val_0
    -0 val_0 2 val_2
    -0 val_0 5 val_5
    -0 val_0 5 val_5
    -0 val_0 8 val_8
    -0 val_0 9 val_9
    -2 val_2 0 val_0
    -2 val_2 5 val_5
    -4 val_4 5 val_5
    -5 val_5 0 val_0
    -5 val_5 2 val_2
    -5 val_5 2 val_2
    -5 val_5 4 val_4
    -5 val_5 5 val_5
    -5 val_5 5 val_5
    -5 val_5 8 val_8
    -5 val_5 9 val_9
    -8 val_8 0 val_0
    +0 val_0 0 val_0
    +0 val_0 0 val_0
    +0 val_0 0 val_0
      0 val_0 0 val_0
      0 val_0 2 val_2
    +0 val_0 2 val_2
    +0 val_0 2 val_2
      0 val_0 4 val_4
      0 val_0 4 val_4
    +0 val_0 4 val_4
    +0 val_0 5 val_5
    +0 val_0 5 val_5
    +0 val_0 5 val_5
    +0 val_0 5 val_5
    +0 val_0 5 val_5
      0 val_0 5 val_5
      0 val_0 5 val_5
    +0 val_0 5 val_5
    +0 val_0 5 val_5
    +0 val_0 8 val_8
    +0 val_0 8 val_8
      0 val_0 8 val_8
      0 val_0 9 val_9
      0 val_0 9 val_9
    +0 val_0 9 val_9
    +2 val_2 0 val_0
      2 val_2 0 val_0
      2 val_2 0 val_0
    +2 val_2 2 val_2
      2 val_2 4 val_4
    +2 val_2 5 val_5
    +2 val_2 5 val_5
    +2 val_2 5 val_5
      2 val_2 8 val_8
      2 val_2 9 val_9
      4 val_4 0 val_0
    +4 val_4 0 val_0
    +4 val_4 0 val_0
    +4 val_4 2 val_2
    +4 val_4 4 val_4
    +4 val_4 5 val_5
    +4 val_4 5 val_5
    +4 val_4 5 val_5
      4 val_4 8 val_8
    +4 val_4 9 val_9
    +5 val_5 0 val_0
    +5 val_5 0 val_0
    +5 val_5 0 val_0
      5 val_5 0 val_0
      5 val_5 0 val_0
      5 val_5 0 val_0
    +5 val_5 0 val_0
    +5 val_5 0 val_0
    +5 val_5 0 val_0
    +5 val_5 2 val_2
      5 val_5 2 val_2
    +5 val_5 2 val_2
    +5 val_5 4 val_4
    +5 val_5 4 val_4
      5 val_5 4 val_4
      5 val_5 5 val_5
      5 val_5 5 val_5
      5 val_5 5 val_5
    -8 val_8 0 val_0
    -8 val_8 4 val_4
    -9 val_9 4 val_4
    -9 val_9 5 val_5
    -9 val_9 8 val_8
    -9 val_9 9 val_9
    -0 val_0 0 val_0
    -0 val_0 0 val_0
    -0 val_0 0 val_0
    -0 val_0 5 val_5
    -0 val_0 5 val_5
    -0 val_0 5 val_5
    -2 val_2 5 val_5
    -4 val_4 0 val_0
    -4 val_4 2 val_2
    -4 val_4 5 val_5
    -5 val_5 0 val_0
    -5 val_5 0 val_0
    -5 val_5 0 val_0
      5 val_5 5 val_5
      5 val_5 5 val_5
      5 val_5 5 val_5
    +5 val_5 5 val_5
    +5 val_5 5 val_5
    +5 val_5 5 val_5
    +5 val_5 8 val_8
    +5 val_5 8 val_8
    +5 val_5 8 val_8
    +5 val_5 9 val_9
    +5 val_5 9 val_9
    +5 val_5 9 val_9
    +8 val_8 0 val_0
    +8 val_8 0 val_0
    +8 val_8 0 val_0
      8 val_8 2 val_2
    +8 val_8 4 val_4
    +8 val_8 5 val_5
    +8 val_8 5 val_5
    +8 val_8 5 val_5
      8 val_8 8 val_8
      8 val_8 9 val_9
    +9 val_9 0 val_0
    +9 val_9 0 val_0
    +9 val_9 0 val_0
    +9 val_9 2 val_2
    +9 val_9 4 val_4
    +9 val_9 5 val_5
    +9 val_9 5 val_5
    +9 val_9 5 val_5
    +9 val_9 8 val_8
    +9 val_9 9 val_9

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/parquet_types.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/parquet_types.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/parquet_types.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/parquet_types.q.out Sat Feb 21 02:36:54 2015
    @@ -215,8 +215,8 @@ SELECT ctinyint,
        MAX(cint),
        MIN(csmallint),
        COUNT(cstring1),
    - AVG(cfloat),
    - STDDEV_POP(cdouble)
    + ROUND(AVG(cfloat), 5),
    + ROUND(STDDEV_POP(cdouble),5)
      FROM parquet_types
      GROUP BY ctinyint
      ORDER BY ctinyint
    @@ -229,17 +229,17 @@ SELECT ctinyint,
        MAX(cint),
        MIN(csmallint),
        COUNT(cstring1),
    - AVG(cfloat),
    - STDDEV_POP(cdouble)
    + ROUND(AVG(cfloat), 5),
    + ROUND(STDDEV_POP(cdouble),5)
      FROM parquet_types
      GROUP BY ctinyint
      ORDER BY ctinyint
      POSTHOOK: type: QUERY
      POSTHOOK: Input: default@parquet_types
      #### A masked pattern was here ####
    -1 121 1 8 1.1749999970197678 2.0621590627301285
    -2 119 1 7 1.2142857142857142 1.8
    -3 120 1 7 1.171428578240531 1.7999999999999996
    +1 121 1 8 1.175 2.06216
    +2 119 1 7 1.21429 1.8
    +3 120 1 7 1.17143 1.8
      PREHOOK: query: SELECT cfloat, count(*) FROM parquet_types GROUP BY cfloat ORDER BY cfloat
      PREHOOK: type: QUERY
      PREHOOK: Input: default@parquet_types

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/pcr.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/pcr.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/pcr.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/pcr.q.out Sat Feb 21 02:36:54 2015
    @@ -235,7 +235,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2
                        columns.types int:string:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -498,7 +498,7 @@ STAGE PLANS:
                        columns _col0,_col1
                        columns.types int:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -763,7 +763,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2
                        columns.types int:string:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -998,7 +998,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2
                        columns.types int:string:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1281,7 +1281,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2
                        columns.types int:string:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1574,7 +1574,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2
                        columns.types int:string:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1820,7 +1820,7 @@ STAGE PLANS:
                        columns _col0,_col1
                        columns.types int:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -2025,7 +2025,7 @@ STAGE PLANS:
                        columns _col0,_col1
                        columns.types int:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -2316,7 +2316,7 @@ STAGE PLANS:
                        columns _col0,_col1
                        columns.types int:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -2603,7 +2603,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2
                        columns.types int:string:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -2865,7 +2865,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2,_col3,_col4,_col5
                        columns.types int:string:string:int:string:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -3208,7 +3208,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2,_col3,_col4,_col5
                        columns.types int:string:string:int:string:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -3578,7 +3578,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2
                        columns.types int:string:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -3895,7 +3895,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2
                        columns.types int:string:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -5349,7 +5349,7 @@ STAGE PLANS:
                          columns _col0,_col1
                          columns.types string:string
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -5576,7 +5576,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2,_col3
                        columns.types string:string:string:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -5799,7 +5799,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2,_col3
                        columns.types string:string:string:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/ppd_join_filter.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/ppd_join_filter.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/ppd_join_filter.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/ppd_join_filter.q.out Sat Feb 21 02:36:54 2015
    @@ -375,7 +375,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2
                          columns.types string:double:double
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -792,7 +792,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2
                          columns.types string:double:double
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1209,7 +1209,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2
                          columns.types string:double:double
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1626,7 +1626,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2
                          columns.types string:double:double
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/ppd_union_view.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/ppd_union_view.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/ppd_union_view.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/ppd_union_view.q.out Sat Feb 21 02:36:54 2015
    @@ -353,7 +353,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2
                              columns.types string:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -471,8 +471,8 @@ STAGE PLANS:
                condition map:
                     Inner Join 0 to 1
                keys:
    - 0 _col0 (type: string), _col2 (type: string)
    - 1 _col1 (type: string), _col2 (type: string)
    + 0 _col0 (type: string)
    + 1 _col1 (type: string)
                outputColumnNames: _col1, _col3
                Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
                Select Operator
    @@ -524,7 +524,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2
                              columns.types string:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -549,7 +549,7 @@ STAGE PLANS:
                            columns _col0,_col1,_col2
                            columns.types string:string:string
                            escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                            serialization.format 1
                            serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/ppd_vc.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/ppd_vc.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/ppd_vc.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/ppd_vc.q.out Sat Feb 21 02:36:54 2015
    @@ -63,7 +63,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2,_col3
                              columns.types string:string:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -773,7 +773,7 @@ STAGE PLANS:
                        columns _col0,_col1,_col2,_col3,_col4
                        columns.types string:string:string:string:bigint
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/ppr_allchildsarenull.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/ppr_allchildsarenull.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/ppr_allchildsarenull.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/ppr_allchildsarenull.q.out Sat Feb 21 02:36:54 2015
    @@ -94,7 +94,7 @@ STAGE PLANS:
                              columns _col0,_col1
                              columns.types int:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -326,7 +326,7 @@ STAGE PLANS:
                              columns _col0,_col1
                              columns.types int:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
  • Prasanthj at Feb 21, 2015 at 2:37 am
    Modified: hive/branches/llap/ql/src/test/results/clientpositive/sample6.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/sample6.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/sample6.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/sample6.q.out Sat Feb 21 02:36:54 2015
    @@ -62,6 +62,7 @@ STAGE PLANS:
                  Filter Operator
                    isSamplingPred: true
                    predicate: (((hash(key) & 2147483647) % 4) = 0) (type: boolean)
    + sampleDesc: BUCKET 1 OUT OF 4
                    Statistics: Num rows: 500 Data size: 5301 Basic stats: COMPLETE Column stats: NONE
                    Select Operator
                      expressions: key (type: int), value (type: string)
    @@ -645,6 +646,7 @@ STAGE PLANS:
                  Filter Operator
                    isSamplingPred: true
                    predicate: (((hash(key) & 2147483647) % 4) = 3) (type: boolean)
    + sampleDesc: BUCKET 4 OUT OF 4
                    Statistics: Num rows: 500 Data size: 5301 Basic stats: COMPLETE Column stats: NONE
                    Select Operator
                      expressions: key (type: int), value (type: string)
    @@ -727,7 +729,7 @@ STAGE PLANS:
                        columns _col0,_col1
                        columns.types int:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1042,6 +1044,7 @@ STAGE PLANS:
                  Filter Operator
                    isSamplingPred: true
                    predicate: (((hash(key) & 2147483647) % 2) = 0) (type: boolean)
    + sampleDesc: BUCKET 1 OUT OF 2
                    Statistics: Num rows: 500 Data size: 5301 Basic stats: COMPLETE Column stats: NONE
                    Select Operator
                      expressions: key (type: int), value (type: string)
    @@ -1124,7 +1127,7 @@ STAGE PLANS:
                        columns _col0,_col1
                        columns.types int:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1775,7 +1778,7 @@ STAGE PLANS:
                        columns _col0,_col1
                        columns.types int:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -2269,7 +2272,7 @@ STAGE PLANS:
                        columns _col0,_col1
                        columns.types int:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -2667,6 +2670,7 @@ STAGE PLANS:
                  Filter Operator
                    isSamplingPred: true
                    predicate: (((hash(key) & 2147483647) % 2) = 0) (type: boolean)
    + sampleDesc: BUCKET 1 OUT OF 2
                    Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
                    Select Operator
                      expressions: key (type: int), value (type: string)
    @@ -2796,7 +2800,7 @@ STAGE PLANS:
                        columns _col0,_col1
                        columns.types int:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -2994,6 +2998,7 @@ STAGE PLANS:
                  Filter Operator
                    isSamplingPred: true
                    predicate: (((hash(key) & 2147483647) % 4) = 1) (type: boolean)
    + sampleDesc: BUCKET 2 OUT OF 4
                    Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
                    Select Operator
                      expressions: key (type: int), value (type: string)
    @@ -3076,7 +3081,7 @@ STAGE PLANS:
                        columns _col0,_col1
                        columns.types int:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -3210,7 +3215,7 @@ STAGE PLANS:
                        columns _col0,_col1
                        columns.types int:string
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/sample8.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/sample8.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/sample8.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/sample8.q.out Sat Feb 21 02:36:54 2015
    @@ -346,7 +346,7 @@ STAGE PLANS:
                            columns _col0,_col1,_col2,_col3
                            columns.types string:string:string:string
                            escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                            serialization.format 1
                            serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/sample9.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/sample9.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/sample9.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/sample9.q.out Sat Feb 21 02:36:54 2015
    @@ -58,6 +58,7 @@ STAGE PLANS:
                  Filter Operator
                    isSamplingPred: true
                    predicate: (((hash(key) & 2147483647) % 2) = 0) (type: boolean)
    + sampleDesc: BUCKET 1 OUT OF 2
                    Statistics: Num rows: 500 Data size: 5301 Basic stats: COMPLETE Column stats: NONE
                    Select Operator
                      expressions: key (type: int), value (type: string)
    @@ -77,7 +78,7 @@ STAGE PLANS:
                              columns _col0,_col1
                              columns.types int:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/script_env_var1.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/script_env_var1.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/script_env_var1.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/script_env_var1.q.out Sat Feb 21 02:36:54 2015
    @@ -1,16 +1,16 @@
      PREHOOK: query: -- Verifies that script operator ID environment variables have unique values
      -- in each instance of the script operator.
      SELECT count(1) FROM
    -( SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM src LIMIT 1 UNION ALL
    - SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM src LIMIT 1 ) a GROUP BY key
    +( SELECT * FROM (SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM src LIMIT 1)x UNION ALL
    + SELECT * FROM (SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM src LIMIT 1)y ) a GROUP BY key
      PREHOOK: type: QUERY
      PREHOOK: Input: default@src
      #### A masked pattern was here ####
      POSTHOOK: query: -- Verifies that script operator ID environment variables have unique values
      -- in each instance of the script operator.
      SELECT count(1) FROM
    -( SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM src LIMIT 1 UNION ALL
    - SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM src LIMIT 1 ) a GROUP BY key
    +( SELECT * FROM (SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM src LIMIT 1)x UNION ALL
    + SELECT * FROM (SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM src LIMIT 1)y ) a GROUP BY key
      POSTHOOK: type: QUERY
      POSTHOOK: Input: default@src
      #### A masked pattern was here ####

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/script_env_var2.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/script_env_var2.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/script_env_var2.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/script_env_var2.q.out Sat Feb 21 02:36:54 2015
    @@ -1,14 +1,14 @@
      PREHOOK: query: -- Same test as script_env_var1, but test setting the variable name
      SELECT count(1) FROM
    -( SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1 UNION ALL
    - SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1 ) a GROUP BY key
    +( SELECT * FROM (SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1)a UNION ALL
    + SELECT * FROM (SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1)b ) a GROUP BY key
      PREHOOK: type: QUERY
      PREHOOK: Input: default@src
      #### A masked pattern was here ####
      POSTHOOK: query: -- Same test as script_env_var1, but test setting the variable name
      SELECT count(1) FROM
    -( SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1 UNION ALL
    - SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1 ) a GROUP BY key
    +( SELECT * FROM (SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1)a UNION ALL
    + SELECT * FROM (SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1)b ) a GROUP BY key
      POSTHOOK: type: QUERY
      POSTHOOK: Input: default@src
      #### A masked pattern was here ####

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/serde_user_properties.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/serde_user_properties.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/serde_user_properties.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/serde_user_properties.q.out Sat Feb 21 02:36:54 2015
    @@ -140,7 +140,7 @@ STAGE PLANS:
                            columns _col0
                            columns.types string
                            escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                            serialization.format 1
                            serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -310,7 +310,7 @@ STAGE PLANS:
                            columns _col0
                            columns.types string
                            escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                            serialization.format 1
                            serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -488,7 +488,7 @@ STAGE PLANS:
                            columns _col0
                            columns.types string
                            escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                            serialization.format 1
                            serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/show_functions.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/show_functions.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/show_functions.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/show_functions.q.out Sat Feb 21 02:36:54 2015
    @@ -105,6 +105,7 @@ lcase
      lead
      least
      length
    +levenshtein
      like
      ln
      locate
    @@ -124,6 +125,7 @@ minute
      month
      named_struct
      negative
    +next_day
      ngrams
      noop
      noopstreaming

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/smb_mapjoin_11.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/smb_mapjoin_11.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/smb_mapjoin_11.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/smb_mapjoin_11.q.out Sat Feb 21 02:36:54 2015
    @@ -1863,6 +1863,7 @@ STAGE PLANS:
                  Filter Operator
                    isSamplingPred: true
                    predicate: (((hash(key) & 2147483647) % 16) = 1) (type: boolean)
    + sampleDesc: BUCKET 2 OUT OF 16
                    Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
                    Select Operator
                      expressions: key (type: int), value (type: string), ds (type: string)
    @@ -1882,7 +1883,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2
                              columns.types int:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1987,6 +1988,7 @@ STAGE PLANS:
                  Filter Operator
                    isSamplingPred: true
                    predicate: (((hash(key) & 2147483647) % 16) = 1) (type: boolean)
    + sampleDesc: BUCKET 2 OUT OF 16
                    Statistics: Num rows: 514 Data size: 5484 Basic stats: COMPLETE Column stats: NONE
                    Select Operator
                      expressions: key (type: int), value (type: string), ds (type: string)
    @@ -2006,7 +2008,7 @@ STAGE PLANS:
                              columns _col0,_col1,_col2
                              columns.types int:string:string
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/smb_mapjoin_13.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/smb_mapjoin_13.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/smb_mapjoin_13.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/smb_mapjoin_13.q.out Sat Feb 21 02:36:54 2015
    @@ -226,7 +226,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3
                          columns.types int:string:int:string
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -451,7 +451,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3
                          columns.types int:string:int:string
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/smb_mapjoin_15.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/smb_mapjoin_15.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/smb_mapjoin_15.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/smb_mapjoin_15.q.out Sat Feb 21 02:36:54 2015
    @@ -198,7 +198,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3
                          columns.types int:string:int:string
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -454,7 +454,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3,_col4,_col5
                          columns.types int:int:string:int:int:string
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -658,7 +658,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3,_col4,_col5
                          columns.types int:int:string:int:int:string
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -901,7 +901,7 @@ STAGE PLANS:
                          columns _col0,_col1,_col2,_col3,_col4,_col5
                          columns.types int:int:string:int:int:string
                          escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                          serialization.format 1
                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out Sat Feb 21 02:36:54 2015
    @@ -215,7 +215,7 @@ STAGE PLANS:
                        columns _col0
                        columns.types bigint
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out Sat Feb 21 02:36:54 2015
    @@ -299,7 +299,7 @@ STAGE PLANS:
                        columns _col0
                        columns.types bigint
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out Sat Feb 21 02:36:54 2015
    @@ -419,7 +419,7 @@ STAGE PLANS:
                        columns _col0
                        columns.types bigint
                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                        serialization.format 1
                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_join_reordering_values.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_join_reordering_values.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_join_reordering_values.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_join_reordering_values.q.out Sat Feb 21 02:36:54 2015
    @@ -601,7 +601,7 @@ STAGE PLANS:
                                  columns _col0,_col1
                                  columns.types string:int
                                  escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                  serialization.format 1
                                  serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_1.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_1.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_1.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_1.q.out Sat Feb 21 02:36:54 2015
    @@ -305,7 +305,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -536,7 +536,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -858,7 +858,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_11.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_11.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_11.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_11.q.out Sat Feb 21 02:36:54 2015
    @@ -382,7 +382,7 @@ STAGE PLANS:
                                columns _col0
                                columns.types bigint
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -713,7 +713,7 @@ STAGE PLANS:
                                columns _col0
                                columns.types bigint
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1043,7 +1043,7 @@ STAGE PLANS:
                                columns _col0
                                columns.types bigint
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1515,7 +1515,7 @@ STAGE PLANS:
                                columns _col0
                                columns.types bigint
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_12.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_12.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_12.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_12.q.out Sat Feb 21 02:36:54 2015
    @@ -610,7 +610,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_2.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_2.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_2.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_2.q.out Sat Feb 21 02:36:54 2015
    @@ -285,7 +285,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -609,7 +609,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_3.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_3.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_3.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_3.q.out Sat Feb 21 02:36:54 2015
    @@ -236,7 +236,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -418,7 +418,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -740,7 +740,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out Sat Feb 21 02:36:54 2015
    @@ -252,7 +252,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -434,7 +434,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -756,7 +756,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out Sat Feb 21 02:36:54 2015
    @@ -213,7 +213,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -385,7 +385,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -644,7 +644,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out Sat Feb 21 02:36:54 2015
    @@ -318,7 +318,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -551,7 +551,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -924,7 +924,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_8.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_8.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_8.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_8.q.out Sat Feb 21 02:36:54 2015
    @@ -318,7 +318,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -551,7 +551,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -926,7 +926,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/bucket_map_join_1.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/bucket_map_join_1.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/bucket_map_join_1.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/bucket_map_join_1.q.out Sat Feb 21 02:36:54 2015
    @@ -298,7 +298,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/bucket_map_join_2.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/bucket_map_join_2.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/bucket_map_join_2.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/bucket_map_join_2.q.out Sat Feb 21 02:36:54 2015
    @@ -298,7 +298,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out Sat Feb 21 02:36:54 2015
    @@ -348,7 +348,7 @@ STAGE PLANS:
                                        columns _col0,_col1,_col2,_col3
                                        columns.types int:string:string:string
                                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                        serialization.format 1
                                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -774,7 +774,7 @@ STAGE PLANS:
                                        columns _col0,_col1,_col2,_col3
                                        columns.types int:string:string:string
                                        escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                        serialization.format 1
                                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out Sat Feb 21 02:36:54 2015
    @@ -480,7 +480,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out Sat Feb 21 02:36:54 2015
    @@ -501,7 +501,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -909,7 +909,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out Sat Feb 21 02:36:54 2015
    @@ -363,7 +363,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -653,7 +653,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin13.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin13.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin13.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin13.q.out Sat Feb 21 02:36:54 2015
    @@ -360,7 +360,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -657,7 +657,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -953,7 +953,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1251,7 +1251,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out Sat Feb 21 02:36:54 2015
    @@ -347,7 +347,7 @@ STAGE PLANS:
                                columns _col0,_col1
                                columns.types int:string
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out Sat Feb 21 02:36:54 2015
    @@ -329,7 +329,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -639,7 +639,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out Sat Feb 21 02:36:54 2015
    @@ -326,7 +326,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -658,7 +658,7 @@ STAGE PLANS:
                              columns _col0
                              columns.types bigint
                              escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                              serialization.format 1
                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out Sat Feb 21 02:36:54 2015
    @@ -320,7 +320,7 @@ STAGE PLANS:
                                      columns _col0,_col1,_col2,_col3
                                      columns.types string:string:string:string
                                      escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                      serialization.format 1
                                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -566,7 +566,7 @@ STAGE PLANS:
                                      columns _col0,_col1,_col2,_col3
                                      columns.types string:string:string:string
                                      escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                      serialization.format 1
                                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -803,7 +803,7 @@ STAGE PLANS:
                                      columns _col0,_col1,_col2,_col3
                                      columns.types string:string:string:string
                                      escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                      serialization.format 1
                                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1038,7 +1038,7 @@ STAGE PLANS:
                                      columns _col0,_col1,_col2,_col3
                                      columns.types string:string:string:string
                                      escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                      serialization.format 1
                                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1273,7 +1273,7 @@ STAGE PLANS:
                                      columns _col0,_col1,_col2,_col3
                                      columns.types string:string:string:string
                                      escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                      serialization.format 1
                                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1508,7 +1508,7 @@ STAGE PLANS:
                                      columns _col0,_col1,_col2,_col3
                                      columns.types string:string:string:string
                                      escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                      serialization.format 1
                                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1743,7 +1743,7 @@ STAGE PLANS:
                                      columns _col0,_col1,_col2,_col3
                                      columns.types string:string:string:string
                                      escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                      serialization.format 1
                                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -1978,7 +1978,7 @@ STAGE PLANS:
                                      columns _col0,_col1,_col2,_col3
                                      columns.types string:string:string:string
                                      escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                      serialization.format 1
                                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
    @@ -2213,7 +2213,7 @@ STAGE PLANS:
                                      columns _col0,_col1,_col2,_col3
                                      columns.types string:string:string:string
                                      escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                      serialization.format 1
                                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/filter_join_breaktask.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/filter_join_breaktask.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/filter_join_breaktask.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/filter_join_breaktask.q.out Sat Feb 21 02:36:54 2015
    @@ -396,7 +396,7 @@ STAGE PLANS:
                                columns _col0,_col1
                                columns.types int:string
                                escape.delim \
    - hive.serialization.extend.nesting.levels true
    + hive.serialization.extend.additional.nesting.levels true
                                serialization.format 1
                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/groupby3_map_skew.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/groupby3_map_skew.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/groupby3_map_skew.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/groupby3_map_skew.q.out Sat Feb 21 02:36:54 2015
    @@ -147,12 +147,14 @@ POSTHOOK: Lineage: dest1.c6 EXPRESSION [
      POSTHOOK: Lineage: dest1.c7 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
      POSTHOOK: Lineage: dest1.c8 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
      POSTHOOK: Lineage: dest1.c9 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
    -PREHOOK: query: SELECT c1, c2, c3, c4, c5, c6, c7, ROUND(c8, 5), ROUND(c9, 5) FROM dest1
    +PREHOOK: query: SELECT ROUND(c1, 1), ROUND(c2, 3), ROUND(c3, 5), ROUND(c4, 1), ROUND(c5, 1), ROUND(c6, 5),
    +ROUND(c7,5), ROUND(c8, 5), ROUND(c9, 5) FROM dest1
      PREHOOK: type: QUERY
      PREHOOK: Input: default@dest1
      #### A masked pattern was here ####
    -POSTHOOK: query: SELECT c1, c2, c3, c4, c5, c6, c7, ROUND(c8, 5), ROUND(c9, 5) FROM dest1
    +POSTHOOK: query: SELECT ROUND(c1, 1), ROUND(c2, 3), ROUND(c3, 5), ROUND(c4, 1), ROUND(c5, 1), ROUND(c6, 5),
    +ROUND(c7,5), ROUND(c8, 5), ROUND(c9, 5) FROM dest1
      POSTHOOK: type: QUERY
      POSTHOOK: Input: default@dest1
      #### A masked pattern was here ####
    -130091.0 260.182 256.10355987055016 98.0 0.0 142.9268095075238 143.06995106518906 20428.07288 20469.0109
    +130091.0 260.182 256.10356 98.0 0.0 142.92681 143.06995 20428.07288 20469.0109

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/groupby7_noskew_multi_single_reducer.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/groupby7_noskew_multi_single_reducer.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/groupby7_noskew_multi_single_reducer.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/groupby7_noskew_multi_single_reducer.q.out Sat Feb 21 02:36:54 2015
    @@ -20,13 +20,13 @@ POSTHOOK: Output: database:default
      POSTHOOK: Output: default@DEST2
      PREHOOK: query: EXPLAIN
      FROM SRC
    -INSERT OVERWRITE TABLE DEST1 SELECT SRC.key, sum(SUBSTR(SRC.value,5)) GROUP BY SRC.key limit 10
    -INSERT OVERWRITE TABLE DEST2 SELECT SRC.key, sum(SUBSTR(SRC.value,5)) GROUP BY SRC.key limit 10
    +INSERT OVERWRITE TABLE DEST1 SELECT SRC.key, sum(SUBSTR(SRC.value,5)) GROUP BY SRC.key ORDER BY SRC.key limit 10
    +INSERT OVERWRITE TABLE DEST2 SELECT SRC.key, sum(SUBSTR(SRC.value,5)) GROUP BY SRC.key ORDER BY SRC.key limit 10
      PREHOOK: type: QUERY
      POSTHOOK: query: EXPLAIN
      FROM SRC
    -INSERT OVERWRITE TABLE DEST1 SELECT SRC.key, sum(SUBSTR(SRC.value,5)) GROUP BY SRC.key limit 10
    -INSERT OVERWRITE TABLE DEST2 SELECT SRC.key, sum(SUBSTR(SRC.value,5)) GROUP BY SRC.key limit 10
    +INSERT OVERWRITE TABLE DEST1 SELECT SRC.key, sum(SUBSTR(SRC.value,5)) GROUP BY SRC.key ORDER BY SRC.key limit 10
    +INSERT OVERWRITE TABLE DEST2 SELECT SRC.key, sum(SUBSTR(SRC.value,5)) GROUP BY SRC.key ORDER BY SRC.key limit 10
      POSTHOOK: type: QUERY
      STAGE DEPENDENCIES:
        Stage-2 is a root stage
    @@ -41,8 +41,8 @@ STAGE PLANS:
            Edges:
              Reducer 5 <- Map 1 (SORT, 31)
              Reducer 6 <- Map 1 (SORT, 31)
    - Reducer 3 <- Reducer 5 (GROUP, 1)
    - Reducer 4 <- Reducer 6 (GROUP, 1)
    + Reducer 3 <- Reducer 5 (SORT, 1)
    + Reducer 4 <- Reducer 6 (SORT, 1)
      #### A masked pattern was here ####
            Vertices:
              Map 1
    @@ -63,9 +63,9 @@ STAGE PLANS:
              Reducer 3
                  Reduce Operator Tree:
                    Select Operator
    - expressions: VALUE._col0 (type: string), VALUE._col1 (type: double)
    + expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: double)
                      outputColumnNames: _col0, _col1
    - Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
                      Limit
                        Number of rows: 10
                        Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
    @@ -84,9 +84,9 @@ STAGE PLANS:
              Reducer 4
                  Reduce Operator Tree:
                    Select Operator
    - expressions: VALUE._col0 (type: string), VALUE._col1 (type: double)
    + expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: double)
                      outputColumnNames: _col0, _col1
    - Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
    + Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
                      Limit
                        Number of rows: 10
                        Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
    @@ -112,13 +112,11 @@ STAGE PLANS:
                        mode: complete
                        outputColumnNames: _col0, _col1
                        Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
    - Limit
    - Number of rows: 10
    - Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
    - Reduce Output Operator
    - sort order:
    - Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
    - value expressions: _col0 (type: string), _col1 (type: double)
    + Reduce Output Operator
    + key expressions: _col0 (type: string)
    + sort order: +
    + Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
    + value expressions: _col1 (type: double)
              Reducer 6
                  Reduce Operator Tree:
                    Forward
    @@ -129,13 +127,11 @@ STAGE PLANS:
                        mode: complete
                        outputColumnNames: _col0, _col1
                        Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
    - Limit
    - Number of rows: 10
    - Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
    - Reduce Output Operator
    - sort order:
    - Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
    - value expressions: _col0 (type: string), _col1 (type: double)
    + Reduce Output Operator
    + key expressions: _col0 (type: string)
    + sort order: +
    + Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
    + value expressions: _col1 (type: double)

        Stage: Stage-0
          Move Operator

    Modified: hive/branches/llap/ql/src/test/results/clientpositive/spark/groupby_multi_single_reducer3.q.out
    URL: http://svn.apache.org/viewvc/hive/branches/llap/ql/src/test/results/clientpositive/spark/groupby_multi_single_reducer3.q.out?rev=1661271&r1=1661270&r2=1661271&view=diff
    ==============================================================================
    --- hive/branches/llap/ql/src/test/results/clientpositive/spark/groupby_multi_single_reducer3.q.out (original)
    +++ hive/branches/llap/ql/src/test/results/clientpositive/spark/groupby_multi_single_reducer3.q.out Sat Feb 21 02:36