hive-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From xu...@apache.org
Subject svn commit: r1660293 [33/48] - in /hive/branches/spark: ./ accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/ accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/ accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/mr/ accumul...
Date Tue, 17 Feb 2015 06:49:34 GMT
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/sample2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/sample2.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/sample2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/sample2.q.out Tue Feb 17
06:49:27 2015
@@ -60,6 +60,7 @@ STAGE PLANS:
                   Filter Operator
                     isSamplingPred: true
                     predicate: (((hash(key) & 2147483647) % 2) = 0) (type: boolean)
+                    sampleDesc: BUCKET 1 OUT OF 2
                     Statistics: Num rows: 500 Data size: 5301 Basic stats: COMPLETE Column
stats: NONE
                     Select Operator
                       expressions: key (type: int), value (type: string)

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/sample4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/sample4.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/sample4.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/sample4.q.out Tue Feb 17
06:49:27 2015
@@ -62,6 +62,7 @@ STAGE PLANS:
                   Filter Operator
                     isSamplingPred: true
                     predicate: (((hash(key) & 2147483647) % 2) = 0) (type: boolean)
+                    sampleDesc: BUCKET 1 OUT OF 2
                     Statistics: Num rows: 500 Data size: 5301 Basic stats: COMPLETE Column
stats: NONE
                     Select Operator
                       expressions: key (type: int), value (type: string)

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/sample6.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/sample6.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/sample6.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/sample6.q.out Tue Feb 17
06:49:27 2015
@@ -60,6 +60,7 @@ STAGE PLANS:
                   Filter Operator
                     isSamplingPred: true
                     predicate: (((hash(key) & 2147483647) % 4) = 0) (type: boolean)
+                    sampleDesc: BUCKET 1 OUT OF 4
                     Statistics: Num rows: 500 Data size: 5301 Basic stats: COMPLETE Column
stats: NONE
                     Select Operator
                       expressions: key (type: int), value (type: string)
@@ -495,6 +496,7 @@ STAGE PLANS:
                   Filter Operator
                     isSamplingPred: true
                     predicate: (((hash(key) & 2147483647) % 4) = 3) (type: boolean)
+                    sampleDesc: BUCKET 4 OUT OF 4
                     Statistics: Num rows: 500 Data size: 5301 Basic stats: COMPLETE Column
stats: NONE
                     Select Operator
                       expressions: key (type: int), value (type: string)
@@ -578,7 +580,7 @@ STAGE PLANS:
                         columns _col0,_col1
                         columns.types int:string
                         escape.delim \
-                        hive.serialization.extend.nesting.levels true
+                        hive.serialization.extend.additional.nesting.levels true
                         serialization.format 1
                         serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                       serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -898,6 +900,7 @@ STAGE PLANS:
                   Filter Operator
                     isSamplingPred: true
                     predicate: (((hash(key) & 2147483647) % 2) = 0) (type: boolean)
+                    sampleDesc: BUCKET 1 OUT OF 2
                     Statistics: Num rows: 500 Data size: 5301 Basic stats: COMPLETE Column
stats: NONE
                     Select Operator
                       expressions: key (type: int), value (type: string)
@@ -981,7 +984,7 @@ STAGE PLANS:
                         columns _col0,_col1
                         columns.types int:string
                         escape.delim \
-                        hive.serialization.extend.nesting.levels true
+                        hive.serialization.extend.additional.nesting.levels true
                         serialization.format 1
                         serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                       serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1638,7 +1641,7 @@ STAGE PLANS:
                         columns _col0,_col1
                         columns.types int:string
                         escape.delim \
-                        hive.serialization.extend.nesting.levels true
+                        hive.serialization.extend.additional.nesting.levels true
                         serialization.format 1
                         serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                       serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -2138,7 +2141,7 @@ STAGE PLANS:
                         columns _col0,_col1
                         columns.types int:string
                         escape.delim \
-                        hive.serialization.extend.nesting.levels true
+                        hive.serialization.extend.additional.nesting.levels true
                         serialization.format 1
                         serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                       serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -2541,6 +2544,7 @@ STAGE PLANS:
                   Filter Operator
                     isSamplingPred: true
                     predicate: (((hash(key) & 2147483647) % 2) = 0) (type: boolean)
+                    sampleDesc: BUCKET 1 OUT OF 2
                     Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column
stats: NONE
                     Select Operator
                       expressions: key (type: int), value (type: string)
@@ -2671,7 +2675,7 @@ STAGE PLANS:
                         columns _col0,_col1
                         columns.types int:string
                         escape.delim \
-                        hive.serialization.extend.nesting.levels true
+                        hive.serialization.extend.additional.nesting.levels true
                         serialization.format 1
                         serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                       serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -2874,6 +2878,7 @@ STAGE PLANS:
                   Filter Operator
                     isSamplingPred: true
                     predicate: (((hash(key) & 2147483647) % 4) = 1) (type: boolean)
+                    sampleDesc: BUCKET 2 OUT OF 4
                     Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column
stats: NONE
                     Select Operator
                       expressions: key (type: int), value (type: string)
@@ -2957,7 +2962,7 @@ STAGE PLANS:
                         columns _col0,_col1
                         columns.types int:string
                         escape.delim \
-                        hive.serialization.extend.nesting.levels true
+                        hive.serialization.extend.additional.nesting.levels true
                         serialization.format 1
                         serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                       serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -3097,7 +3102,7 @@ STAGE PLANS:
                         columns _col0,_col1
                         columns.types int:string
                         escape.delim \
-                        hive.serialization.extend.nesting.levels true
+                        hive.serialization.extend.additional.nesting.levels true
                         serialization.format 1
                         serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                       serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/sample8.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/sample8.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/sample8.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/sample8.q.out Tue Feb 17
06:49:27 2015
@@ -405,7 +405,7 @@ STAGE PLANS:
                             columns _col0,_col1,_col2,_col3
                             columns.types string:string:string:string
                             escape.delim \
-                            hive.serialization.extend.nesting.levels true
+                            hive.serialization.extend.additional.nesting.levels true
                             serialization.format 1
                             serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                           serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/sample9.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/sample9.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/sample9.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/sample9.q.out Tue Feb 17
06:49:27 2015
@@ -58,6 +58,7 @@ STAGE PLANS:
           Filter Operator
             isSamplingPred: true
             predicate: (((hash(key) & 2147483647) % 2) = 0) (type: boolean)
+            sampleDesc: BUCKET 1 OUT OF 2
             Statistics: Num rows: 500 Data size: 5301 Basic stats: COMPLETE Column stats:
NONE
             Select Operator
               expressions: key (type: int), value (type: string)

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/script_env_var1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/script_env_var1.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/script_env_var1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/script_env_var1.q.out Tue
Feb 17 06:49:27 2015
@@ -1,16 +1,16 @@
 PREHOOK: query: -- Verifies that script operator ID environment variables have unique values
 -- in each instance of the script operator.
 SELECT count(1) FROM
-( SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM src LIMIT 1 UNION
ALL
-  SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM src LIMIT 1 )
a GROUP BY key
+( SELECT * FROM (SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM
src LIMIT 1)x UNION ALL
+  SELECT * FROM (SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM
src LIMIT 1)y ) a GROUP BY key
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
 #### A masked pattern was here ####
 POSTHOOK: query: -- Verifies that script operator ID environment variables have unique values
 -- in each instance of the script operator.
 SELECT count(1) FROM
-( SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM src LIMIT 1 UNION
ALL
-  SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM src LIMIT 1 )
a GROUP BY key
+( SELECT * FROM (SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM
src LIMIT 1)x UNION ALL
+  SELECT * FROM (SELECT TRANSFORM('echo $HIVE_SCRIPT_OPERATOR_ID') USING 'sh' AS key FROM
src LIMIT 1)y ) a GROUP BY key
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
 #### A masked pattern was here ####

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/script_env_var2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/script_env_var2.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/script_env_var2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/script_env_var2.q.out Tue
Feb 17 06:49:27 2015
@@ -1,14 +1,14 @@
 PREHOOK: query: -- Same test as script_env_var1, but test setting the variable name
 SELECT count(1) FROM
-( SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1 UNION ALL
-  SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1 ) a GROUP BY key
+( SELECT * FROM (SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1)a UNION
ALL
+  SELECT * FROM (SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1)b ) a
GROUP BY key
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
 #### A masked pattern was here ####
 POSTHOOK: query: -- Same test as script_env_var1, but test setting the variable name
 SELECT count(1) FROM
-( SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1 UNION ALL
-  SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1 ) a GROUP BY key
+( SELECT * FROM (SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1)a UNION
ALL
+  SELECT * FROM (SELECT TRANSFORM('echo $MY_ID') USING 'sh' AS key FROM src LIMIT 1)b ) a
GROUP BY key
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
 #### A masked pattern was here ####

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_11.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_11.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_11.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_11.q.out Tue
Feb 17 06:49:27 2015
@@ -1916,6 +1916,7 @@ STAGE PLANS:
           Filter Operator
             isSamplingPred: true
             predicate: (((hash(key) & 2147483647) % 16) = 1) (type: boolean)
+            sampleDesc: BUCKET 2 OUT OF 16
             Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats:
NONE
             Select Operator
               expressions: key (type: int), value (type: string), ds (type: string)
@@ -2007,6 +2008,7 @@ STAGE PLANS:
           Filter Operator
             isSamplingPred: true
             predicate: (((hash(key) & 2147483647) % 16) = 1) (type: boolean)
+            sampleDesc: BUCKET 2 OUT OF 16
             Statistics: Num rows: 514 Data size: 5484 Basic stats: COMPLETE Column stats:
NONE
             Select Operator
               expressions: key (type: int), value (type: string), ds (type: string)

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out Tue
Feb 17 06:49:27 2015
@@ -238,7 +238,7 @@ STAGE PLANS:
                           columns _col0,_col1,_col2,_col3
                           columns.types int:string:int:string
                           escape.delim \
-                          hive.serialization.extend.nesting.levels true
+                          hive.serialization.extend.additional.nesting.levels true
                           serialization.format 1
                           serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                         serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -524,7 +524,7 @@ STAGE PLANS:
                           columns _col0,_col1,_col2,_col3
                           columns.types int:string:int:string
                           escape.delim \
-                          hive.serialization.extend.nesting.levels true
+                          hive.serialization.extend.additional.nesting.levels true
                           serialization.format 1
                           serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                         serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_15.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_15.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_15.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_15.q.out Tue
Feb 17 06:49:27 2015
@@ -210,7 +210,7 @@ STAGE PLANS:
                           columns _col0,_col1,_col2,_col3
                           columns.types int:string:int:string
                           escape.delim \
-                          hive.serialization.extend.nesting.levels true
+                          hive.serialization.extend.additional.nesting.levels true
                           serialization.format 1
                           serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                         serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -478,7 +478,7 @@ STAGE PLANS:
                           columns _col0,_col1,_col2,_col3,_col4,_col5
                           columns.types int:int:string:int:int:string
                           escape.delim \
-                          hive.serialization.extend.nesting.levels true
+                          hive.serialization.extend.additional.nesting.levels true
                           serialization.format 1
                           serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                         serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -694,7 +694,7 @@ STAGE PLANS:
                           columns _col0,_col1,_col2,_col3,_col4,_col5
                           columns.types int:int:string:int:int:string
                           escape.delim \
-                          hive.serialization.extend.nesting.levels true
+                          hive.serialization.extend.additional.nesting.levels true
                           serialization.format 1
                           serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                         serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1001,7 +1001,7 @@ STAGE PLANS:
                           columns _col0,_col1,_col2,_col3,_col4,_col5
                           columns.types int:int:string:int:int:string
                           escape.delim \
-                          hive.serialization.extend.nesting.levels true
+                          hive.serialization.extend.additional.nesting.levels true
                           serialization.format 1
                           serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                         serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_5.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_5.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_5.q.out
(original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_5.q.out
Tue Feb 17 06:49:27 2015
@@ -230,7 +230,7 @@ STAGE PLANS:
                           columns _col0
                           columns.types bigint
                           escape.delim \
-                          hive.serialization.extend.nesting.levels true
+                          hive.serialization.extend.additional.nesting.levels true
                           serialization.format 1
                           serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                         serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_6.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_6.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_6.q.out
(original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_6.q.out
Tue Feb 17 06:49:27 2015
@@ -320,7 +320,7 @@ STAGE PLANS:
                           columns _col0
                           columns.types bigint
                           escape.delim \
-                          hive.serialization.extend.nesting.levels true
+                          hive.serialization.extend.additional.nesting.levels true
                           serialization.format 1
                           serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                         serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_7.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_7.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_7.q.out
(original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_7.q.out
Tue Feb 17 06:49:27 2015
@@ -440,7 +440,7 @@ STAGE PLANS:
                           columns _col0
                           columns.types bigint
                           escape.delim \
-                          hive.serialization.extend.nesting.levels true
+                          hive.serialization.extend.additional.nesting.levels true
                           serialization.format 1
                           serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                         serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_in.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_in.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_in.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_in.q.out Tue Feb
17 06:49:27 2015
@@ -314,6 +314,25 @@ STAGE PLANS:
                 outputColumnNames: _col2, _col5
                 Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats:
NONE
                 PTF Operator
+                  Function definitions:
+                      Input definition
+                        input alias: ptf_0
+                        output shape: _col2: string, _col5: int
+                        type: WINDOWING
+                      Windowing table definition
+                        input alias: ptf_1
+                        name: windowingtablefunction
+                        order by: _col5
+                        partition by: _col2
+                        raw input shape:
+                        window functions:
+                            window function definition
+                              alias: _wcol0
+                              arguments: _col5
+                              name: rank
+                              window function: GenericUDAFRankEvaluator
+                              window frame: PRECEDING(MAX)~FOLLOWING(MAX)
+                              isPivotResult: true
                   Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats:
NONE
                   Filter Operator
                     predicate: (_wcol0 <= 2) (type: boolean)
@@ -463,6 +482,25 @@ STAGE PLANS:
                 outputColumnNames: _col2, _col5
                 Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats:
NONE
                 PTF Operator
+                  Function definitions:
+                      Input definition
+                        input alias: ptf_0
+                        output shape: _col2: string, _col5: int
+                        type: WINDOWING
+                      Windowing table definition
+                        input alias: ptf_1
+                        name: windowingtablefunction
+                        order by: _col5
+                        partition by: _col2
+                        raw input shape:
+                        window functions:
+                            window function definition
+                              alias: _wcol0
+                              arguments: _col5
+                              name: rank
+                              window function: GenericUDAFRankEvaluator
+                              window frame: PRECEDING(MAX)~FOLLOWING(MAX)
+                              isPivotResult: true
                   Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats:
NONE
                   Filter Operator
                     predicate: ((_wcol0 <= 2) and _col2 is not null) (type: boolean)

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/transform_ppr1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/transform_ppr1.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/transform_ppr1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/transform_ppr1.q.out Tue
Feb 17 06:49:27 2015
@@ -354,7 +354,7 @@ STAGE PLANS:
                         columns _col0,_col1
                         columns.types string:string
                         escape.delim \
-                        hive.serialization.extend.nesting.levels true
+                        hive.serialization.extend.additional.nesting.levels true
                         serialization.format 1
                         serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                       serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/transform_ppr2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/transform_ppr2.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/transform_ppr2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/transform_ppr2.q.out Tue
Feb 17 06:49:27 2015
@@ -262,7 +262,7 @@ STAGE PLANS:
                         columns _col0,_col1
                         columns.types string:string
                         escape.delim \
-                        hive.serialization.extend.nesting.levels true
+                        hive.serialization.extend.additional.nesting.levels true
                         serialization.format 1
                         serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                       serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union3.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/union3.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union3.q.out Tue Feb 17 06:49:27
2015
@@ -5,17 +5,16 @@ SELECT *
 FROM (
   SELECT 1 AS id
   FROM (SELECT * FROM src LIMIT 1) s1
-  CLUSTER BY id
   UNION ALL
   SELECT 2 AS id
   FROM (SELECT * FROM src LIMIT 1) s1
-  CLUSTER BY id
   UNION ALL
   SELECT 3 AS id
   FROM (SELECT * FROM src LIMIT 1) s2
   UNION ALL
   SELECT 4 AS id
   FROM (SELECT * FROM src LIMIT 1) s2
+  CLUSTER BY id
 ) a
 PREHOOK: type: QUERY
 POSTHOOK: query: -- SORT_QUERY_RESULTS
@@ -25,17 +24,16 @@ SELECT *
 FROM (
   SELECT 1 AS id
   FROM (SELECT * FROM src LIMIT 1) s1
-  CLUSTER BY id
   UNION ALL
   SELECT 2 AS id
   FROM (SELECT * FROM src LIMIT 1) s1
-  CLUSTER BY id
   UNION ALL
   SELECT 3 AS id
   FROM (SELECT * FROM src LIMIT 1) s2
   UNION ALL
   SELECT 4 AS id
   FROM (SELECT * FROM src LIMIT 1) s2
+  CLUSTER BY id
 ) a
 POSTHOOK: type: QUERY
 STAGE DEPENDENCIES:
@@ -46,12 +44,11 @@ STAGE PLANS:
   Stage: Stage-1
     Spark
       Edges:
-        Reducer 10 <- Map 9 (GROUP, 1)
         Reducer 2 <- Map 1 (GROUP, 1)
         Reducer 5 <- Map 4 (GROUP, 1)
-        Reducer 8 <- Map 7 (GROUP, 1)
-        Reducer 3 <- Reducer 2 (SORT, 2)
-        Reducer 6 <- Reducer 5 (SORT, 2)
+        Reducer 7 <- Map 6 (GROUP, 1)
+        Reducer 9 <- Map 8 (GROUP, 1)
+        Reducer 3 <- Reducer 2 (SORT, 2), Reducer 5 (SORT, 2), Reducer 7 (SORT, 2), Reducer
9 (SORT, 2)
 #### A masked pattern was here ####
       Vertices:
         Map 1 
@@ -80,7 +77,7 @@ STAGE PLANS:
                       Reduce Output Operator
                         sort order: 
                         Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column
stats: COMPLETE
-        Map 7 
+        Map 6 
             Map Operator Tree:
                 TableScan
                   alias: src
@@ -93,7 +90,7 @@ STAGE PLANS:
                       Reduce Output Operator
                         sort order: 
                         Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column
stats: COMPLETE
-        Map 9 
+        Map 8 
             Map Operator Tree:
                 TableScan
                   alias: src
@@ -106,36 +103,26 @@ STAGE PLANS:
                       Reduce Output Operator
                         sort order: 
                         Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column
stats: COMPLETE
-        Reducer 10 
+        Reducer 2 
             Reduce Operator Tree:
               Limit
                 Number of rows: 1
                 Select Operator
-                  expressions: 4 (type: int)
+                  expressions: 1 (type: int)
                   outputColumnNames: _col0
-                  File Output Operator
-                    compressed: false
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-        Reducer 2 
-            Reduce Operator Tree:
-              Limit
-                Number of rows: 1
-                Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: COMPLETE
-                Reduce Output Operator
-                  key expressions: 1 (type: int)
-                  sort order: +
-                  Map-reduce partition columns: 1 (type: int)
-                  Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats:
COMPLETE
+                  Reduce Output Operator
+                    key expressions: _col0 (type: int)
+                    sort order: +
+                    Map-reduce partition columns: _col0 (type: int)
         Reducer 3 
             Reduce Operator Tree:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: int)
                 outputColumnNames: _col0
+                Statistics: Num rows: 4 Data size: 0 Basic stats: PARTIAL Column stats: COMPLETE
                 File Output Operator
                   compressed: false
+                  Statistics: Num rows: 4 Data size: 0 Basic stats: PARTIAL Column stats:
COMPLETE
                   table:
                       input format: org.apache.hadoop.mapred.TextInputFormat
                       output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -144,36 +131,35 @@ STAGE PLANS:
             Reduce Operator Tree:
               Limit
                 Number of rows: 1
-                Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: COMPLETE
-                Reduce Output Operator
-                  key expressions: 2 (type: int)
-                  sort order: +
-                  Map-reduce partition columns: 2 (type: int)
-                  Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats:
COMPLETE
-        Reducer 6 
-            Reduce Operator Tree:
-              Select Operator
-                expressions: KEY.reducesinkkey0 (type: int)
-                outputColumnNames: _col0
-                File Output Operator
-                  compressed: false
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-        Reducer 8 
+                Select Operator
+                  expressions: 2 (type: int)
+                  outputColumnNames: _col0
+                  Reduce Output Operator
+                    key expressions: _col0 (type: int)
+                    sort order: +
+                    Map-reduce partition columns: _col0 (type: int)
+        Reducer 7 
             Reduce Operator Tree:
               Limit
                 Number of rows: 1
                 Select Operator
                   expressions: 3 (type: int)
                   outputColumnNames: _col0
-                  File Output Operator
-                    compressed: false
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  Reduce Output Operator
+                    key expressions: _col0 (type: int)
+                    sort order: +
+                    Map-reduce partition columns: _col0 (type: int)
+        Reducer 9 
+            Reduce Operator Tree:
+              Limit
+                Number of rows: 1
+                Select Operator
+                  expressions: 4 (type: int)
+                  outputColumnNames: _col0
+                  Reduce Output Operator
+                    key expressions: _col0 (type: int)
+                    sort order: +
+                    Map-reduce partition columns: _col0 (type: int)
 
   Stage: Stage-0
     Fetch Operator
@@ -194,17 +180,16 @@ SELECT *
 FROM (
   SELECT 1 AS id
   FROM (SELECT * FROM src LIMIT 1) s1
-  CLUSTER BY id
   UNION ALL
   SELECT 2 AS id
   FROM (SELECT * FROM src LIMIT 1) s1
-  CLUSTER BY id
   UNION ALL
   SELECT 3 AS id
   FROM (SELECT * FROM src LIMIT 1) s2
   UNION ALL
   SELECT 4 AS id
   FROM (SELECT * FROM src LIMIT 1) s2
+  CLUSTER BY id
 ) a
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
@@ -214,31 +199,30 @@ SELECT *
 FROM (
   SELECT 1 AS id
   FROM (SELECT * FROM src LIMIT 1) s1
-  CLUSTER BY id
   UNION ALL
   SELECT 2 AS id
   FROM (SELECT * FROM src LIMIT 1) s1
-  CLUSTER BY id
   UNION ALL
   SELECT 3 AS id
   FROM (SELECT * FROM src LIMIT 1) s2
   UNION ALL
   SELECT 4 AS id
   FROM (SELECT * FROM src LIMIT 1) s2
+  CLUSTER BY id
 ) a
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
 POSTHOOK: Output: default@union_out
 POSTHOOK: Lineage: union_out.id EXPRESSION []
-PREHOOK: query: select * from union_out cluster by id
+PREHOOK: query: select * from union_out
 PREHOOK: type: QUERY
 PREHOOK: Input: default@union_out
 #### A masked pattern was here ####
-POSTHOOK: query: select * from union_out cluster by id
+POSTHOOK: query: select * from union_out
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@union_out
 #### A masked pattern was here ####
-4
 1
 2
 3
+4

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union_null.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union_null.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/union_null.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union_null.q.out Tue Feb
17 06:49:27 2015
@@ -1,34 +1,34 @@
 PREHOOK: query: -- SORT_BEFORE_DIFF
 
 -- HIVE-2901
-select x from (select value as x from src order by x limit 5 union all select NULL as x from
src limit 5) a
+select x from (select * from (select value as x from src order by x limit 5)a union all select
* from (select NULL as x from src limit 5)b )a
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
 #### A masked pattern was here ####
 POSTHOOK: query: -- SORT_BEFORE_DIFF
 
 -- HIVE-2901
-select x from (select value as x from src order by x limit 5 union all select NULL as x from
src limit 5) a
+select x from (select * from (select value as x from src order by x limit 5)a union all select
* from (select NULL as x from src limit 5)b )a
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
 #### A masked pattern was here ####
-NULL
-NULL
-NULL
-NULL
-NULL
 val_0
 val_0
 val_0
 val_10
 val_100
+NULL
+NULL
+NULL
+NULL
+NULL
 PREHOOK: query: -- HIVE-4837
-select * from (select null as N from src1 group by key UNION ALL select null as N from src1
group by key ) a
+select * from (select * from (select null as N from src1 group by key)a UNION ALL select
* from (select null as N from src1 group by key)b ) a
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src1
 #### A masked pattern was here ####
 POSTHOOK: query: -- HIVE-4837
-select * from (select null as N from src1 group by key UNION ALL select null as N from src1
group by key ) a
+select * from (select * from (select null as N from src1 group by key)a UNION ALL select
* from (select null as N from src1 group by key)b ) a
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src1
 #### A masked pattern was here ####

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union_ppr.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union_ppr.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/union_ppr.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union_ppr.q.out Tue Feb 17
06:49:27 2015
@@ -25,7 +25,7 @@ ABSTRACT SYNTAX TREE:
 TOK_QUERY
    TOK_FROM
       TOK_SUBQUERY
-         TOK_UNION
+         TOK_UNIONALL
             TOK_QUERY
                TOK_FROM
                   TOK_TABREF
@@ -374,7 +374,7 @@ STAGE PLANS:
                         columns _col0,_col1,_col2,_col3
                         columns.types string:string:string:string
                         escape.delim \
-                        hive.serialization.extend.nesting.levels true
+                        hive.serialization.extend.additional.nesting.levels true
                         serialization.format 1
                         serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                       serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union_remove_25.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union_remove_25.q.out?rev=1660293&r1=1660292&r2=1660293&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/union_remove_25.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union_remove_25.q.out Tue
Feb 17 06:49:27 2015
@@ -258,18 +258,18 @@ PREHOOK: query: explain
 insert overwrite table outputTbl2 partition(ds)
 SELECT *
 FROM (
-  SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500
+  select * from (SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500)a
   UNION ALL
-  SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500
+  select * from (SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500)b
 ) a
 PREHOOK: type: QUERY
 POSTHOOK: query: explain 
 insert overwrite table outputTbl2 partition(ds)
 SELECT *
 FROM (
-  SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500
+  select * from (SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500)a
   UNION ALL
-  SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500
+  select * from (SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500)b
 ) a
 POSTHOOK: type: QUERY
 STAGE DEPENDENCIES:
@@ -366,9 +366,9 @@ STAGE PLANS:
 PREHOOK: query: insert overwrite table outputTbl2 partition(ds)
 SELECT *
 FROM (
-  SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500
+  select * from (SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500)a
   UNION ALL
-  SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500
+  select * from (SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500)b
 ) a
 PREHOOK: type: QUERY
 PREHOOK: Input: default@srcpart
@@ -378,9 +378,9 @@ PREHOOK: Output: default@outputtbl2
 POSTHOOK: query: insert overwrite table outputTbl2 partition(ds)
 SELECT *
 FROM (
-  SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500
+  select * from (SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500)a
   UNION ALL
-  SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500
+  select * from (SELECT key, value, ds from srcpart where ds='2008-04-08' limit 500)b
 ) a
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@srcpart
@@ -440,17 +440,17 @@ Storage Desc Params:
 PREHOOK: query: explain insert overwrite table outputTbl3 partition(ds, hr)
 SELECT *
 FROM (
-  SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000
+  select * from (SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000)a
   UNION ALL
-  SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000
+  select * from (SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000)b
 ) a
 PREHOOK: type: QUERY
 POSTHOOK: query: explain insert overwrite table outputTbl3 partition(ds, hr)
 SELECT *
 FROM (
-  SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000
+  select * from (SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000)a
   UNION ALL
-  SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000
+  select * from (SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000)b
 ) a
 POSTHOOK: type: QUERY
 STAGE DEPENDENCIES:
@@ -548,9 +548,9 @@ STAGE PLANS:
 PREHOOK: query: insert overwrite table outputTbl3 partition(ds, hr)
 SELECT *
 FROM (
-  SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000
+  select * from (SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000)a
   UNION ALL
-  SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000
+  select * from (SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000)b
 ) a
 PREHOOK: type: QUERY
 PREHOOK: Input: default@srcpart
@@ -560,9 +560,9 @@ PREHOOK: Output: default@outputtbl3
 POSTHOOK: query: insert overwrite table outputTbl3 partition(ds, hr)
 SELECT *
 FROM (
-  SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000
+  select * from (SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000)a
   UNION ALL
-  SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000
+  select * from (SELECT key, value, ds, hr from srcpart where ds='2008-04-08' limit 1000)b
 ) a
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@srcpart



Mime
View raw message