hive-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From xu...@apache.org
Subject svn commit: r1669775 [23/35] - in /hive/branches/spark: ./ ant/src/org/apache/hadoop/hive/ant/ beeline/src/java/org/apache/hive/beeline/ common/src/java/org/apache/hadoop/hive/common/ common/src/java/org/apache/hadoop/hive/common/type/ common/src/java/...
Date Sat, 28 Mar 2015 14:03:49 GMT
Modified: hive/branches/spark/ql/src/test/results/clientpositive/columnstats_tbllvl.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/columnstats_tbllvl.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/columnstats_tbllvl.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/columnstats_tbllvl.q.out Sat Mar 28 14:03:43 2015
@@ -81,7 +81,7 @@ STAGE PLANS:
       Column Stats Desc:
           Columns: sourceIP, avgTimeOnSite, adRevenue
           Column Types: string, int, float
-          Table: uservisits_web_text_none
+          Table: default.uservisits_web_text_none
 
 PREHOOK: query: explain extended
 analyze table UserVisits_web_text_none compute statistics for columns sourceIP, avgTimeOnSite, adRevenue
@@ -204,7 +204,7 @@ STAGE PLANS:
       Column Stats Desc:
           Columns: sourceIP, avgTimeOnSite, adRevenue
           Column Types: string, int, float
-          Table: uservisits_web_text_none
+          Table: default.uservisits_web_text_none
           Is Table Level Stats: true
 
 PREHOOK: query: analyze table UserVisits_web_text_none compute statistics for columns sourceIP, avgTimeOnSite, adRevenue
@@ -216,10 +216,10 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@uservisits_web_text_none
 #### A masked pattern was here ####
 PREHOOK: query: explain 
-analyze table UserVisits_web_text_none compute statistics for columns
+analyze table default.UserVisits_web_text_none compute statistics for columns
 PREHOOK: type: QUERY
 POSTHOOK: query: explain 
-analyze table UserVisits_web_text_none compute statistics for columns
+analyze table default.UserVisits_web_text_none compute statistics for columns
 POSTHOOK: type: QUERY
 STAGE DEPENDENCIES:
   Stage-0 is a root stage
@@ -258,13 +258,13 @@ STAGE PLANS:
       Column Stats Desc:
           Columns: sourceip, desturl, visitdate, adrevenue, useragent, ccode, lcode, skeyword, avgtimeonsite
           Column Types: string, string, string, float, string, string, string, string, int
-          Table: uservisits_web_text_none
+          Table: default.uservisits_web_text_none
 
-PREHOOK: query: analyze table UserVisits_web_text_none compute statistics for columns
+PREHOOK: query: analyze table default.UserVisits_web_text_none compute statistics for columns
 PREHOOK: type: QUERY
 PREHOOK: Input: default@uservisits_web_text_none
 #### A masked pattern was here ####
-POSTHOOK: query: analyze table UserVisits_web_text_none compute statistics for columns
+POSTHOOK: query: analyze table default.UserVisits_web_text_none compute statistics for columns
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@uservisits_web_text_none
 #### A masked pattern was here ####
@@ -358,7 +358,7 @@ STAGE PLANS:
       Column Stats Desc:
           Columns: a, b, c, d, e
           Column Types: int, double, string, boolean, binary
-          Table: empty_tab
+          Table: default.empty_tab
 
 PREHOOK: query: analyze table empty_tab compute statistics for columns a,b,c,d,e
 PREHOOK: type: QUERY
@@ -368,3 +368,348 @@ POSTHOOK: query: analyze table empty_tab
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@empty_tab
 #### A masked pattern was here ####
+PREHOOK: query: create database if not exists dummydb
+PREHOOK: type: CREATEDATABASE
+PREHOOK: Output: database:dummydb
+POSTHOOK: query: create database if not exists dummydb
+POSTHOOK: type: CREATEDATABASE
+POSTHOOK: Output: database:dummydb
+PREHOOK: query: use dummydb
+PREHOOK: type: SWITCHDATABASE
+PREHOOK: Input: database:dummydb
+POSTHOOK: query: use dummydb
+POSTHOOK: type: SWITCHDATABASE
+POSTHOOK: Input: database:dummydb
+PREHOOK: query: analyze table default.UserVisits_web_text_none compute statistics for columns destURL
+PREHOOK: type: QUERY
+PREHOOK: Input: default@uservisits_web_text_none
+#### A masked pattern was here ####
+POSTHOOK: query: analyze table default.UserVisits_web_text_none compute statistics for columns destURL
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@uservisits_web_text_none
+#### A masked pattern was here ####
+PREHOOK: query: describe formatted default.UserVisits_web_text_none destURL
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@uservisits_web_text_none
+POSTHOOK: query: describe formatted default.UserVisits_web_text_none destURL
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@uservisits_web_text_none
+# col_name            	data_type           	min                 	max                 	num_nulls           	distinct_count      	avg_col_len         	max_col_len         	num_trues           	num_falses          	comment             
+	 	 	 	 	 	 	 	 	 	 
+destURL             	string              	                    	                    	0                   	56                  	48.945454545454545  	96                  	                    	                    	from deserializer   
+PREHOOK: query: CREATE TABLE UserVisits_in_dummy_db (
+  sourceIP string,
+  destURL string,
+  visitDate string,
+  adRevenue float,
+  userAgent string,
+  cCode string,
+  lCode string,
+  sKeyword string,
+  avgTimeOnSite int)
+row format delimited fields terminated by '|'  stored as textfile
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:dummydb
+PREHOOK: Output: dummydb@UserVisits_in_dummy_db
+POSTHOOK: query: CREATE TABLE UserVisits_in_dummy_db (
+  sourceIP string,
+  destURL string,
+  visitDate string,
+  adRevenue float,
+  userAgent string,
+  cCode string,
+  lCode string,
+  sKeyword string,
+  avgTimeOnSite int)
+row format delimited fields terminated by '|'  stored as textfile
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:dummydb
+POSTHOOK: Output: dummydb@UserVisits_in_dummy_db
+PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/UserVisits.dat" INTO TABLE UserVisits_in_dummy_db
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: dummydb@uservisits_in_dummy_db
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/UserVisits.dat" INTO TABLE UserVisits_in_dummy_db
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: dummydb@uservisits_in_dummy_db
+PREHOOK: query: use default
+PREHOOK: type: SWITCHDATABASE
+PREHOOK: Input: database:default
+POSTHOOK: query: use default
+POSTHOOK: type: SWITCHDATABASE
+POSTHOOK: Input: database:default
+PREHOOK: query: explain 
+analyze table dummydb.UserVisits_in_dummy_db compute statistics for columns sourceIP, avgTimeOnSite, adRevenue
+PREHOOK: type: QUERY
+POSTHOOK: query: explain 
+analyze table dummydb.UserVisits_in_dummy_db compute statistics for columns sourceIP, avgTimeOnSite, adRevenue
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-0 is a root stage
+  Stage-1 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-0
+    Map Reduce
+      Map Operator Tree:
+          TableScan
+            alias: uservisits_in_dummy_db
+            Select Operator
+              expressions: sourceip (type: string), avgtimeonsite (type: int), adrevenue (type: float)
+              outputColumnNames: sourceip, avgtimeonsite, adrevenue
+              Group By Operator
+                aggregations: compute_stats(sourceip, 16), compute_stats(avgtimeonsite, 16), compute_stats(adrevenue, 16)
+                mode: hash
+                outputColumnNames: _col0, _col1, _col2
+                Reduce Output Operator
+                  sort order: 
+                  value expressions: _col0 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>), _col1 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>), _col2 (type: struct<columntype:string,min:double,max:double,countnulls:bigint,bitvector:string,numbitvectors:int>)
+      Reduce Operator Tree:
+        Group By Operator
+          aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2)
+          mode: mergepartial
+          outputColumnNames: _col0, _col1, _col2
+          File Output Operator
+            compressed: false
+            table:
+                input format: org.apache.hadoop.mapred.TextInputFormat
+                output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+  Stage: Stage-1
+    Column Stats Work
+      Column Stats Desc:
+          Columns: sourceIP, avgTimeOnSite, adRevenue
+          Column Types: string, int, float
+          Table: dummydb.uservisits_in_dummy_db
+
+PREHOOK: query: explain extended
+analyze table dummydb.UserVisits_in_dummy_db compute statistics for columns sourceIP, avgTimeOnSite, adRevenue
+PREHOOK: type: QUERY
+POSTHOOK: query: explain extended
+analyze table dummydb.UserVisits_in_dummy_db compute statistics for columns sourceIP, avgTimeOnSite, adRevenue
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+  
+TOK_ANALYZE
+   TOK_TAB
+      TOK_TABNAME
+         dummydb
+         UserVisits_in_dummy_db
+   columns
+   TOK_TABCOLNAME
+      sourceIP
+      avgTimeOnSite
+      adRevenue
+
+
+STAGE DEPENDENCIES:
+  Stage-0 is a root stage
+  Stage-1 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-0
+    Map Reduce
+      Map Operator Tree:
+          TableScan
+            alias: uservisits_in_dummy_db
+            GatherStats: false
+            Select Operator
+              expressions: sourceip (type: string), avgtimeonsite (type: int), adrevenue (type: float)
+              outputColumnNames: sourceip, avgtimeonsite, adrevenue
+              Group By Operator
+                aggregations: compute_stats(sourceip, 16), compute_stats(avgtimeonsite, 16), compute_stats(adrevenue, 16)
+                mode: hash
+                outputColumnNames: _col0, _col1, _col2
+                Reduce Output Operator
+                  sort order: 
+                  tag: -1
+                  value expressions: _col0 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>), _col1 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>), _col2 (type: struct<columntype:string,min:double,max:double,countnulls:bigint,bitvector:string,numbitvectors:int>)
+                  auto parallelism: false
+      Path -> Alias:
+#### A masked pattern was here ####
+      Path -> Partition:
+#### A masked pattern was here ####
+          Partition
+            base file name: uservisits_in_dummy_db
+            input format: org.apache.hadoop.mapred.TextInputFormat
+            output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+            properties:
+              COLUMN_STATS_ACCURATE true
+              bucket_count -1
+              columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite
+              columns.comments 
+              columns.types string:string:string:float:string:string:string:string:int
+              field.delim |
+#### A masked pattern was here ####
+              name dummydb.uservisits_in_dummy_db
+              numFiles 1
+              serialization.ddl struct uservisits_in_dummy_db { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite}
+              serialization.format |
+              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+              totalSize 7060
+#### A masked pattern was here ####
+            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+          
+              input format: org.apache.hadoop.mapred.TextInputFormat
+              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              properties:
+                COLUMN_STATS_ACCURATE true
+                bucket_count -1
+                columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite
+                columns.comments 
+                columns.types string:string:string:float:string:string:string:string:int
+                field.delim |
+#### A masked pattern was here ####
+                name dummydb.uservisits_in_dummy_db
+                numFiles 1
+                serialization.ddl struct uservisits_in_dummy_db { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite}
+                serialization.format |
+                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                totalSize 7060
+#### A masked pattern was here ####
+              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+              name: dummydb.uservisits_in_dummy_db
+            name: dummydb.uservisits_in_dummy_db
+      Truncated Path -> Alias:
+        /dummydb.db/uservisits_in_dummy_db [uservisits_in_dummy_db]
+      Needs Tagging: false
+      Reduce Operator Tree:
+        Group By Operator
+          aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2)
+          mode: mergepartial
+          outputColumnNames: _col0, _col1, _col2
+          File Output Operator
+            compressed: false
+            GlobalTableId: 0
+#### A masked pattern was here ####
+            NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+            table:
+                input format: org.apache.hadoop.mapred.TextInputFormat
+                output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                properties:
+                  columns _col0,_col1,_col2
+                  columns.types struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint>:struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint>:struct<columntype:string,min:double,max:double,countnulls:bigint,numdistinctvalues:bigint>
+                  escape.delim \
+                  hive.serialization.extend.additional.nesting.levels true
+                  serialization.format 1
+                  serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+            TotalFiles: 1
+            GatherStats: false
+            MultiFileSpray: false
+
+  Stage: Stage-1
+    Column Stats Work
+      Column Stats Desc:
+          Columns: sourceIP, avgTimeOnSite, adRevenue
+          Column Types: string, int, float
+          Table: dummydb.uservisits_in_dummy_db
+          Is Table Level Stats: true
+
+PREHOOK: query: analyze table dummydb.UserVisits_in_dummy_db compute statistics for columns sourceIP, avgTimeOnSite, adRevenue
+PREHOOK: type: QUERY
+PREHOOK: Input: dummydb@uservisits_in_dummy_db
+#### A masked pattern was here ####
+POSTHOOK: query: analyze table dummydb.UserVisits_in_dummy_db compute statistics for columns sourceIP, avgTimeOnSite, adRevenue
+POSTHOOK: type: QUERY
+POSTHOOK: Input: dummydb@uservisits_in_dummy_db
+#### A masked pattern was here ####
+PREHOOK: query: explain 
+analyze table dummydb.UserVisits_in_dummy_db compute statistics for columns
+PREHOOK: type: QUERY
+POSTHOOK: query: explain 
+analyze table dummydb.UserVisits_in_dummy_db compute statistics for columns
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-0 is a root stage
+  Stage-1 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-0
+    Map Reduce
+      Map Operator Tree:
+          TableScan
+            alias: uservisits_in_dummy_db
+            Select Operator
+              expressions: sourceip (type: string), desturl (type: string), visitdate (type: string), adrevenue (type: float), useragent (type: string), ccode (type: string), lcode (type: string), skeyword (type: string), avgtimeonsite (type: int)
+              outputColumnNames: sourceip, desturl, visitdate, adrevenue, useragent, ccode, lcode, skeyword, avgtimeonsite
+              Group By Operator
+                aggregations: compute_stats(sourceip, 16), compute_stats(desturl, 16), compute_stats(visitdate, 16), compute_stats(adrevenue, 16), compute_stats(useragent, 16), compute_stats(ccode, 16), compute_stats(lcode, 16), compute_stats(skeyword, 16), compute_stats(avgtimeonsite, 16)
+                mode: hash
+                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8
+                Reduce Output Operator
+                  sort order: 
+                  value expressions: _col0 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>), _col1 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>), _col2 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>), _col3 (type: struct<columntype:string,min:double,max:double,countnulls:bigint,bitvector:string,numbitvectors:int>), _col4 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>), _col5 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>), _col6 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>)
 , _col7 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>), _col8 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>)
+      Reduce Operator Tree:
+        Group By Operator
+          aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3), compute_stats(VALUE._col4), compute_stats(VALUE._col5), compute_stats(VALUE._col6), compute_stats(VALUE._col7), compute_stats(VALUE._col8)
+          mode: mergepartial
+          outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8
+          File Output Operator
+            compressed: false
+            table:
+                input format: org.apache.hadoop.mapred.TextInputFormat
+                output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+  Stage: Stage-1
+    Column Stats Work
+      Column Stats Desc:
+          Columns: sourceip, desturl, visitdate, adrevenue, useragent, ccode, lcode, skeyword, avgtimeonsite
+          Column Types: string, string, string, float, string, string, string, string, int
+          Table: dummydb.uservisits_in_dummy_db
+
+PREHOOK: query: analyze table dummydb.UserVisits_in_dummy_db compute statistics for columns
+PREHOOK: type: QUERY
+PREHOOK: Input: dummydb@uservisits_in_dummy_db
+#### A masked pattern was here ####
+POSTHOOK: query: analyze table dummydb.UserVisits_in_dummy_db compute statistics for columns
+POSTHOOK: type: QUERY
+POSTHOOK: Input: dummydb@uservisits_in_dummy_db
+#### A masked pattern was here ####
+PREHOOK: query: describe formatted dummydb.UserVisits_in_dummy_db destURL
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: dummydb@uservisits_in_dummy_db
+POSTHOOK: query: describe formatted dummydb.UserVisits_in_dummy_db destURL
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: dummydb@uservisits_in_dummy_db
+# col_name            	data_type           	min                 	max                 	num_nulls           	distinct_count      	avg_col_len         	max_col_len         	num_trues           	num_falses          	comment             
+	 	 	 	 	 	 	 	 	 	 
+destURL             	string              	                    	                    	0                   	56                  	48.945454545454545  	96                  	                    	                    	from deserializer   
+PREHOOK: query: describe formatted dummydb.UserVisits_in_dummy_db adRevenue
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: dummydb@uservisits_in_dummy_db
+POSTHOOK: query: describe formatted dummydb.UserVisits_in_dummy_db adRevenue
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: dummydb@uservisits_in_dummy_db
+# col_name            	data_type           	min                 	max                 	num_nulls           	distinct_count      	avg_col_len         	max_col_len         	num_trues           	num_falses          	comment             
+	 	 	 	 	 	 	 	 	 	 
+adRevenue           	float               	13.099044799804688  	492.98870849609375  	0                   	58                  	                    	                    	                    	                    	from deserializer   
+PREHOOK: query: describe formatted dummydb.UserVisits_in_dummy_db avgTimeOnSite
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: dummydb@uservisits_in_dummy_db
+POSTHOOK: query: describe formatted dummydb.UserVisits_in_dummy_db avgTimeOnSite
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: dummydb@uservisits_in_dummy_db
+# col_name            	data_type           	min                 	max                 	num_nulls           	distinct_count      	avg_col_len         	max_col_len         	num_trues           	num_falses          	comment             
+	 	 	 	 	 	 	 	 	 	 
+avgTimeOnSite       	int                 	1                   	9                   	0                   	11                  	                    	                    	                    	                    	from deserializer   
+PREHOOK: query: drop table dummydb.UserVisits_in_dummy_db
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: dummydb@uservisits_in_dummy_db
+PREHOOK: Output: dummydb@uservisits_in_dummy_db
+POSTHOOK: query: drop table dummydb.UserVisits_in_dummy_db
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: dummydb@uservisits_in_dummy_db
+POSTHOOK: Output: dummydb@uservisits_in_dummy_db
+PREHOOK: query: drop database dummydb
+PREHOOK: type: DROPDATABASE
+PREHOOK: Input: database:dummydb
+PREHOOK: Output: database:dummydb
+POSTHOOK: query: drop database dummydb
+POSTHOOK: type: DROPDATABASE
+POSTHOOK: Input: database:dummydb
+POSTHOOK: Output: database:dummydb

Modified: hive/branches/spark/ql/src/test/results/clientpositive/decimal_10_0.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/decimal_10_0.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/decimal_10_0.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/decimal_10_0.q.out Sat Mar 28 14:03:43 2015
@@ -1,38 +1,38 @@
-PREHOOK: query: DROP TABLE IF EXISTS DECIMAL
+PREHOOK: query: DROP TABLE IF EXISTS `DECIMAL`
 PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE IF EXISTS DECIMAL
+POSTHOOK: query: DROP TABLE IF EXISTS `DECIMAL`
 POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE DECIMAL (dec decimal)
+PREHOOK: query: CREATE TABLE `DECIMAL` (dec decimal)
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:default
 PREHOOK: Output: default@DECIMAL
-POSTHOOK: query: CREATE TABLE DECIMAL (dec decimal)
+POSTHOOK: query: CREATE TABLE `DECIMAL` (dec decimal)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: database:default
 POSTHOOK: Output: default@DECIMAL
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/decimal_10_0.txt' OVERWRITE INTO TABLE DECIMAL
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/decimal_10_0.txt' OVERWRITE INTO TABLE `DECIMAL`
 PREHOOK: type: LOAD
 #### A masked pattern was here ####
 PREHOOK: Output: default@decimal
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/decimal_10_0.txt' OVERWRITE INTO TABLE DECIMAL
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/decimal_10_0.txt' OVERWRITE INTO TABLE `DECIMAL`
 POSTHOOK: type: LOAD
 #### A masked pattern was here ####
 POSTHOOK: Output: default@decimal
-PREHOOK: query: SELECT dec FROM DECIMAL
+PREHOOK: query: SELECT dec FROM `DECIMAL`
 PREHOOK: type: QUERY
 PREHOOK: Input: default@decimal
 #### A masked pattern was here ####
-POSTHOOK: query: SELECT dec FROM DECIMAL
+POSTHOOK: query: SELECT dec FROM `DECIMAL`
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@decimal
 #### A masked pattern was here ####
 1000000000
 NULL
-PREHOOK: query: DROP TABLE DECIMAL
+PREHOOK: query: DROP TABLE `DECIMAL`
 PREHOOK: type: DROPTABLE
 PREHOOK: Input: default@decimal
 PREHOOK: Output: default@decimal
-POSTHOOK: query: DROP TABLE DECIMAL
+POSTHOOK: query: DROP TABLE `DECIMAL`
 POSTHOOK: type: DROPTABLE
 POSTHOOK: Input: default@decimal
 POSTHOOK: Output: default@decimal

Modified: hive/branches/spark/ql/src/test/results/clientpositive/decimal_udf.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/decimal_udf.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/decimal_udf.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/decimal_udf.q.out Sat Mar 28 14:03:43 2015
@@ -649,7 +649,7 @@ STAGE PLANS:
           alias: decimal_udf
           Statistics: Num rows: 3 Data size: 359 Basic stats: COMPLETE Column stats: NONE
           Filter Operator
-            predicate: ((key * CAST( value AS decimal(10,0))) > CAST( 0 AS decimal(31,10))) (type: boolean)
+            predicate: ((key * CAST( value AS decimal(10,0))) > 0) (type: boolean)
             Statistics: Num rows: 1 Data size: 119 Basic stats: COMPLETE Column stats: NONE
             Select Operator
               expressions: key (type: decimal(20,10)), value (type: int)
@@ -907,7 +907,7 @@ STAGE PLANS:
           alias: decimal_udf
           Statistics: Num rows: 3 Data size: 359 Basic stats: COMPLETE Column stats: NONE
           Select Operator
-            expressions: (key / CAST( 0 AS decimal(10,0))) (type: decimal(22,12))
+            expressions: (key / 0) (type: decimal(22,12))
             outputColumnNames: _col0
             Statistics: Num rows: 3 Data size: 359 Basic stats: COMPLETE Column stats: NONE
             Limit
@@ -973,7 +973,7 @@ STAGE PLANS:
           alias: decimal_udf
           Statistics: Num rows: 3 Data size: 359 Basic stats: COMPLETE Column stats: NONE
           Filter Operator
-            predicate: (key <> CAST( 0 AS decimal(20,10))) (type: boolean)
+            predicate: (key <> 0) (type: boolean)
             Statistics: Num rows: 3 Data size: 359 Basic stats: COMPLETE Column stats: NONE
             Select Operator
               expressions: (key / key) (type: decimal(38,24))
@@ -1808,7 +1808,7 @@ STAGE PLANS:
           alias: decimal_udf
           Statistics: Num rows: 3 Data size: 359 Basic stats: COMPLETE Column stats: NONE
           Select Operator
-            expressions: ((key + CAST( 1 AS decimal(10,0))) % (key / CAST( 2 AS decimal(10,0)))) (type: decimal(22,12))
+            expressions: ((key + 1) % (key / 2)) (type: decimal(22,12))
             outputColumnNames: _col0
             Statistics: Num rows: 3 Data size: 359 Basic stats: COMPLETE Column stats: NONE
             ListSink

Modified: hive/branches/spark/ql/src/test/results/clientpositive/decimal_udf2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/decimal_udf2.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/decimal_udf2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/decimal_udf2.q.out Sat Mar 28 14:03:43 2015
@@ -44,10 +44,10 @@ STAGE PLANS:
             alias: decimal_udf2
             Statistics: Num rows: 3 Data size: 359 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
-              predicate: (key = CAST( 10 AS decimal(20,10))) (type: boolean)
+              predicate: (key = 10) (type: boolean)
               Statistics: Num rows: 1 Data size: 119 Basic stats: COMPLETE Column stats: NONE
               Select Operator
-                expressions: acos(key) (type: double), asin(key) (type: double), atan(key) (type: double), cos(key) (type: double), sin(key) (type: double), tan(key) (type: double), radians(key) (type: double)
+                expressions: NaN (type: double), NaN (type: double), 1.4711276743037347 (type: double), -0.8390715290764524 (type: double), -0.5440211108893698 (type: double), 0.6483608274590866 (type: double), 0.17453292519943295 (type: double)
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6
                 Statistics: Num rows: 1 Data size: 119 Basic stats: COMPLETE Column stats: NONE
                 File Output Operator
@@ -101,10 +101,10 @@ STAGE PLANS:
             alias: decimal_udf2
             Statistics: Num rows: 3 Data size: 359 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
-              predicate: (key = CAST( 10 AS decimal(20,10))) (type: boolean)
+              predicate: (key = 10) (type: boolean)
               Statistics: Num rows: 1 Data size: 119 Basic stats: COMPLETE Column stats: NONE
               Select Operator
-                expressions: exp(key) (type: double), ln(key) (type: double), log(key) (type: double), log(key, key) (type: double), log(key, value) (type: double), log(value, key) (type: double), log10(key) (type: double), sqrt(key) (type: double)
+                expressions: 22026.465794806718 (type: double), 2.302585092994046 (type: double), 2.302585092994046 (type: double), 1.0 (type: double), log(10, value) (type: double), log(value, 10) (type: double), 1.0 (type: double), 3.1622776601683795 (type: double)
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7
                 Statistics: Num rows: 1 Data size: 119 Basic stats: COMPLETE Column stats: NONE
                 File Output Operator

Modified: hive/branches/spark/ql/src/test/results/clientpositive/display_colstats_tbllvl.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/display_colstats_tbllvl.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/display_colstats_tbllvl.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/display_colstats_tbllvl.q.out Sat Mar 28 14:03:43 2015
@@ -97,7 +97,7 @@ STAGE PLANS:
       Column Stats Desc:
           Columns: sourceIP, avgTimeOnSite, adRevenue
           Column Types: string, int, float
-          Table: uservisits_web_text_none
+          Table: default.uservisits_web_text_none
 
 PREHOOK: query: explain extended
 analyze table UserVisits_web_text_none compute statistics for columns sourceIP, avgTimeOnSite, adRevenue
@@ -220,7 +220,7 @@ STAGE PLANS:
       Column Stats Desc:
           Columns: sourceIP, avgTimeOnSite, adRevenue
           Column Types: string, int, float
-          Table: uservisits_web_text_none
+          Table: default.uservisits_web_text_none
           Is Table Level Stats: true
 
 PREHOOK: query: analyze table UserVisits_web_text_none compute statistics for columns sourceIP, avgTimeOnSite, adRevenue
@@ -330,7 +330,7 @@ STAGE PLANS:
       Column Stats Desc:
           Columns: a, b, c, d, e
           Column Types: int, double, string, boolean, binary
-          Table: empty_tab
+          Table: default.empty_tab
 
 PREHOOK: query: analyze table empty_tab compute statistics for columns a,b,c,d,e
 PREHOOK: type: QUERY

Modified: hive/branches/spark/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out Sat Mar 28 14:03:43 2015
@@ -896,7 +896,7 @@ Partition Parameters:
 	numFiles            	2                   
 	numRows             	32                  
 	rawDataSize         	640                 
-	totalSize           	1400                
+	totalSize           	1424                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -940,7 +940,7 @@ Partition Parameters:
 	numFiles            	2                   
 	numRows             	6                   
 	rawDataSize         	120                 
-	totalSize           	1102                
+	totalSize           	1128                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -984,7 +984,7 @@ Partition Parameters:
 	numFiles            	2                   
 	numRows             	14                  
 	rawDataSize         	280                 
-	totalSize           	1216                
+	totalSize           	1240                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -1028,7 +1028,7 @@ Partition Parameters:
 	numFiles            	2                   
 	numRows             	6                   
 	rawDataSize         	120                 
-	totalSize           	1102                
+	totalSize           	1128                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -1071,7 +1071,7 @@ Partition Parameters:
 	numFiles            	8                   
 	numRows             	32                  
 	rawDataSize         	640                 
-	totalSize           	4548                
+	totalSize           	4644                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -1114,7 +1114,7 @@ Partition Parameters:
 	numFiles            	8                   
 	numRows             	6                   
 	rawDataSize         	120                 
-	totalSize           	2212                
+	totalSize           	2262                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -1157,7 +1157,7 @@ Partition Parameters:
 	numFiles            	8                   
 	numRows             	32                  
 	rawDataSize         	640                 
-	totalSize           	4534                
+	totalSize           	4630                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -1200,7 +1200,7 @@ Partition Parameters:
 	numFiles            	8                   
 	numRows             	6                   
 	rawDataSize         	120                 
-	totalSize           	2212                
+	totalSize           	2262                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 

Modified: hive/branches/spark/ql/src/test/results/clientpositive/dynpart_sort_optimization2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/dynpart_sort_optimization2.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/dynpart_sort_optimization2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/dynpart_sort_optimization2.q.out Sat Mar 28 14:03:43 2015
@@ -1171,7 +1171,7 @@ Partition Parameters:
 	numFiles            	1                   
 	numRows             	11                  
 	rawDataSize         	88                  
-	totalSize           	433                 
+	totalSize           	445                 
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -1233,7 +1233,7 @@ Partition Parameters:
 	numFiles            	1                   
 	numRows             	13                  
 	rawDataSize         	104                 
-	totalSize           	456                 
+	totalSize           	468                 
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -1392,7 +1392,7 @@ Partition Parameters:
 	numFiles            	1                   
 	numRows             	11                  
 	rawDataSize         	88                  
-	totalSize           	433                 
+	totalSize           	445                 
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -1454,7 +1454,7 @@ Partition Parameters:
 	numFiles            	1                   
 	numRows             	13                  
 	rawDataSize         	104                 
-	totalSize           	456                 
+	totalSize           	468                 
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 

Modified: hive/branches/spark/ql/src/test/results/clientpositive/encrypted/encryption_insert_partition_dynamic.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/encrypted/encryption_insert_partition_dynamic.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/encrypted/encryption_insert_partition_dynamic.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/encrypted/encryption_insert_partition_dynamic.q.out Sat Mar 28 14:03:43 2015
@@ -552,16 +552,16 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: encryptedtable
-            Statistics: Num rows: 21 Data size: 2372 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 24 Data size: 2447 Basic stats: COMPLETE Column stats: NONE
             GatherStats: false
             Select Operator
               expressions: value (type: string), key (type: string)
               outputColumnNames: _col0, _col1
-              Statistics: Num rows: 21 Data size: 2372 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 24 Data size: 2447 Basic stats: COMPLETE Column stats: NONE
               Reduce Output Operator
                 sort order: 
                 Map-reduce partition columns: _col0 (type: string)
-                Statistics: Num rows: 21 Data size: 2372 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 24 Data size: 2447 Basic stats: COMPLETE Column stats: NONE
                 tag: -1
                 value expressions: _col0 (type: string), _col1 (type: string)
                 auto parallelism: false
@@ -592,7 +592,7 @@ STAGE PLANS:
               serialization.ddl struct encryptedtable { string value}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 600
+              totalSize 618
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -640,7 +640,7 @@ STAGE PLANS:
               serialization.ddl struct encryptedtable { string value}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 592
+              totalSize 611
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -688,7 +688,7 @@ STAGE PLANS:
               serialization.ddl struct encryptedtable { string value}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 592
+              totalSize 611
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -736,7 +736,7 @@ STAGE PLANS:
               serialization.ddl struct encryptedtable { string value}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 588
+              totalSize 607
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -770,13 +770,13 @@ STAGE PLANS:
         Select Operator
           expressions: VALUE._col0 (type: string), VALUE._col1 (type: string)
           outputColumnNames: _col0, _col1
-          Statistics: Num rows: 21 Data size: 2372 Basic stats: COMPLETE Column stats: NONE
+          Statistics: Num rows: 24 Data size: 2447 Basic stats: COMPLETE Column stats: NONE
           File Output Operator
             compressed: false
             GlobalTableId: 1
 #### A PARTIAL masked pattern was here #### data/warehouse/unencryptedtable/.hive-staging
             NumFilesPerFileSink: 1
-            Statistics: Num rows: 21 Data size: 2372 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 24 Data size: 2447 Basic stats: COMPLETE Column stats: NONE
 #### A PARTIAL masked pattern was here #### data/warehouse/unencryptedtable/.hive-staging
             table:
                 input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat

Modified: hive/branches/spark/ql/src/test/results/clientpositive/encrypted/encryption_insert_partition_static.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/encrypted/encryption_insert_partition_static.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/encrypted/encryption_insert_partition_static.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/encrypted/encryption_insert_partition_static.q.out Sat Mar 28 14:03:43 2015
@@ -555,16 +555,16 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: encryptedtable
-            Statistics: Num rows: 12 Data size: 2695 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 12 Data size: 2777 Basic stats: COMPLETE Column stats: NONE
             GatherStats: false
             Select Operator
               expressions: key (type: string), value (type: string)
               outputColumnNames: _col0, _col1
-              Statistics: Num rows: 12 Data size: 2695 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 12 Data size: 2777 Basic stats: COMPLETE Column stats: NONE
               Reduce Output Operator
                 sort order: 
                 Map-reduce partition columns: _col0 (type: string)
-                Statistics: Num rows: 12 Data size: 2695 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 12 Data size: 2777 Basic stats: COMPLETE Column stats: NONE
                 tag: -1
                 value expressions: _col0 (type: string), _col1 (type: string)
                 auto parallelism: false
@@ -595,7 +595,7 @@ STAGE PLANS:
               serialization.ddl struct encryptedtable { string key, string value}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 1351
+              totalSize 1392
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -643,7 +643,7 @@ STAGE PLANS:
               serialization.ddl struct encryptedtable { string key, string value}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 1344
+              totalSize 1385
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -675,14 +675,14 @@ STAGE PLANS:
         Select Operator
           expressions: VALUE._col0 (type: string), VALUE._col1 (type: string)
           outputColumnNames: _col0, _col1
-          Statistics: Num rows: 12 Data size: 2695 Basic stats: COMPLETE Column stats: NONE
+          Statistics: Num rows: 12 Data size: 2777 Basic stats: COMPLETE Column stats: NONE
           File Output Operator
             compressed: false
             GlobalTableId: 1
 #### A PARTIAL masked pattern was here #### data/warehouse/unencryptedtable/ds=today/.hive-staging
             NumFilesPerFileSink: 1
             Static Partition Specification: ds=today/
-            Statistics: Num rows: 12 Data size: 2695 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 12 Data size: 2777 Basic stats: COMPLETE Column stats: NONE
 #### A PARTIAL masked pattern was here #### data/warehouse/unencryptedtable/ds=today/.hive-staging
             table:
                 input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat

Modified: hive/branches/spark/ql/src/test/results/clientpositive/extrapolate_part_stats_full.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/extrapolate_part_stats_full.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/extrapolate_part_stats_full.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/extrapolate_part_stats_full.q.out Sat Mar 28 14:03:43 2015
@@ -136,7 +136,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 362
+              totalSize 375
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -179,7 +179,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 382
+              totalSize 397
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -269,7 +269,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 362
+              totalSize 375
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -312,7 +312,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 382
+              totalSize 397
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -469,7 +469,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 277
+              totalSize 290
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -513,7 +513,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 279
+              totalSize 291
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -557,7 +557,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 261
+              totalSize 274
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -601,7 +601,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 261
+              totalSize 274
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -688,7 +688,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 277
+              totalSize 290
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -732,7 +732,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 279
+              totalSize 291
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -776,7 +776,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 261
+              totalSize 274
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -820,7 +820,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 261
+              totalSize 274
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           

Modified: hive/branches/spark/ql/src/test/results/clientpositive/extrapolate_part_stats_partial.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/extrapolate_part_stats_partial.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/extrapolate_part_stats_partial.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/extrapolate_part_stats_partial.q.out Sat Mar 28 14:03:43 2015
@@ -153,7 +153,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 362
+              totalSize 375
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -196,7 +196,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 383
+              totalSize 398
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -239,7 +239,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 404
+              totalSize 418
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -282,7 +282,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 412
+              totalSize 427
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -372,7 +372,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 362
+              totalSize 375
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -415,7 +415,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 383
+              totalSize 398
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -458,7 +458,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 404
+              totalSize 418
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -501,7 +501,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 412
+              totalSize 427
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -604,7 +604,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 362
+              totalSize 375
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -647,7 +647,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 383
+              totalSize 398
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -690,7 +690,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 404
+              totalSize 418
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -733,7 +733,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 412
+              totalSize 427
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -819,7 +819,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 362
+              totalSize 375
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -862,7 +862,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 383
+              totalSize 398
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -905,7 +905,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 404
+              totalSize 418
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -948,7 +948,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 412
+              totalSize 427
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1106,7 +1106,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 281
+              totalSize 293
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1150,7 +1150,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 291
+              totalSize 303
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1194,7 +1194,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 298
+              totalSize 312
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1238,7 +1238,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 277
+              totalSize 290
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1282,7 +1282,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 272
+              totalSize 285
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1326,7 +1326,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 283
+              totalSize 295
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1370,7 +1370,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 292
+              totalSize 305
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1414,7 +1414,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 261
+              totalSize 274
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1458,7 +1458,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 261
+              totalSize 274
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1502,7 +1502,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 294
+              totalSize 306
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1546,7 +1546,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 291
+              totalSize 305
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1633,7 +1633,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 281
+              totalSize 293
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1677,7 +1677,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 291
+              totalSize 303
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1721,7 +1721,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 298
+              totalSize 312
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1765,7 +1765,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 277
+              totalSize 290
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1809,7 +1809,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 272
+              totalSize 285
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1853,7 +1853,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 283
+              totalSize 295
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1897,7 +1897,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 292
+              totalSize 305
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1941,7 +1941,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 261
+              totalSize 274
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -1985,7 +1985,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 261
+              totalSize 274
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -2029,7 +2029,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 294
+              totalSize 306
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           
@@ -2073,7 +2073,7 @@ STAGE PLANS:
               serialization.ddl struct loc_orc_2d { string state, i32 locid}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde
-              totalSize 291
+              totalSize 305
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
           

Modified: hive/branches/spark/ql/src/test/results/clientpositive/groupby_cube1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/groupby_cube1.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/groupby_cube1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/groupby_cube1.q.out Sat Mar 28 14:03:43 2015
@@ -206,11 +206,11 @@ STAGE PLANS:
             Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
             Select Operator
               expressions: key (type: string), val (type: string)
-              outputColumnNames: key, val
+              outputColumnNames: _col0, _col1
               Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
               Group By Operator
-                aggregations: count(DISTINCT val)
-                keys: key (type: string), '0' (type: string), val (type: string)
+                aggregations: count(DISTINCT _col1)
+                keys: _col0 (type: string), '0' (type: string), _col1 (type: string)
                 mode: hash
                 outputColumnNames: _col0, _col1, _col2, _col3
                 Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
@@ -388,11 +388,11 @@ STAGE PLANS:
             Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
             Select Operator
               expressions: key (type: string), val (type: string)
-              outputColumnNames: key, val
+              outputColumnNames: _col0, _col1
               Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
               Group By Operator
-                aggregations: count(DISTINCT val)
-                keys: key (type: string), '0' (type: string), val (type: string)
+                aggregations: count(DISTINCT _col1)
+                keys: _col0 (type: string), '0' (type: string), _col1 (type: string)
                 mode: hash
                 outputColumnNames: _col0, _col1, _col2, _col3
                 Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE

Modified: hive/branches/spark/ql/src/test/results/clientpositive/groupby_rollup1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/groupby_rollup1.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/groupby_rollup1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/groupby_rollup1.q.out Sat Mar 28 14:03:43 2015
@@ -116,11 +116,11 @@ STAGE PLANS:
             Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
             Select Operator
               expressions: key (type: string), val (type: string)
-              outputColumnNames: key, val
+              outputColumnNames: _col0, _col1
               Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
               Group By Operator
-                aggregations: count(DISTINCT val)
-                keys: key (type: string), '0' (type: string), val (type: string)
+                aggregations: count(DISTINCT _col1)
+                keys: _col0 (type: string), '0' (type: string), _col1 (type: string)
                 mode: hash
                 outputColumnNames: _col0, _col1, _col2, _col3
                 Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
@@ -292,11 +292,11 @@ STAGE PLANS:
             Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
             Select Operator
               expressions: key (type: string), val (type: string)
-              outputColumnNames: key, val
+              outputColumnNames: _col0, _col1
               Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
               Group By Operator
-                aggregations: count(DISTINCT val)
-                keys: key (type: string), '0' (type: string), val (type: string)
+                aggregations: count(DISTINCT _col1)
+                keys: _col0 (type: string), '0' (type: string), _col1 (type: string)
                 mode: hash
                 outputColumnNames: _col0, _col1, _col2, _col3
                 Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE

Modified: hive/branches/spark/ql/src/test/results/clientpositive/innerjoin.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/innerjoin.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/innerjoin.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/innerjoin.q.out Sat Mar 28 14:03:43 2015
@@ -1142,44 +1142,46 @@ POSTHOOK: Input: default@dest_j1
 98	val_98
 98	val_98
 PREHOOK: query: -- verify that INNER is a non-reserved word for backwards compatibility
-create table inner(i int)
+-- change from HIVE-6617, inner is a SQL2011 reserved keyword.
+create table `inner`(i int)
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:default
 PREHOOK: Output: default@inner
 POSTHOOK: query: -- verify that INNER is a non-reserved word for backwards compatibility
-create table inner(i int)
+-- change from HIVE-6617, inner is a SQL2011 reserved keyword.
+create table `inner`(i int)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: database:default
 POSTHOOK: Output: default@inner
-PREHOOK: query: select i from inner
+PREHOOK: query: select i from `inner`
 PREHOOK: type: QUERY
 PREHOOK: Input: default@inner
 #### A masked pattern was here ####
-POSTHOOK: query: select i from inner
+POSTHOOK: query: select i from `inner`
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@inner
 #### A masked pattern was here ####
-PREHOOK: query: create table i(inner int)
+PREHOOK: query: create table i(`inner` int)
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:default
 PREHOOK: Output: default@i
-POSTHOOK: query: create table i(inner int)
+POSTHOOK: query: create table i(`inner` int)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: database:default
 POSTHOOK: Output: default@i
-PREHOOK: query: select inner from i
+PREHOOK: query: select `inner` from i
 PREHOOK: type: QUERY
 PREHOOK: Input: default@i
 #### A masked pattern was here ####
-POSTHOOK: query: select inner from i
+POSTHOOK: query: select `inner` from i
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@i
 #### A masked pattern was here ####
-PREHOOK: query: explain select * from (select * from src) inner left outer join src
-on inner.key=src.key
+PREHOOK: query: explain select * from (select * from src) `inner` left outer join src
+on `inner`.key=src.key
 PREHOOK: type: QUERY
-POSTHOOK: query: explain select * from (select * from src) inner left outer join src
-on inner.key=src.key
+POSTHOOK: query: explain select * from (select * from src) `inner` left outer join src
+on `inner`.key=src.key
 POSTHOOK: type: QUERY
 STAGE DEPENDENCIES:
   Stage-1 is a root stage

Modified: hive/branches/spark/ql/src/test/results/clientpositive/keyword_1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/keyword_1.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/keyword_1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/keyword_1.q.out Sat Mar 28 14:03:43 2015
@@ -1,12 +1,12 @@
 PREHOOK: query: -- SORT_BEFORE_DIFF
 
-create table test_user (user string, `group` string)
+create table test_user (`user` string, `group` string)
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:default
 PREHOOK: Output: default@test_user
 POSTHOOK: query: -- SORT_BEFORE_DIFF
 
-create table test_user (user string, `group` string)
+create table test_user (`user` string, `group` string)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: database:default
 POSTHOOK: Output: default@test_user
@@ -16,9 +16,9 @@ PREHOOK: Output: default@test_user
 POSTHOOK: query: grant select on table test_user to user hive_test
 POSTHOOK: type: GRANT_PRIVILEGE
 POSTHOOK: Output: default@test_user
-PREHOOK: query: explain select user from test_user
+PREHOOK: query: explain select `user` from test_user
 PREHOOK: type: QUERY
-POSTHOOK: query: explain select user from test_user
+POSTHOOK: query: explain select `user` from test_user
 POSTHOOK: type: QUERY
 STAGE DEPENDENCIES:
   Stage-0 is a root stage

Modified: hive/branches/spark/ql/src/test/results/clientpositive/literal_decimal.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/literal_decimal.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/literal_decimal.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/literal_decimal.q.out Sat Mar 28 14:03:43 2015
@@ -14,7 +14,7 @@ STAGE PLANS:
           alias: src
           Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE
           Select Operator
-            expressions: (- 1) (type: decimal(1,0)), 0 (type: decimal(1,0)), 1 (type: decimal(1,0)), 3.14 (type: decimal(3,2)), (- 3.14) (type: decimal(3,2)), 99999999999999999 (type: decimal(17,0)), 99999999999999999.9999999999999 (type: decimal(30,13)), 1E99 (type: decimal(1,0))
+            expressions: -1 (type: decimal(1,0)), 0 (type: decimal(1,0)), 1 (type: decimal(1,0)), 3.14 (type: decimal(3,2)), -3.14 (type: decimal(3,2)), 99999999999999999 (type: decimal(17,0)), 99999999999999999.9999999999999 (type: decimal(30,13)), 1E99 (type: decimal(1,0))
             outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7
             Statistics: Num rows: 500 Data size: 448000 Basic stats: COMPLETE Column stats: COMPLETE
             Limit

Modified: hive/branches/spark/ql/src/test/results/clientpositive/nonreserved_keywords_input37.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/nonreserved_keywords_input37.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/nonreserved_keywords_input37.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/nonreserved_keywords_input37.q.out Sat Mar 28 14:03:43 2015
@@ -1,38 +1,38 @@
-PREHOOK: query: CREATE TABLE table(string string) STORED AS TEXTFILE
+PREHOOK: query: CREATE TABLE `table`(`string` string) STORED AS TEXTFILE
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:default
 PREHOOK: Output: default@table
-POSTHOOK: query: CREATE TABLE table(string string) STORED AS TEXTFILE
+POSTHOOK: query: CREATE TABLE `table`(`string` string) STORED AS TEXTFILE
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: database:default
 POSTHOOK: Output: default@table
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/docurl.txt' INTO TABLE table
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/docurl.txt' INTO TABLE `table`
 PREHOOK: type: LOAD
 #### A masked pattern was here ####
 PREHOOK: Output: default@table
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/docurl.txt' INTO TABLE table
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/docurl.txt' INTO TABLE `table`
 POSTHOOK: type: LOAD
 #### A masked pattern was here ####
 POSTHOOK: Output: default@table
-PREHOOK: query: SELECT table, count(1)
+PREHOOK: query: SELECT `table`, count(1)
 FROM
 (
-  FROM table
-  SELECT TRANSFORM (table.string)
+  FROM `table`
+  SELECT TRANSFORM (`table`.`string`)
 #### A masked pattern was here ####
 ) subq
-GROUP BY table
+GROUP BY `table`
 PREHOOK: type: QUERY
 PREHOOK: Input: default@table
 #### A masked pattern was here ####
-POSTHOOK: query: SELECT table, count(1)
+POSTHOOK: query: SELECT `table`, count(1)
 FROM
 (
-  FROM table
-  SELECT TRANSFORM (table.string)
+  FROM `table`
+  SELECT TRANSFORM (`table`.`string`)
 #### A masked pattern was here ####
 ) subq
-GROUP BY table
+GROUP BY `table`
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@table
 #### A masked pattern was here ####

Modified: hive/branches/spark/ql/src/test/results/clientpositive/nonreserved_keywords_insert_into1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/nonreserved_keywords_insert_into1.q.out?rev=1669775&r1=1669774&r2=1669775&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/nonreserved_keywords_insert_into1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/nonreserved_keywords_insert_into1.q.out Sat Mar 28 14:03:43 2015
@@ -1,18 +1,18 @@
-PREHOOK: query: DROP TABLE insert
+PREHOOK: query: DROP TABLE `insert`
 PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE insert
+POSTHOOK: query: DROP TABLE `insert`
 POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE insert (key INT, as STRING)
+PREHOOK: query: CREATE TABLE `insert` (key INT, `as` STRING)
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:default
 PREHOOK: Output: default@insert
-POSTHOOK: query: CREATE TABLE insert (key INT, as STRING)
+POSTHOOK: query: CREATE TABLE `insert` (key INT, `as` STRING)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: database:default
 POSTHOOK: Output: default@insert
-PREHOOK: query: EXPLAIN INSERT INTO TABLE insert SELECT * FROM src LIMIT 100
+PREHOOK: query: EXPLAIN INSERT INTO TABLE `insert` SELECT * FROM src LIMIT 100
 PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN INSERT INTO TABLE insert SELECT * FROM src LIMIT 100
+POSTHOOK: query: EXPLAIN INSERT INTO TABLE `insert` SELECT * FROM src LIMIT 100
 POSTHOOK: type: QUERY
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
@@ -71,32 +71,32 @@ STAGE PLANS:
   Stage: Stage-2
     Stats-Aggr Operator
 
-PREHOOK: query: INSERT INTO TABLE insert SELECT * FROM src LIMIT 100
+PREHOOK: query: INSERT INTO TABLE `insert` SELECT * FROM src LIMIT 100
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
 PREHOOK: Output: default@insert
-POSTHOOK: query: INSERT INTO TABLE insert SELECT * FROM src LIMIT 100
+POSTHOOK: query: INSERT INTO TABLE `insert` SELECT * FROM src LIMIT 100
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
 POSTHOOK: Output: default@insert
 POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
 POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
 PREHOOK: query: SELECT SUM(HASH(hash)) FROM (
-    SELECT TRANSFORM(*) USING 'tr \t _' AS (hash) FROM insert
+    SELECT TRANSFORM(*) USING 'tr \t _' AS (hash) FROM `insert`
 ) t
 PREHOOK: type: QUERY
 PREHOOK: Input: default@insert
 #### A masked pattern was here ####
 POSTHOOK: query: SELECT SUM(HASH(hash)) FROM (
-    SELECT TRANSFORM(*) USING 'tr \t _' AS (hash) FROM insert
+    SELECT TRANSFORM(*) USING 'tr \t _' AS (hash) FROM `insert`
 ) t
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@insert
 #### A masked pattern was here ####
 10226524244
-PREHOOK: query: EXPLAIN INSERT INTO TABLE insert SELECT * FROM src LIMIT 100
+PREHOOK: query: EXPLAIN INSERT INTO TABLE `insert` SELECT * FROM src LIMIT 100
 PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN INSERT INTO TABLE insert SELECT * FROM src LIMIT 100
+POSTHOOK: query: EXPLAIN INSERT INTO TABLE `insert` SELECT * FROM src LIMIT 100
 POSTHOOK: type: QUERY
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
@@ -155,41 +155,41 @@ STAGE PLANS:
   Stage: Stage-2
     Stats-Aggr Operator
 
-PREHOOK: query: INSERT INTO TABLE insert SELECT * FROM src LIMIT 100
+PREHOOK: query: INSERT INTO TABLE `insert` SELECT * FROM src LIMIT 100
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
 PREHOOK: Output: default@insert
-POSTHOOK: query: INSERT INTO TABLE insert SELECT * FROM src LIMIT 100
+POSTHOOK: query: INSERT INTO TABLE `insert` SELECT * FROM src LIMIT 100
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
 POSTHOOK: Output: default@insert
 POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
 POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
 PREHOOK: query: SELECT SUM(HASH(sum)) FROM (
-    SELECT TRANSFORM(*) USING 'tr \t _' AS (sum) FROM insert
+    SELECT TRANSFORM(*) USING 'tr \t _' AS (sum) FROM `insert`
 ) t
 PREHOOK: type: QUERY
 PREHOOK: Input: default@insert
 #### A masked pattern was here ####
 POSTHOOK: query: SELECT SUM(HASH(sum)) FROM (
-    SELECT TRANSFORM(*) USING 'tr \t _' AS (sum) FROM insert
+    SELECT TRANSFORM(*) USING 'tr \t _' AS (sum) FROM `insert`
 ) t
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@insert
 #### A masked pattern was here ####
 20453048488
-PREHOOK: query: SELECT COUNT(*) FROM insert
+PREHOOK: query: SELECT COUNT(*) FROM `insert`
 PREHOOK: type: QUERY
 PREHOOK: Input: default@insert
 #### A masked pattern was here ####
-POSTHOOK: query: SELECT COUNT(*) FROM insert
+POSTHOOK: query: SELECT COUNT(*) FROM `insert`
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@insert
 #### A masked pattern was here ####
 200
-PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE insert SELECT * FROM src LIMIT 10
+PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE `insert` SELECT * FROM src LIMIT 10
 PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN INSERT OVERWRITE TABLE insert SELECT * FROM src LIMIT 10
+POSTHOOK: query: EXPLAIN INSERT OVERWRITE TABLE `insert` SELECT * FROM src LIMIT 10
 POSTHOOK: type: QUERY
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
@@ -248,34 +248,34 @@ STAGE PLANS:
   Stage: Stage-2
     Stats-Aggr Operator
 
-PREHOOK: query: INSERT OVERWRITE TABLE insert SELECT * FROM src LIMIT 10
+PREHOOK: query: INSERT OVERWRITE TABLE `insert` SELECT * FROM src LIMIT 10
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
 PREHOOK: Output: default@insert
-POSTHOOK: query: INSERT OVERWRITE TABLE insert SELECT * FROM src LIMIT 10
+POSTHOOK: query: INSERT OVERWRITE TABLE `insert` SELECT * FROM src LIMIT 10
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
 POSTHOOK: Output: default@insert
 POSTHOOK: Lineage: insert.as SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
 POSTHOOK: Lineage: insert.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
 PREHOOK: query: SELECT SUM(HASH(add)) FROM (
-    SELECT TRANSFORM(*) USING 'tr \t _' AS (add) FROM insert
+    SELECT TRANSFORM(*) USING 'tr \t _' AS (add) FROM `insert`
 ) t
 PREHOOK: type: QUERY
 PREHOOK: Input: default@insert
 #### A masked pattern was here ####
 POSTHOOK: query: SELECT SUM(HASH(add)) FROM (
-    SELECT TRANSFORM(*) USING 'tr \t _' AS (add) FROM insert
+    SELECT TRANSFORM(*) USING 'tr \t _' AS (add) FROM `insert`
 ) t
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@insert
 #### A masked pattern was here ####
 -826625916
-PREHOOK: query: DROP TABLE insert
+PREHOOK: query: DROP TABLE `insert`
 PREHOOK: type: DROPTABLE
 PREHOOK: Input: default@insert
 PREHOOK: Output: default@insert
-POSTHOOK: query: DROP TABLE insert
+POSTHOOK: query: DROP TABLE `insert`
 POSTHOOK: type: DROPTABLE
 POSTHOOK: Input: default@insert
 POSTHOOK: Output: default@insert



Mime
View raw message