hive-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From pxi...@apache.org
Subject [15/21] hive git commit: HIVE-13341: Stats state is not captured correctly: differentiate load table and create table (Pengcheng Xiong, reviewed by Ashutosh Chauhan)
Date Mon, 09 May 2016 17:41:57 GMT
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out b/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out
index d40a693..de1305f 100644
--- a/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out
+++ b/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out
@@ -439,6 +439,8 @@ Table:              	list_bucketing_dynamic_part
 #### A masked pattern was here ####
 Partition Parameters:	 	 
 	numFiles            	3                   
+	numRows             	984                 
+	rawDataSize         	9488                
 	totalSize           	10586               
 #### A masked pattern was here ####
 	 	 
@@ -555,8 +557,10 @@ STAGE PLANS:
 #### A masked pattern was here ####
               name default.list_bucketing_dynamic_part
               numFiles 3
+              numRows 984
               partition_columns ds/hr
               partition_columns.types string:string
+              rawDataSize 9488
               serialization.ddl struct list_bucketing_dynamic_part { string key, string value}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
@@ -585,16 +589,16 @@ STAGE PLANS:
       Processor Tree:
         TableScan
           alias: list_bucketing_dynamic_part
-          Statistics: Num rows: 16 Data size: 136 Basic stats: PARTIAL Column stats: NONE
+          Statistics: Num rows: 1000 Data size: 9624 Basic stats: COMPLETE Column stats: NONE
           GatherStats: false
           Filter Operator
             isSamplingPred: false
             predicate: ((key = '484') and (value = 'val_484')) (type: boolean)
-            Statistics: Num rows: 4 Data size: 34 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 250 Data size: 2406 Basic stats: COMPLETE Column stats: NONE
             Select Operator
               expressions: '484' (type: string), 'val_484' (type: string), ds (type: string), hr (type: string)
               outputColumnNames: _col0, _col1, _col2, _col3
-              Statistics: Num rows: 4 Data size: 34 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 250 Data size: 2406 Basic stats: COMPLETE Column stats: NONE
               ListSink
 
 PREHOOK: query: select * from list_bucketing_dynamic_part where key = '484' and value = 'val_484'

http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/mapjoin_memcheck.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/mapjoin_memcheck.q.out b/ql/src/test/results/clientpositive/mapjoin_memcheck.q.out
index 8f5492f..65d896b 100644
--- a/ql/src/test/results/clientpositive/mapjoin_memcheck.q.out
+++ b/ql/src/test/results/clientpositive/mapjoin_memcheck.q.out
@@ -44,14 +44,14 @@ STAGE PLANS:
         $hdt$_0:src1 
           TableScan
             alias: src1
-            Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: key is not null (type: boolean)
-              Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: key (type: string), value (type: string)
                 outputColumnNames: _col0, _col1
-                Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE
                 HashTable Sink Operator
                   keys:
                     0 _col0 (type: string)
@@ -62,14 +62,14 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: src1
-            Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: key is not null (type: boolean)
-              Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: key (type: string), value (type: string)
                 outputColumnNames: _col0, _col1
-                Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE
                 Map Join Operator
                   condition map:
                        Inner Join 0 to 1
@@ -77,10 +77,10 @@ STAGE PLANS:
                     0 _col0 (type: string)
                     1 _col0 (type: string)
                   outputColumnNames: _col0, _col1, _col2, _col3
-                  Statistics: Num rows: 11 Data size: 77 Basic stats: COMPLETE Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: NONE
                   File Output Operator
                     compressed: false
-                    Statistics: Num rows: 11 Data size: 77 Basic stats: COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: NONE
                     table:
                         input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                         output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/metadataonly1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/metadataonly1.q.out b/ql/src/test/results/clientpositive/metadataonly1.q.out
index 0281407..4b4c57c 100644
--- a/ql/src/test/results/clientpositive/metadataonly1.q.out
+++ b/ql/src/test/results/clientpositive/metadataonly1.q.out
@@ -132,17 +132,22 @@ STAGE PLANS:
             partition values:
               ds 1
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test1
+              numFiles 0
+              numRows 0
               partition_columns ds
               partition_columns.types string
+              rawDataSize 0
               serialization.ddl struct test1 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.NullStructSerDe
           
@@ -256,17 +261,22 @@ STAGE PLANS:
             partition values:
               ds 1
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test1
+              numFiles 0
+              numRows 0
               partition_columns ds
               partition_columns.types string
+              rawDataSize 0
               serialization.ddl struct test1 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.NullStructSerDe
           
@@ -380,17 +390,22 @@ STAGE PLANS:
             partition values:
               ds 1
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test1
+              numFiles 0
+              numRows 0
               partition_columns ds
               partition_columns.types string
+              rawDataSize 0
               serialization.ddl struct test1 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
@@ -515,17 +530,22 @@ STAGE PLANS:
             partition values:
               ds 1
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test1
+              numFiles 0
+              numRows 0
               partition_columns ds
               partition_columns.types string
+              rawDataSize 0
               serialization.ddl struct test1 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.NullStructSerDe
           
@@ -554,17 +574,22 @@ STAGE PLANS:
             partition values:
               ds 2
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test1
+              numFiles 0
+              numRows 0
               partition_columns ds
               partition_columns.types string
+              rawDataSize 0
               serialization.ddl struct test1 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.NullStructSerDe
           
@@ -678,17 +703,22 @@ STAGE PLANS:
             partition values:
               ds 1
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test1
+              numFiles 0
+              numRows 0
               partition_columns ds
               partition_columns.types string
+              rawDataSize 0
               serialization.ddl struct test1 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
@@ -718,17 +748,22 @@ STAGE PLANS:
             partition values:
               ds 2
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test1
+              numFiles 0
+              numRows 0
               partition_columns ds
               partition_columns.types string
+              rawDataSize 0
               serialization.ddl struct test1 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
@@ -948,17 +983,22 @@ STAGE PLANS:
               ds 1
               hr 1
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test2
+              numFiles 0
+              numRows 0
               partition_columns ds/hr
               partition_columns.types string:string
+              rawDataSize 0
               serialization.ddl struct test2 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.NullStructSerDe
           
@@ -988,17 +1028,22 @@ STAGE PLANS:
               ds 1
               hr 2
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test2
+              numFiles 0
+              numRows 0
               partition_columns ds/hr
               partition_columns.types string:string
+              rawDataSize 0
               serialization.ddl struct test2 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.NullStructSerDe
           
@@ -1028,17 +1073,22 @@ STAGE PLANS:
               ds 1
               hr 3
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test2
+              numFiles 0
+              numRows 0
               partition_columns ds/hr
               partition_columns.types string:string
+              rawDataSize 0
               serialization.ddl struct test2 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.NullStructSerDe
           
@@ -1163,17 +1213,22 @@ STAGE PLANS:
               ds 1
               hr 1
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test2
+              numFiles 0
+              numRows 0
               partition_columns ds/hr
               partition_columns.types string:string
+              rawDataSize 0
               serialization.ddl struct test2 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
@@ -1204,17 +1259,22 @@ STAGE PLANS:
               ds 1
               hr 2
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test2
+              numFiles 0
+              numRows 0
               partition_columns ds/hr
               partition_columns.types string:string
+              rawDataSize 0
               serialization.ddl struct test2 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
@@ -1245,17 +1305,22 @@ STAGE PLANS:
               ds 1
               hr 3
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test2
+              numFiles 0
+              numRows 0
               partition_columns ds/hr
               partition_columns.types string:string
+              rawDataSize 0
               serialization.ddl struct test2 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
@@ -1375,17 +1440,22 @@ STAGE PLANS:
             partition values:
               ds 1
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test1
+              numFiles 0
+              numRows 0
               partition_columns ds
               partition_columns.types string
+              rawDataSize 0
               serialization.ddl struct test1 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.NullStructSerDe
           
@@ -1414,17 +1484,22 @@ STAGE PLANS:
             partition values:
               ds 2
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test1
+              numFiles 0
+              numRows 0
               partition_columns ds
               partition_columns.types string
+              rawDataSize 0
               serialization.ddl struct test1 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.NullStructSerDe
           
@@ -1598,17 +1673,22 @@ STAGE PLANS:
               ds 01:10:10
               hr 01
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test2
+              numFiles 0
+              numRows 0
               partition_columns ds/hr
               partition_columns.types string:string
+              rawDataSize 0
               serialization.ddl struct test2 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.NullStructSerDe
           
@@ -1638,17 +1718,22 @@ STAGE PLANS:
               ds 01:10:20
               hr 02
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test2
+              numFiles 0
+              numRows 0
               partition_columns ds/hr
               partition_columns.types string:string
+              rawDataSize 0
               serialization.ddl struct test2 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.NullStructSerDe
           
@@ -1678,17 +1763,22 @@ STAGE PLANS:
               ds 1
               hr 1
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test2
+              numFiles 0
+              numRows 0
               partition_columns ds/hr
               partition_columns.types string:string
+              rawDataSize 0
               serialization.ddl struct test2 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.NullStructSerDe
           
@@ -1718,17 +1808,22 @@ STAGE PLANS:
               ds 1
               hr 2
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test2
+              numFiles 0
+              numRows 0
               partition_columns ds/hr
               partition_columns.types string:string
+              rawDataSize 0
               serialization.ddl struct test2 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.NullStructSerDe
           
@@ -1758,17 +1853,22 @@ STAGE PLANS:
               ds 1
               hr 3
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns a,b
               columns.comments 
               columns.types int:double
 #### A masked pattern was here ####
               name default.test2
+              numFiles 0
+              numRows 0
               partition_columns ds/hr
               partition_columns.types string:string
+              rawDataSize 0
               serialization.ddl struct test2 { i32 a, double b}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.NullStructSerDe
           

http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/nullformat.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/nullformat.q.out b/ql/src/test/results/clientpositive/nullformat.q.out
index af91470..2a80359 100644
--- a/ql/src/test/results/clientpositive/nullformat.q.out
+++ b/ql/src/test/results/clientpositive/nullformat.q.out
@@ -93,6 +93,11 @@ OUTPUTFORMAT
 LOCATION
 #### A masked pattern was here ####
 TBLPROPERTIES (
+  'COLUMN_STATS_ACCURATE'='{\"BASIC_STATS\":\"true\"}', 
+  'numFiles'='0', 
+  'numRows'='0', 
+  'rawDataSize'='0', 
+  'totalSize'='0', 
 #### A masked pattern was here ####
 PREHOOK: query: -- load null data from another table and verify that the null is stored in the expected format
 INSERT OVERWRITE TABLE null_tab1 SELECT a,b FROM base_tab

http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/orc_create.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/orc_create.q.out b/ql/src/test/results/clientpositive/orc_create.q.out
index 3013fda..20c3fce 100644
--- a/ql/src/test/results/clientpositive/orc_create.q.out
+++ b/ql/src/test/results/clientpositive/orc_create.q.out
@@ -70,6 +70,11 @@ Retention:          	0
 #### A masked pattern was here ####
 Table Type:         	MANAGED_TABLE       	 
 Table Parameters:	 	 
+	COLUMN_STATS_ACCURATE	{\"BASIC_STATS\":\"true\"}
+	numFiles            	0                   
+	numRows             	0                   
+	rawDataSize         	0                   
+	totalSize           	0                   
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -319,6 +324,11 @@ Retention:          	0
 #### A masked pattern was here ####
 Table Type:         	MANAGED_TABLE       	 
 Table Parameters:	 	 
+	COLUMN_STATS_ACCURATE	{\"BASIC_STATS\":\"true\"}
+	numFiles            	0                   
+	numRows             	0                   
+	rawDataSize         	0                   
+	totalSize           	0                   
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 

http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/orc_llap.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/orc_llap.q.out b/ql/src/test/results/clientpositive/orc_llap.q.out
index 6fc73b7..bae69bb 100644
--- a/ql/src/test/results/clientpositive/orc_llap.q.out
+++ b/ql/src/test/results/clientpositive/orc_llap.q.out
@@ -719,17 +719,17 @@ STAGE PLANS:
           TableScan
             alias: orc_llap
             filterExpr: ((cint > 10) and cbigint is not null) (type: boolean)
-            Statistics: Num rows: 98779 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: ((cint > 10) and cbigint is not null) (type: boolean)
-              Statistics: Num rows: 32926 Data size: 526817 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: cint (type: int), csmallint (type: smallint), cbigint (type: bigint)
                 outputColumnNames: _col0, _col1, _col2
-                Statistics: Num rows: 32926 Data size: 526817 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE
                 File Output Operator
                   compressed: false
-                  Statistics: Num rows: 32926 Data size: 526817 Basic stats: COMPLETE Column stats: NONE
+                  Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE
                   table:
                       input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                       output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -792,17 +792,17 @@ STAGE PLANS:
           TableScan
             alias: orc_llap
             filterExpr: ((cint > 10) and cbigint is not null) (type: boolean)
-            Statistics: Num rows: 4938 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: ((cint > 10) and cbigint is not null) (type: boolean)
-              Statistics: Num rows: 1646 Data size: 526823 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: ctinyint (type: tinyint), csmallint (type: smallint), cint (type: int), cbigint (type: bigint), cfloat (type: float), cdouble (type: double), cstring1 (type: string), cstring2 (type: string), ctimestamp1 (type: timestamp), ctimestamp2 (type: timestamp), cboolean1 (type: boolean), cboolean2 (type: boolean)
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11
-                Statistics: Num rows: 1646 Data size: 526823 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE
                 File Output Operator
                   compressed: false
-                  Statistics: Num rows: 1646 Data size: 526823 Basic stats: COMPLETE Column stats: NONE
+                  Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE
                   table:
                       input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                       output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -874,17 +874,17 @@ STAGE PLANS:
           TableScan
             alias: orc_llap
             filterExpr: ((cint > 5) and (cint < 10)) (type: boolean)
-            Statistics: Num rows: 15196 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: ((cint > 5) and (cint < 10)) (type: boolean)
-              Statistics: Num rows: 1688 Data size: 175561 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 27306 Data size: 6462051 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: cstring2 (type: string)
                 outputColumnNames: _col0
-                Statistics: Num rows: 1688 Data size: 175561 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 27306 Data size: 6462051 Basic stats: COMPLETE Column stats: NONE
                 File Output Operator
                   compressed: false
-                  Statistics: Num rows: 1688 Data size: 175561 Basic stats: COMPLETE Column stats: NONE
+                  Statistics: Num rows: 27306 Data size: 6462051 Basic stats: COMPLETE Column stats: NONE
                   table:
                       input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                       output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -944,22 +944,22 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: orc_llap
-            Statistics: Num rows: 7902 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
             Select Operator
               expressions: cstring1 (type: string), cstring2 (type: string)
               outputColumnNames: cstring1, cstring2
-              Statistics: Num rows: 7902 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
               Group By Operator
                 aggregations: count()
                 keys: cstring1 (type: string), cstring2 (type: string)
                 mode: hash
                 outputColumnNames: _col0, _col1, _col2
-                Statistics: Num rows: 7902 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
                 Reduce Output Operator
                   key expressions: _col0 (type: string), _col1 (type: string)
                   sort order: ++
                   Map-reduce partition columns: _col0 (type: string), _col1 (type: string)
-                  Statistics: Num rows: 7902 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+                  Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
                   value expressions: _col2 (type: bigint)
       Execution mode: vectorized
       LLAP IO: all inputs
@@ -969,10 +969,10 @@ STAGE PLANS:
           keys: KEY._col0 (type: string), KEY._col1 (type: string)
           mode: mergepartial
           outputColumnNames: _col0, _col1, _col2
-          Statistics: Num rows: 3951 Data size: 790234 Basic stats: COMPLETE Column stats: NONE
+          Statistics: Num rows: 122880 Data size: 29079940 Basic stats: COMPLETE Column stats: NONE
           File Output Operator
             compressed: false
-            Statistics: Num rows: 3951 Data size: 790234 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 122880 Data size: 29079940 Basic stats: COMPLETE Column stats: NONE
             table:
                 input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                 output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -1039,14 +1039,14 @@ STAGE PLANS:
           TableScan
             alias: o1
             filterExpr: (csmallint is not null and cbigint is not null) (type: boolean)
-            Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: (csmallint is not null and cbigint is not null) (type: boolean)
-              Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: csmallint (type: smallint), cstring1 (type: string)
                 outputColumnNames: _col0, _col2
-                Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
                 HashTable Sink Operator
                   keys:
                     0 _col0 (type: smallint)
@@ -1058,14 +1058,14 @@ STAGE PLANS:
           TableScan
             alias: o1
             filterExpr: (csmallint is not null and cbigint is not null) (type: boolean)
-            Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: (csmallint is not null and cbigint is not null) (type: boolean)
-              Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: csmallint (type: smallint), cstring2 (type: string)
                 outputColumnNames: _col0, _col2
-                Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
                 Map Join Operator
                   condition map:
                        Inner Join 0 to 1
@@ -1073,14 +1073,14 @@ STAGE PLANS:
                     0 _col0 (type: smallint)
                     1 _col0 (type: smallint)
                   outputColumnNames: _col2, _col5
-                  Statistics: Num rows: 15522 Data size: 1738515 Basic stats: COMPLETE Column stats: NONE
+                  Statistics: Num rows: 270336 Data size: 63975869 Basic stats: COMPLETE Column stats: NONE
                   Select Operator
                     expressions: _col2 (type: string), _col5 (type: string)
                     outputColumnNames: _col0, _col1
-                    Statistics: Num rows: 15522 Data size: 1738515 Basic stats: COMPLETE Column stats: NONE
+                    Statistics: Num rows: 270336 Data size: 63975869 Basic stats: COMPLETE Column stats: NONE
                     File Output Operator
                       compressed: false
-                      Statistics: Num rows: 15522 Data size: 1738515 Basic stats: COMPLETE Column stats: NONE
+                      Statistics: Num rows: 270336 Data size: 63975869 Basic stats: COMPLETE Column stats: NONE
                       table:
                           input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                           output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out b/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out
index 7b361b7..38321e9 100644
--- a/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out
+++ b/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out
@@ -135,11 +135,11 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: orc_pred
-            Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
             Select Operator
               expressions: hash(t) (type: int)
               outputColumnNames: _col0
-              Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
               Group By Operator
                 aggregations: sum(_col0)
                 mode: hash
@@ -183,11 +183,11 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: orc_pred
-            Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
             Select Operator
               expressions: hash(t) (type: int)
               outputColumnNames: _col0
-              Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
               Group By Operator
                 aggregations: sum(_col0)
                 mode: hash
@@ -311,14 +311,14 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: orc_pred
-            Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: ((t < 0) and (UDFToInteger(t) > -2)) (type: boolean)
-              Statistics: Num rows: 116 Data size: 34409 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 670 Data size: 2680 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: hash(t) (type: int)
                 outputColumnNames: _col0
-                Statistics: Num rows: 116 Data size: 34409 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 670 Data size: 2680 Basic stats: COMPLETE Column stats: NONE
                 Group By Operator
                   aggregations: sum(_col0)
                   mode: hash
@@ -369,14 +369,14 @@ STAGE PLANS:
           TableScan
             alias: orc_pred
             filterExpr: ((t < 0) and (UDFToInteger(t) > -2)) (type: boolean)
-            Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: ((t < 0) and (UDFToInteger(t) > -2)) (type: boolean)
-              Statistics: Num rows: 116 Data size: 34409 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 670 Data size: 2680 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: hash(t) (type: int)
                 outputColumnNames: _col0
-                Statistics: Num rows: 116 Data size: 34409 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 670 Data size: 2680 Basic stats: COMPLETE Column stats: NONE
                 Group By Operator
                   aggregations: sum(_col0)
                   mode: hash
@@ -458,17 +458,17 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: orc_pred
-            Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 232 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: ((t = -1) and s is not null and (s like 'bob%')) (type: boolean)
-              Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: -1 (type: tinyint), s (type: string)
                 outputColumnNames: _col0, _col1
-                Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE
                 File Output Operator
                   compressed: false
-                  Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE
+                  Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE
                   table:
                       input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                       output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -501,17 +501,17 @@ STAGE PLANS:
           TableScan
             alias: orc_pred
             filterExpr: ((t = -1) and s is not null and (s like 'bob%')) (type: boolean)
-            Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 232 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: ((t = -1) and s is not null and (s like 'bob%')) (type: boolean)
-              Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: -1 (type: tinyint), s (type: string)
                 outputColumnNames: _col0, _col1
-                Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE
                 File Output Operator
                   compressed: false
-                  Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE
+                  Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE
                   table:
                       input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                       output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -589,26 +589,26 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: orc_pred
-            Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 232 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: (s is not null and (s like 'bob%') and (not (t) IN (-1, -2, -3)) and t BETWEEN 25 AND 30) (type: boolean)
-              Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: t (type: tinyint), s (type: string)
                 outputColumnNames: _col0, _col1
-                Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
                 Reduce Output Operator
                   key expressions: _col0 (type: tinyint), _col1 (type: string)
                   sort order: ++
-                  Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+                  Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
       Reduce Operator Tree:
         Select Operator
           expressions: KEY.reducesinkkey0 (type: tinyint), KEY.reducesinkkey1 (type: string)
           outputColumnNames: _col0, _col1
-          Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+          Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
           File Output Operator
             compressed: false
-            Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
             table:
                 input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                 output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -645,26 +645,26 @@ STAGE PLANS:
           TableScan
             alias: orc_pred
             filterExpr: (s is not null and (s like 'bob%') and (not (t) IN (-1, -2, -3)) and t BETWEEN 25 AND 30) (type: boolean)
-            Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 232 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: (s is not null and (s like 'bob%') and (not (t) IN (-1, -2, -3)) and t BETWEEN 25 AND 30) (type: boolean)
-              Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: t (type: tinyint), s (type: string)
                 outputColumnNames: _col0, _col1
-                Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
                 Reduce Output Operator
                   key expressions: _col0 (type: tinyint), _col1 (type: string)
                   sort order: ++
-                  Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+                  Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
       Reduce Operator Tree:
         Select Operator
           expressions: KEY.reducesinkkey0 (type: tinyint), KEY.reducesinkkey1 (type: string)
           outputColumnNames: _col0, _col1
-          Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+          Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
           File Output Operator
             compressed: false
-            Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
             table:
                 input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                 output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -766,31 +766,31 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: orc_pred
-            Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 208 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: ((d >= 10.0) and (d < 12.0) and (s like '%son') and (t > 0) and si BETWEEN 300 AND 400 and (not (s like '%car%'))) (type: boolean)
-              Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: t (type: tinyint), si (type: smallint), d (type: double), s (type: string)
                 outputColumnNames: _col0, _col1, _col2, _col3
-                Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
                 Reduce Output Operator
                   key expressions: _col3 (type: string)
                   sort order: -
-                  Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
                   TopN Hash Memory Usage: 0.1
                   value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double)
       Reduce Operator Tree:
         Select Operator
           expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string)
           outputColumnNames: _col0, _col1, _col2, _col3
-          Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE
+          Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
           Limit
             Number of rows: 3
-            Statistics: Num rows: 3 Data size: 888 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
             File Output Operator
               compressed: false
-              Statistics: Num rows: 3 Data size: 888 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
               table:
                   input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                   output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -835,31 +835,31 @@ STAGE PLANS:
           TableScan
             alias: orc_pred
             filterExpr: ((d >= 10.0) and (d < 12.0) and (s like '%son') and (t > 0) and si BETWEEN 300 AND 400 and (not (s like '%car%'))) (type: boolean)
-            Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 208 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: ((d >= 10.0) and (d < 12.0) and (s like '%son') and (t > 0) and si BETWEEN 300 AND 400 and (not (s like '%car%'))) (type: boolean)
-              Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: t (type: tinyint), si (type: smallint), d (type: double), s (type: string)
                 outputColumnNames: _col0, _col1, _col2, _col3
-                Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
                 Reduce Output Operator
                   key expressions: _col3 (type: string)
                   sort order: -
-                  Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
                   TopN Hash Memory Usage: 0.1
                   value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double)
       Reduce Operator Tree:
         Select Operator
           expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string)
           outputColumnNames: _col0, _col1, _col2, _col3
-          Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE
+          Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
           Limit
             Number of rows: 3
-            Statistics: Num rows: 3 Data size: 888 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
             File Output Operator
               compressed: false
-              Statistics: Num rows: 3 Data size: 888 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
               table:
                   input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                   output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -970,28 +970,28 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: orc_pred
-            Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 208 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: ((t > 10) and (t <> 101) and (d >= 10.0) and (d < 12.0) and (s like '%son') and (not (s like '%car%')) and (t > 0) and si BETWEEN 300 AND 400) (type: boolean)
-              Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: t (type: tinyint), si (type: smallint), d (type: double), s (type: string)
                 outputColumnNames: _col0, _col1, _col2, _col3
-                Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
                 Reduce Output Operator
                   key expressions: _col3 (type: string)
                   sort order: -
-                  Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
                   TopN Hash Memory Usage: 0.1
                   value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double)
       Reduce Operator Tree:
         Select Operator
           expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string)
           outputColumnNames: _col0, _col1, _col2, _col3
-          Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+          Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
           Limit
             Number of rows: 3
-            Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
             File Output Operator
               compressed: false
               table:
@@ -1006,20 +1006,20 @@ STAGE PLANS:
             Reduce Output Operator
               key expressions: _col3 (type: string)
               sort order: -
-              Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
               TopN Hash Memory Usage: 0.1
               value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double)
       Reduce Operator Tree:
         Select Operator
           expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string)
           outputColumnNames: _col0, _col1, _col2, _col3
-          Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+          Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
           Limit
             Number of rows: 3
-            Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
             File Output Operator
               compressed: false
-              Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
               table:
                   input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                   output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -1069,28 +1069,28 @@ STAGE PLANS:
           TableScan
             alias: orc_pred
             filterExpr: ((t > 10) and (t <> 101) and (d >= 10.0) and (d < 12.0) and (s like '%son') and (not (s like '%car%')) and (t > 0) and si BETWEEN 300 AND 400) (type: boolean)
-            Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 208 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
             Filter Operator
               predicate: ((t > 10) and (t <> 101) and (d >= 10.0) and (d < 12.0) and (s like '%son') and (not (s like '%car%')) and (t > 0) and si BETWEEN 300 AND 400) (type: boolean)
-              Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
               Select Operator
                 expressions: t (type: tinyint), si (type: smallint), d (type: double), s (type: string)
                 outputColumnNames: _col0, _col1, _col2, _col3
-                Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
                 Reduce Output Operator
                   key expressions: _col3 (type: string)
                   sort order: -
-                  Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
                   TopN Hash Memory Usage: 0.1
                   value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double)
       Reduce Operator Tree:
         Select Operator
           expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string)
           outputColumnNames: _col0, _col1, _col2, _col3
-          Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+          Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
           Limit
             Number of rows: 3
-            Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
             File Output Operator
               compressed: false
               table:
@@ -1105,20 +1105,20 @@ STAGE PLANS:
             Reduce Output Operator
               key expressions: _col3 (type: string)
               sort order: -
-              Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
               TopN Hash Memory Usage: 0.1
               value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double)
       Reduce Operator Tree:
         Select Operator
           expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string)
           outputColumnNames: _col0, _col1, _col2, _col3
-          Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+          Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
           Limit
             Number of rows: 3
-            Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+            Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
             File Output Operator
               compressed: false
-              Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+              Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
               table:
                   input format: org.apache.hadoop.mapred.SequenceFileInputFormat
                   output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/parquet_array_null_element.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/parquet_array_null_element.q.out b/ql/src/test/results/clientpositive/parquet_array_null_element.q.out
index 387f01e..75d2d27 100644
--- a/ql/src/test/results/clientpositive/parquet_array_null_element.q.out
+++ b/ql/src/test/results/clientpositive/parquet_array_null_element.q.out
@@ -70,6 +70,11 @@ Retention:          	0
 #### A masked pattern was here ####
 Table Type:         	MANAGED_TABLE       	 
 Table Parameters:	 	 
+	COLUMN_STATS_ACCURATE	{\"BASIC_STATS\":\"true\"}
+	numFiles            	0                   
+	numRows             	0                   
+	rawDataSize         	0                   
+	totalSize           	0                   
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 

http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/parquet_create.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/parquet_create.q.out b/ql/src/test/results/clientpositive/parquet_create.q.out
index c6d33ff..cc4b735 100644
--- a/ql/src/test/results/clientpositive/parquet_create.q.out
+++ b/ql/src/test/results/clientpositive/parquet_create.q.out
@@ -73,6 +73,11 @@ Retention:          	0
 #### A masked pattern was here ####
 Table Type:         	MANAGED_TABLE       	 
 Table Parameters:	 	 
+	COLUMN_STATS_ACCURATE	{\"BASIC_STATS\":\"true\"}
+	numFiles            	0                   
+	numRows             	0                   
+	rawDataSize         	0                   
+	totalSize           	0                   
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 

http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/parquet_mixed_partition_formats.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/parquet_mixed_partition_formats.q.out b/ql/src/test/results/clientpositive/parquet_mixed_partition_formats.q.out
index d6affd6..e96aa80 100644
--- a/ql/src/test/results/clientpositive/parquet_mixed_partition_formats.q.out
+++ b/ql/src/test/results/clientpositive/parquet_mixed_partition_formats.q.out
@@ -126,6 +126,8 @@ Table:              	parquet_mixed_partition_formats
 #### A masked pattern was here ####
 Partition Parameters:	 	 
 	numFiles            	1                   
+	numRows             	0                   
+	rawDataSize         	0                   
 	totalSize           	2521                
 #### A masked pattern was here ####
 	 	 
@@ -247,6 +249,8 @@ Table:              	parquet_mixed_partition_formats
 #### A masked pattern was here ####
 Partition Parameters:	 	 
 	numFiles            	1                   
+	numRows             	0                   
+	rawDataSize         	0                   
 	totalSize           	2521                
 #### A masked pattern was here ####
 	 	 

http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/parquet_serde.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/parquet_serde.q.out b/ql/src/test/results/clientpositive/parquet_serde.q.out
index c1e594a..6d5f0f8 100644
--- a/ql/src/test/results/clientpositive/parquet_serde.q.out
+++ b/ql/src/test/results/clientpositive/parquet_serde.q.out
@@ -72,6 +72,8 @@ Table:              	parquet_mixed_fileformat
 #### A masked pattern was here ####
 Partition Parameters:	 	 
 	numFiles            	1                   
+	numRows             	0                   
+	rawDataSize         	0                   
 	totalSize           	36                  
 #### A masked pattern was here ####
 	 	 
@@ -173,6 +175,8 @@ Table:              	parquet_mixed_fileformat
 #### A masked pattern was here ####
 Partition Parameters:	 	 
 	numFiles            	1                   
+	numRows             	0                   
+	rawDataSize         	0                   
 	totalSize           	36                  
 #### A masked pattern was here ####
 	 	 

http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/part_inherit_tbl_props.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/part_inherit_tbl_props.q.out b/ql/src/test/results/clientpositive/part_inherit_tbl_props.q.out
index 82f4750..876d0db 100644
--- a/ql/src/test/results/clientpositive/part_inherit_tbl_props.q.out
+++ b/ql/src/test/results/clientpositive/part_inherit_tbl_props.q.out
@@ -38,8 +38,13 @@ Database:           	default
 Table:              	mytbl               	 
 #### A masked pattern was here ####
 Partition Parameters:	 	 
+	COLUMN_STATS_ACCURATE	{\"BASIC_STATS\":\"true\"}
 	a                   	myval               
 	b                   	yourval             
+	numFiles            	0                   
+	numRows             	0                   
+	rawDataSize         	0                   
+	totalSize           	0                   
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 

http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/part_inherit_tbl_props_empty.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/part_inherit_tbl_props_empty.q.out b/ql/src/test/results/clientpositive/part_inherit_tbl_props_empty.q.out
index e4b8003..95c2d6e 100644
--- a/ql/src/test/results/clientpositive/part_inherit_tbl_props_empty.q.out
+++ b/ql/src/test/results/clientpositive/part_inherit_tbl_props_empty.q.out
@@ -34,6 +34,11 @@ Database:           	default
 Table:              	mytbl               	 
 #### A masked pattern was here ####
 Partition Parameters:	 	 
+	COLUMN_STATS_ACCURATE	{\"BASIC_STATS\":\"true\"}
+	numFiles            	0                   
+	numRows             	0                   
+	rawDataSize         	0                   
+	totalSize           	0                   
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 

http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/part_inherit_tbl_props_with_star.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/part_inherit_tbl_props_with_star.q.out b/ql/src/test/results/clientpositive/part_inherit_tbl_props_with_star.q.out
index 106448a..0be588a 100644
--- a/ql/src/test/results/clientpositive/part_inherit_tbl_props_with_star.q.out
+++ b/ql/src/test/results/clientpositive/part_inherit_tbl_props_with_star.q.out
@@ -38,9 +38,14 @@ Database:           	default
 Table:              	mytbl               	 
 #### A masked pattern was here ####
 Partition Parameters:	 	 
+	COLUMN_STATS_ACCURATE	{\"BASIC_STATS\":\"true\"}
 	a                   	myval               
 	b                   	yourval             
 	c                   	noval               
+	numFiles            	0                   
+	numRows             	0                   
+	rawDataSize         	0                   
+	totalSize           	0                   
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 

http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/partition_coltype_literals.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/partition_coltype_literals.q.out b/ql/src/test/results/clientpositive/partition_coltype_literals.q.out
index bc159eb..b7a09d2 100644
--- a/ql/src/test/results/clientpositive/partition_coltype_literals.q.out
+++ b/ql/src/test/results/clientpositive/partition_coltype_literals.q.out
@@ -45,6 +45,11 @@ Database:           	default
 Table:              	partcoltypenum      	 
 #### A masked pattern was here ####
 Partition Parameters:	 	 
+	COLUMN_STATS_ACCURATE	{\"BASIC_STATS\":\"true\"}
+	numFiles            	0                   
+	numRows             	0                   
+	rawDataSize         	0                   
+	totalSize           	0                   
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -92,8 +97,11 @@ Database:           	default
 Table:              	partcoltypenum      	 
 #### A masked pattern was here ####
 Partition Parameters:	 	 
+	COLUMN_STATS_ACCURATE	{\"BASIC_STATS\":\"true\"}
 #### A masked pattern was here ####
 	numFiles            	0                   
+	numRows             	0                   
+	rawDataSize         	0                   
 	totalSize           	0                   
 #### A masked pattern was here ####
 	 	 
@@ -142,8 +150,11 @@ Database:           	default
 Table:              	partcoltypenum      	 
 #### A masked pattern was here ####
 Partition Parameters:	 	 
+	COLUMN_STATS_ACCURATE	{\"BASIC_STATS\":\"true\"}
 #### A masked pattern was here ####
 	numFiles            	0                   
+	numRows             	0                   
+	rawDataSize         	0                   
 	totalSize           	0                   
 #### A masked pattern was here ####
 	 	 
@@ -193,8 +204,11 @@ Database:           	default
 Table:              	partcoltypenum      	 
 #### A masked pattern was here ####
 Partition Parameters:	 	 
+	COLUMN_STATS_ACCURATE	{\"BASIC_STATS\":\"true\"}
 #### A masked pattern was here ####
 	numFiles            	0                   
+	numRows             	0                   
+	rawDataSize         	0                   
 	totalSize           	0                   
 #### A masked pattern was here ####
 	 	 

http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/pcr.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/pcr.q.out b/ql/src/test/results/clientpositive/pcr.q.out
index b53226e..7222617 100644
--- a/ql/src/test/results/clientpositive/pcr.q.out
+++ b/ql/src/test/results/clientpositive/pcr.q.out
@@ -3419,15 +3419,20 @@ STAGE PLANS:
                     input format: org.apache.hadoop.mapred.TextInputFormat
                     output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                     properties:
+                      COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
                       bucket_count -1
                       columns key,value
                       columns.comments 
                       columns.types int:string
 #### A masked pattern was here ####
                       name default.pcr_t2
+                      numFiles 0
+                      numRows 0
+                      rawDataSize 0
                       serialization.ddl struct pcr_t2 { i32 key, string value}
                       serialization.format 1
                       serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                      totalSize 0
 #### A masked pattern was here ####
                     serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                     name: default.pcr_t2
@@ -3449,15 +3454,20 @@ STAGE PLANS:
                     input format: org.apache.hadoop.mapred.TextInputFormat
                     output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                     properties:
+                      COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
                       bucket_count -1
                       columns key,value
                       columns.comments 
                       columns.types int:string
 #### A masked pattern was here ####
                       name default.pcr_t3
+                      numFiles 0
+                      numRows 0
+                      rawDataSize 0
                       serialization.ddl struct pcr_t3 { i32 key, string value}
                       serialization.format 1
                       serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                      totalSize 0
 #### A masked pattern was here ####
                     serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                     name: default.pcr_t3
@@ -3533,15 +3543,20 @@ STAGE PLANS:
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
+                COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
                 bucket_count -1
                 columns key,value
                 columns.comments 
                 columns.types int:string
 #### A masked pattern was here ####
                 name default.pcr_t2
+                numFiles 0
+                numRows 0
+                rawDataSize 0
                 serialization.ddl struct pcr_t2 { i32 key, string value}
                 serialization.format 1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                totalSize 0
 #### A masked pattern was here ####
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: default.pcr_t2
@@ -3564,15 +3579,20 @@ STAGE PLANS:
                   input format: org.apache.hadoop.mapred.TextInputFormat
                   output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                   properties:
+                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
                     bucket_count -1
                     columns key,value
                     columns.comments 
                     columns.types int:string
 #### A masked pattern was here ####
                     name default.pcr_t2
+                    numFiles 0
+                    numRows 0
+                    rawDataSize 0
                     serialization.ddl struct pcr_t2 { i32 key, string value}
                     serialization.format 1
                     serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                    totalSize 0
 #### A masked pattern was here ####
                   serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                   name: default.pcr_t2
@@ -3588,30 +3608,40 @@ STAGE PLANS:
             input format: org.apache.hadoop.mapred.TextInputFormat
             output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns key,value
               columns.comments 
               columns.types int:string
 #### A masked pattern was here ####
               name default.pcr_t2
+              numFiles 0
+              numRows 0
+              rawDataSize 0
               serialization.ddl struct pcr_t2 { i32 key, string value}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
+                COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
                 bucket_count -1
                 columns key,value
                 columns.comments 
                 columns.types int:string
 #### A masked pattern was here ####
                 name default.pcr_t2
+                numFiles 0
+                numRows 0
+                rawDataSize 0
                 serialization.ddl struct pcr_t2 { i32 key, string value}
                 serialization.format 1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                totalSize 0
 #### A masked pattern was here ####
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: default.pcr_t2
@@ -3633,15 +3663,20 @@ STAGE PLANS:
                   input format: org.apache.hadoop.mapred.TextInputFormat
                   output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                   properties:
+                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
                     bucket_count -1
                     columns key,value
                     columns.comments 
                     columns.types int:string
 #### A masked pattern was here ####
                     name default.pcr_t2
+                    numFiles 0
+                    numRows 0
+                    rawDataSize 0
                     serialization.ddl struct pcr_t2 { i32 key, string value}
                     serialization.format 1
                     serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                    totalSize 0
 #### A masked pattern was here ####
                   serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                   name: default.pcr_t2
@@ -3657,30 +3692,40 @@ STAGE PLANS:
             input format: org.apache.hadoop.mapred.TextInputFormat
             output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns key,value
               columns.comments 
               columns.types int:string
 #### A masked pattern was here ####
               name default.pcr_t2
+              numFiles 0
+              numRows 0
+              rawDataSize 0
               serialization.ddl struct pcr_t2 { i32 key, string value}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
+                COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
                 bucket_count -1
                 columns key,value
                 columns.comments 
                 columns.types int:string
 #### A masked pattern was here ####
                 name default.pcr_t2
+                numFiles 0
+                numRows 0
+                rawDataSize 0
                 serialization.ddl struct pcr_t2 { i32 key, string value}
                 serialization.format 1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                totalSize 0
 #### A masked pattern was here ####
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: default.pcr_t2
@@ -3712,15 +3757,20 @@ STAGE PLANS:
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
+                COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
                 bucket_count -1
                 columns key,value
                 columns.comments 
                 columns.types int:string
 #### A masked pattern was here ####
                 name default.pcr_t3
+                numFiles 0
+                numRows 0
+                rawDataSize 0
                 serialization.ddl struct pcr_t3 { i32 key, string value}
                 serialization.format 1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                totalSize 0
 #### A masked pattern was here ####
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: default.pcr_t3
@@ -3743,15 +3793,20 @@ STAGE PLANS:
                   input format: org.apache.hadoop.mapred.TextInputFormat
                   output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                   properties:
+                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
                     bucket_count -1
                     columns key,value
                     columns.comments 
                     columns.types int:string
 #### A masked pattern was here ####
                     name default.pcr_t3
+                    numFiles 0
+                    numRows 0
+                    rawDataSize 0
                     serialization.ddl struct pcr_t3 { i32 key, string value}
                     serialization.format 1
                     serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                    totalSize 0
 #### A masked pattern was here ####
                   serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                   name: default.pcr_t3
@@ -3767,30 +3822,40 @@ STAGE PLANS:
             input format: org.apache.hadoop.mapred.TextInputFormat
             output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns key,value
               columns.comments 
               columns.types int:string
 #### A masked pattern was here ####
               name default.pcr_t3
+              numFiles 0
+              numRows 0
+              rawDataSize 0
               serialization.ddl struct pcr_t3 { i32 key, string value}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
+                COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
                 bucket_count -1
                 columns key,value
                 columns.comments 
                 columns.types int:string
 #### A masked pattern was here ####
                 name default.pcr_t3
+                numFiles 0
+                numRows 0
+                rawDataSize 0
                 serialization.ddl struct pcr_t3 { i32 key, string value}
                 serialization.format 1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                totalSize 0
 #### A masked pattern was here ####
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: default.pcr_t3
@@ -3812,15 +3877,20 @@ STAGE PLANS:
                   input format: org.apache.hadoop.mapred.TextInputFormat
                   output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                   properties:
+                    COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
                     bucket_count -1
                     columns key,value
                     columns.comments 
                     columns.types int:string
 #### A masked pattern was here ####
                     name default.pcr_t3
+                    numFiles 0
+                    numRows 0
+                    rawDataSize 0
                     serialization.ddl struct pcr_t3 { i32 key, string value}
                     serialization.format 1
                     serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                    totalSize 0
 #### A masked pattern was here ####
                   serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                   name: default.pcr_t3
@@ -3836,30 +3906,40 @@ STAGE PLANS:
             input format: org.apache.hadoop.mapred.TextInputFormat
             output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
             properties:
+              COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
               bucket_count -1
               columns key,value
               columns.comments 
               columns.types int:string
 #### A masked pattern was here ####
               name default.pcr_t3
+              numFiles 0
+              numRows 0
+              rawDataSize 0
               serialization.ddl struct pcr_t3 { i32 key, string value}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+              totalSize 0
 #### A masked pattern was here ####
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
+                COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
                 bucket_count -1
                 columns key,value
                 columns.comments 
                 columns.types int:string
 #### A masked pattern was here ####
                 name default.pcr_t3
+                numFiles 0
+                numRows 0
+                rawDataSize 0
                 serialization.ddl struct pcr_t3 { i32 key, string value}
                 serialization.format 1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                totalSize 0
 #### A masked pattern was here ####
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: default.pcr_t3


Mime
View raw message