hive-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From hashut...@apache.org
Subject [22/51] [partial] hive git commit: HIVE-13567 : Enable auto-gather column stats by default (Zoltan Haindrich, Pengcheng Xiong via Ashutosh Chauhan)
Date Tue, 12 Dec 2017 00:00:20 GMT
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/llap/dp_counter_non_mm.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/dp_counter_non_mm.q.out b/ql/src/test/results/clientpositive/llap/dp_counter_non_mm.q.out
index 85f9b9d..e2ff9a6 100644
--- a/ql/src/test/results/clientpositive/llap/dp_counter_non_mm.q.out
+++ b/ql/src/test/results/clientpositive/llap/dp_counter_non_mm.q.out
@@ -17,10 +17,12 @@ PREHOOK: Output: default@src2
 Stage-1 FILE SYSTEM COUNTERS:
 Stage-1 HIVE COUNTERS:
    CREATED_DYNAMIC_PARTITIONS: 57
-   CREATED_FILES: 57
+   CREATED_FILES: 61
    DESERIALIZE_ERRORS: 0
    RECORDS_IN_Map_1: 500
+   RECORDS_OUT_0: 57
    RECORDS_OUT_1_default.src2: 84
+   RECORDS_OUT_INTERMEDIATE_Map_1: 57
 PREHOOK: query: insert into table src2 partition (value) select * from src where key <
200
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
@@ -28,10 +30,12 @@ PREHOOK: Output: default@src2
 Stage-1 FILE SYSTEM COUNTERS:
 Stage-1 HIVE COUNTERS:
    CREATED_DYNAMIC_PARTITIONS: 64
-   CREATED_FILES: 121
+   CREATED_FILES: 125
    DESERIALIZE_ERRORS: 0
    RECORDS_IN_Map_1: 500
+   RECORDS_OUT_0: 121
    RECORDS_OUT_1_default.src2: 189
+   RECORDS_OUT_INTERMEDIATE_Map_1: 121
 PREHOOK: query: drop table src2
 PREHOOK: type: DROPTABLE
 PREHOOK: Input: default@src2
@@ -47,10 +51,12 @@ PREHOOK: Output: default@src2
 Stage-1 FILE SYSTEM COUNTERS:
 Stage-1 HIVE COUNTERS:
    CREATED_DYNAMIC_PARTITIONS: 121
-   CREATED_FILES: 121
+   CREATED_FILES: 125
    DESERIALIZE_ERRORS: 0
    RECORDS_IN_Map_1: 500
+   RECORDS_OUT_0: 121
    RECORDS_OUT_1_default.src2: 189
+   RECORDS_OUT_INTERMEDIATE_Map_1: 121
 PREHOOK: query: insert into table src2 partition (value) select * from src where key <
300
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
@@ -58,10 +64,12 @@ PREHOOK: Output: default@src2
 Stage-1 FILE SYSTEM COUNTERS:
 Stage-1 HIVE COUNTERS:
    CREATED_DYNAMIC_PARTITIONS: 63
-   CREATED_FILES: 184
+   CREATED_FILES: 188
    DESERIALIZE_ERRORS: 0
    RECORDS_IN_Map_1: 500
+   RECORDS_OUT_0: 184
    RECORDS_OUT_1_default.src2: 292
+   RECORDS_OUT_INTERMEDIATE_Map_1: 184
 PREHOOK: query: drop table src2
 PREHOOK: type: DROPTABLE
 PREHOOK: Input: default@src2
@@ -86,11 +94,13 @@ PREHOOK: Output: default@src3
 Stage-2 FILE SYSTEM COUNTERS:
 Stage-2 HIVE COUNTERS:
    CREATED_DYNAMIC_PARTITIONS: 121
-   CREATED_FILES: 121
+   CREATED_FILES: 129
    DESERIALIZE_ERRORS: 0
    RECORDS_IN_Map_1: 500
+   RECORDS_OUT_0: 121
    RECORDS_OUT_1_default.src2: 84
    RECORDS_OUT_2_default.src3: 105
+   RECORDS_OUT_INTERMEDIATE_Map_1: 121
 PREHOOK: query: from src
 insert into table src2 partition (value) select * where key < 100
 insert into table src3 partition (value) select * where key >= 100 and key < 300
@@ -101,11 +111,13 @@ PREHOOK: Output: default@src3
 Stage-2 FILE SYSTEM COUNTERS:
 Stage-2 HIVE COUNTERS:
    CREATED_DYNAMIC_PARTITIONS: 63
-   CREATED_FILES: 184
+   CREATED_FILES: 192
    DESERIALIZE_ERRORS: 0
    RECORDS_IN_Map_1: 500
+   RECORDS_OUT_0: 184
    RECORDS_OUT_1_default.src2: 84
    RECORDS_OUT_2_default.src3: 208
+   RECORDS_OUT_INTERMEDIATE_Map_1: 184
 PREHOOK: query: drop table src2
 PREHOOK: type: DROPTABLE
 PREHOOK: Input: default@src2
@@ -125,11 +137,14 @@ PREHOOK: Output: default@src2
 Stage-1 FILE SYSTEM COUNTERS:
 Stage-1 HIVE COUNTERS:
    CREATED_DYNAMIC_PARTITIONS: 121
-   CREATED_FILES: 121
+   CREATED_FILES: 125
    DESERIALIZE_ERRORS: 0
    RECORDS_IN_Map_1: 500
-   RECORDS_IN_Map_3: 500
+   RECORDS_IN_Map_4: 500
+   RECORDS_OUT_0: 121
    RECORDS_OUT_1_default.src2: 189
+   RECORDS_OUT_INTERMEDIATE_Map_1: 57
+   RECORDS_OUT_INTERMEDIATE_Map_4: 64
 PREHOOK: query: insert into table src2 partition (value)
 select temps.* from (
   select * from src where key < 100
@@ -141,8 +156,11 @@ PREHOOK: Output: default@src2
 Stage-1 FILE SYSTEM COUNTERS:
 Stage-1 HIVE COUNTERS:
    CREATED_DYNAMIC_PARTITIONS: 63
-   CREATED_FILES: 184
+   CREATED_FILES: 188
    DESERIALIZE_ERRORS: 0
    RECORDS_IN_Map_1: 500
-   RECORDS_IN_Map_3: 500
+   RECORDS_IN_Map_4: 500
+   RECORDS_OUT_0: 184
    RECORDS_OUT_1_default.src2: 292
+   RECORDS_OUT_INTERMEDIATE_Map_1: 57
+   RECORDS_OUT_INTERMEDIATE_Map_4: 127

http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction_3.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction_3.q.out b/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction_3.q.out
index 6772e5d..91557e8 100644
--- a/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction_3.q.out
+++ b/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction_3.q.out
@@ -41,13 +41,14 @@ STAGE PLANS:
     Tez
 #### A masked pattern was here ####
       Edges:
-        Map 1 <- Reducer 8 (BROADCAST_EDGE)
-        Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 7 (SIMPLE_EDGE)
+        Map 1 <- Reducer 9 (BROADCAST_EDGE)
+        Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 8 (SIMPLE_EDGE)
         Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
         Reducer 4 <- Reducer 2 (SIMPLE_EDGE)
         Reducer 5 <- Reducer 2 (SIMPLE_EDGE)
         Reducer 6 <- Reducer 2 (CUSTOM_SIMPLE_EDGE)
-        Reducer 8 <- Map 7 (CUSTOM_SIMPLE_EDGE)
+        Reducer 7 <- Reducer 6 (CUSTOM_SIMPLE_EDGE)
+        Reducer 9 <- Map 8 (CUSTOM_SIMPLE_EDGE)
 #### A masked pattern was here ####
       Vertices:
         Map 1 
@@ -66,7 +67,7 @@ STAGE PLANS:
                       value expressions: ROW__ID (type: struct<transactionid:bigint,bucketid:int,rowid:bigint>)
             Execution mode: llap
             LLAP IO: may be used (ACID table)
-        Map 7 
+        Map 8 
             Map Operator Tree:
                 TableScan
                   alias: s
@@ -215,6 +216,26 @@ STAGE PLANS:
                           output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                           serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                           name: default.merge_tmp_table
+                    Select Operator
+                      expressions: _col0 (type: int)
+                      outputColumnNames: val
+                      Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats:
NONE
+                      Group By Operator
+                        aggregations: compute_stats(val, 'hll')
+                        mode: complete
+                        outputColumnNames: _col0
+                        Statistics: Num rows: 1 Data size: 432 Basic stats: COMPLETE Column
stats: NONE
+                        Select Operator
+                          expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>)
+                          outputColumnNames: _col0
+                          Statistics: Num rows: 1 Data size: 432 Basic stats: COMPLETE Column
stats: NONE
+                          File Output Operator
+                            compressed: false
+                            Statistics: Num rows: 1 Data size: 432 Basic stats: COMPLETE
Column stats: NONE
+                            table:
+                                input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+                                output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
         Reducer 6 
             Execution mode: llap
             Reduce Operator Tree:
@@ -231,7 +252,35 @@ STAGE PLANS:
                       serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
                       name: default.acidtbl
                   Write Type: INSERT
-        Reducer 8 
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int)
+                  outputColumnNames: a, b
+                  Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats:
NONE
+                  Group By Operator
+                    aggregations: compute_stats(a, 'hll'), compute_stats(b, 'hll')
+                    mode: hash
+                    outputColumnNames: _col0, _col1
+                    Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats:
NONE
+                    Reduce Output Operator
+                      sort order: 
+                      Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column
stats: NONE
+                      value expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>),
_col1 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>)
+        Reducer 7 
+            Execution mode: llap
+            Reduce Operator Tree:
+              Group By Operator
+                aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1)
+                mode: mergepartial
+                outputColumnNames: _col0, _col1
+                Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats:
NONE
+                File Output Operator
+                  compressed: false
+                  Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats:
NONE
+                  table:
+                      input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+                      output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+        Reducer 9 
             Execution mode: llap
             Reduce Operator Tree:
               Group By Operator
@@ -290,6 +339,10 @@ STAGE PLANS:
   Stage: Stage-8
     Stats Work
       Basic Stats Work:
+      Column Stats Desc:
+          Columns: val
+          Column Types: int
+          Table: default.merge_tmp_table
 
   Stage: Stage-1
     Move Operator
@@ -305,6 +358,10 @@ STAGE PLANS:
   Stage: Stage-9
     Stats Work
       Basic Stats Work:
+      Column Stats Desc:
+          Columns: a, b
+          Column Types: int, int
+          Table: default.acidtbl
 
 PREHOOK: query: explain merge into acidTbl as t using nonAcidOrcTbl s ON t.a = s.a
 WHEN NOT MATCHED THEN INSERT VALUES(s.a, s.b)
@@ -323,10 +380,11 @@ STAGE PLANS:
     Tez
 #### A masked pattern was here ####
       Edges:
-        Map 1 <- Reducer 5 (BROADCAST_EDGE)
-        Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE)
+        Map 1 <- Reducer 6 (BROADCAST_EDGE)
+        Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE)
         Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE)
-        Reducer 5 <- Map 4 (CUSTOM_SIMPLE_EDGE)
+        Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE)
+        Reducer 6 <- Map 5 (CUSTOM_SIMPLE_EDGE)
 #### A masked pattern was here ####
       Vertices:
         Map 1 
@@ -344,7 +402,7 @@ STAGE PLANS:
                       Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats:
NONE
             Execution mode: llap
             LLAP IO: may be used (ACID table)
-        Map 4 
+        Map 5 
             Map Operator Tree:
                 TableScan
                   alias: s
@@ -409,7 +467,35 @@ STAGE PLANS:
                       serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
                       name: default.acidtbl
                   Write Type: INSERT
-        Reducer 5 
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int)
+                  outputColumnNames: a, b
+                  Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats:
NONE
+                  Group By Operator
+                    aggregations: compute_stats(a, 'hll'), compute_stats(b, 'hll')
+                    mode: hash
+                    outputColumnNames: _col0, _col1
+                    Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats:
NONE
+                    Reduce Output Operator
+                      sort order: 
+                      Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column
stats: NONE
+                      value expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>),
_col1 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>)
+        Reducer 4 
+            Execution mode: llap
+            Reduce Operator Tree:
+              Group By Operator
+                aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1)
+                mode: mergepartial
+                outputColumnNames: _col0, _col1
+                Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats:
NONE
+                File Output Operator
+                  compressed: false
+                  Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats:
NONE
+                  table:
+                      input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+                      output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+        Reducer 6 
             Execution mode: llap
             Reduce Operator Tree:
               Group By Operator
@@ -439,6 +525,10 @@ STAGE PLANS:
   Stage: Stage-3
     Stats Work
       Basic Stats Work:
+      Column Stats Desc:
+          Columns: a, b
+          Column Types: int, int
+          Table: default.acidtbl
 
 PREHOOK: query: explain merge into acidTbl as t using (
   select * from nonAcidOrcTbl where a > 0
@@ -476,14 +566,15 @@ STAGE PLANS:
 #### A masked pattern was here ####
       Edges:
         Map 1 <- Union 2 (CONTAINS)
-        Map 10 <- Reducer 8 (BROADCAST_EDGE)
-        Map 9 <- Union 2 (CONTAINS)
-        Reducer 3 <- Map 10 (SIMPLE_EDGE), Union 2 (SIMPLE_EDGE)
+        Map 10 <- Union 2 (CONTAINS)
+        Map 11 <- Reducer 9 (BROADCAST_EDGE)
+        Reducer 3 <- Map 11 (SIMPLE_EDGE), Union 2 (SIMPLE_EDGE)
         Reducer 4 <- Reducer 3 (SIMPLE_EDGE)
         Reducer 5 <- Reducer 3 (SIMPLE_EDGE)
         Reducer 6 <- Reducer 3 (SIMPLE_EDGE)
         Reducer 7 <- Reducer 3 (CUSTOM_SIMPLE_EDGE)
-        Reducer 8 <- Union 2 (CUSTOM_SIMPLE_EDGE)
+        Reducer 8 <- Reducer 7 (CUSTOM_SIMPLE_EDGE)
+        Reducer 9 <- Union 2 (CUSTOM_SIMPLE_EDGE)
 #### A masked pattern was here ####
       Vertices:
         Map 1 
@@ -523,22 +614,6 @@ STAGE PLANS:
         Map 10 
             Map Operator Tree:
                 TableScan
-                  alias: t
-                  Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats:
NONE
-                  Filter Operator
-                    predicate: (a BETWEEN DynamicValue(RS_10_nonacidorctbl__col0_min) AND
DynamicValue(RS_10_nonacidorctbl__col0_max) and in_bloom_filter(a, DynamicValue(RS_10_nonacidorctbl__col0_bloom_filter)))
(type: boolean)
-                    Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats:
NONE
-                    Reduce Output Operator
-                      key expressions: a (type: int)
-                      sort order: +
-                      Map-reduce partition columns: a (type: int)
-                      Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats:
NONE
-                      value expressions: ROW__ID (type: struct<transactionid:bigint,bucketid:int,rowid:bigint>)
-            Execution mode: llap
-            LLAP IO: may be used (ACID table)
-        Map 9 
-            Map Operator Tree:
-                TableScan
                   alias: nonacidorctbl
                   filterExpr: (b > 0) (type: boolean)
                   Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats:
NONE
@@ -570,6 +645,22 @@ STAGE PLANS:
                             value expressions: _col0 (type: int), _col1 (type: int), _col2
(type: binary)
             Execution mode: llap
             LLAP IO: all inputs
+        Map 11 
+            Map Operator Tree:
+                TableScan
+                  alias: t
+                  Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats:
NONE
+                  Filter Operator
+                    predicate: (a BETWEEN DynamicValue(RS_10_nonacidorctbl__col0_min) AND
DynamicValue(RS_10_nonacidorctbl__col0_max) and in_bloom_filter(a, DynamicValue(RS_10_nonacidorctbl__col0_bloom_filter)))
(type: boolean)
+                    Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats:
NONE
+                    Reduce Output Operator
+                      key expressions: a (type: int)
+                      sort order: +
+                      Map-reduce partition columns: a (type: int)
+                      Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats:
NONE
+                      value expressions: ROW__ID (type: struct<transactionid:bigint,bucketid:int,rowid:bigint>)
+            Execution mode: llap
+            LLAP IO: may be used (ACID table)
         Reducer 3 
             Execution mode: llap
             Reduce Operator Tree:
@@ -693,6 +784,26 @@ STAGE PLANS:
                           output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                           serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                           name: default.merge_tmp_table
+                    Select Operator
+                      expressions: _col0 (type: int)
+                      outputColumnNames: val
+                      Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats:
NONE
+                      Group By Operator
+                        aggregations: compute_stats(val, 'hll')
+                        mode: complete
+                        outputColumnNames: _col0
+                        Statistics: Num rows: 1 Data size: 432 Basic stats: COMPLETE Column
stats: NONE
+                        Select Operator
+                          expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>)
+                          outputColumnNames: _col0
+                          Statistics: Num rows: 1 Data size: 432 Basic stats: COMPLETE Column
stats: NONE
+                          File Output Operator
+                            compressed: false
+                            Statistics: Num rows: 1 Data size: 432 Basic stats: COMPLETE
Column stats: NONE
+                            table:
+                                input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+                                output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
         Reducer 7 
             Execution mode: llap
             Reduce Operator Tree:
@@ -709,10 +820,38 @@ STAGE PLANS:
                       serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde
                       name: default.acidtbl
                   Write Type: INSERT
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int)
+                  outputColumnNames: a, b
+                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats:
NONE
+                  Group By Operator
+                    aggregations: compute_stats(a, 'hll'), compute_stats(b, 'hll')
+                    mode: hash
+                    outputColumnNames: _col0, _col1
+                    Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats:
NONE
+                    Reduce Output Operator
+                      sort order: 
+                      Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column
stats: NONE
+                      value expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>),
_col1 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>)
         Reducer 8 
             Execution mode: llap
             Reduce Operator Tree:
               Group By Operator
+                aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1)
+                mode: mergepartial
+                outputColumnNames: _col0, _col1
+                Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats:
NONE
+                File Output Operator
+                  compressed: false
+                  Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats:
NONE
+                  table:
+                      input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+                      output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+        Reducer 9 
+            Execution mode: llap
+            Reduce Operator Tree:
+              Group By Operator
                 aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2,
expectedEntries=2)
                 mode: final
                 outputColumnNames: _col0, _col1, _col2
@@ -770,6 +909,10 @@ STAGE PLANS:
   Stage: Stage-8
     Stats Work
       Basic Stats Work:
+      Column Stats Desc:
+          Columns: val
+          Column Types: int
+          Table: default.merge_tmp_table
 
   Stage: Stage-1
     Move Operator
@@ -785,6 +928,10 @@ STAGE PLANS:
   Stage: Stage-9
     Stats Work
       Basic Stats Work:
+      Column Stats Desc:
+          Columns: a, b
+          Column Types: int, int
+          Table: default.acidtbl
 
 PREHOOK: query: drop database if exists type2_scd_helper cascade
 PREHOOK: type: DROPDATABASE

http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction_sw.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction_sw.q.out b/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction_sw.q.out
index d6103a8..81b6bde 100644
--- a/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction_sw.q.out
+++ b/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction_sw.q.out
@@ -319,49 +319,49 @@ STAGE PLANS:
                 TableScan
                   alias: srcpart_small
                   filterExpr: key1 is not null (type: boolean)
-                  Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats:
NONE
+                  Statistics: Num rows: 1 Data size: 87 Basic stats: PARTIAL Column stats:
COMPLETE
                   Filter Operator
                     predicate: key1 is not null (type: boolean)
-                    Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats:
NONE
+                    Statistics: Num rows: 1 Data size: 87 Basic stats: PARTIAL Column stats:
COMPLETE
                     Select Operator
                       expressions: key1 (type: string)
                       outputColumnNames: _col0
-                      Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column
stats: NONE
+                      Statistics: Num rows: 1 Data size: 87 Basic stats: PARTIAL Column stats:
COMPLETE
                       Reduce Output Operator
                         key expressions: _col0 (type: string)
                         sort order: +
                         Map-reduce partition columns: _col0 (type: string)
-                        Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column
stats: NONE
+                        Statistics: Num rows: 1 Data size: 87 Basic stats: PARTIAL Column
stats: COMPLETE
                       Select Operator
                         expressions: _col0 (type: string)
                         outputColumnNames: _col0
-                        Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column
stats: NONE
+                        Statistics: Num rows: 1 Data size: 87 Basic stats: PARTIAL Column
stats: COMPLETE
                         Group By Operator
                           aggregations: min(_col0), max(_col0), bloom_filter(_col0, expectedEntries=1)
                           mode: hash
                           outputColumnNames: _col0, _col1, _col2
-                          Statistics: Num rows: 1 Data size: 736 Basic stats: PARTIAL Column
stats: NONE
+                          Statistics: Num rows: 1 Data size: 639 Basic stats: PARTIAL Column
stats: COMPLETE
                           Reduce Output Operator
                             sort order: 
-                            Statistics: Num rows: 1 Data size: 736 Basic stats: PARTIAL Column
stats: NONE
+                            Statistics: Num rows: 1 Data size: 639 Basic stats: PARTIAL Column
stats: COMPLETE
                             value expressions: _col0 (type: string), _col1 (type: string),
_col2 (type: binary)
                       Reduce Output Operator
                         key expressions: _col0 (type: string)
                         sort order: +
                         Map-reduce partition columns: _col0 (type: string)
-                        Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column
stats: NONE
+                        Statistics: Num rows: 1 Data size: 87 Basic stats: PARTIAL Column
stats: COMPLETE
                       Select Operator
                         expressions: _col0 (type: string)
                         outputColumnNames: _col0
-                        Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column
stats: NONE
+                        Statistics: Num rows: 1 Data size: 87 Basic stats: PARTIAL Column
stats: COMPLETE
                         Group By Operator
                           aggregations: min(_col0), max(_col0), bloom_filter(_col0, expectedEntries=1)
                           mode: hash
                           outputColumnNames: _col0, _col1, _col2
-                          Statistics: Num rows: 1 Data size: 736 Basic stats: PARTIAL Column
stats: NONE
+                          Statistics: Num rows: 1 Data size: 639 Basic stats: PARTIAL Column
stats: COMPLETE
                           Reduce Output Operator
                             sort order: 
-                            Statistics: Num rows: 1 Data size: 736 Basic stats: PARTIAL Column
stats: NONE
+                            Statistics: Num rows: 1 Data size: 639 Basic stats: PARTIAL Column
stats: COMPLETE
                             value expressions: _col0 (type: string), _col1 (type: string),
_col2 (type: binary)
             Execution mode: llap
             LLAP IO: all inputs
@@ -445,14 +445,14 @@ STAGE PLANS:
                 aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2,
expectedEntries=1)
                 mode: final
                 outputColumnNames: _col0, _col1, _col2
-                Statistics: Num rows: 1 Data size: 736 Basic stats: PARTIAL Column stats:
NONE
+                Statistics: Num rows: 1 Data size: 639 Basic stats: PARTIAL Column stats:
COMPLETE
                 Reduce Output Operator
                   sort order: 
-                  Statistics: Num rows: 1 Data size: 736 Basic stats: PARTIAL Column stats:
NONE
+                  Statistics: Num rows: 1 Data size: 639 Basic stats: PARTIAL Column stats:
COMPLETE
                   value expressions: _col0 (type: string), _col1 (type: string), _col2 (type:
binary)
                 Reduce Output Operator
                   sort order: 
-                  Statistics: Num rows: 1 Data size: 736 Basic stats: PARTIAL Column stats:
NONE
+                  Statistics: Num rows: 1 Data size: 639 Basic stats: PARTIAL Column stats:
COMPLETE
                   value expressions: _col0 (type: string), _col1 (type: string), _col2 (type:
binary)
         Reducer 8 
             Execution mode: llap
@@ -476,14 +476,14 @@ STAGE PLANS:
                 aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2,
expectedEntries=1)
                 mode: final
                 outputColumnNames: _col0, _col1, _col2
-                Statistics: Num rows: 1 Data size: 736 Basic stats: PARTIAL Column stats:
NONE
+                Statistics: Num rows: 1 Data size: 639 Basic stats: PARTIAL Column stats:
COMPLETE
                 Reduce Output Operator
                   sort order: 
-                  Statistics: Num rows: 1 Data size: 736 Basic stats: PARTIAL Column stats:
NONE
+                  Statistics: Num rows: 1 Data size: 639 Basic stats: PARTIAL Column stats:
COMPLETE
                   value expressions: _col0 (type: string), _col1 (type: string), _col2 (type:
binary)
                 Reduce Output Operator
                   sort order: 
-                  Statistics: Num rows: 1 Data size: 736 Basic stats: PARTIAL Column stats:
NONE
+                  Statistics: Num rows: 1 Data size: 639 Basic stats: PARTIAL Column stats:
COMPLETE
                   value expressions: _col0 (type: string), _col1 (type: string), _col2 (type:
binary)
 
   Stage: Stage-0


Mime
View raw message