hive-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From hashut...@apache.org
Subject svn commit: r1449981 [7/19] - in /hive/trunk: contrib/src/test/results/clientpositive/ hbase-handler/src/test/results/positive/ ql/src/java/org/apache/hadoop/hive/ql/exec/ ql/src/java/org/apache/hadoop/hive/ql/optimizer/ ql/src/java/org/apache/hadoop/h...
Date Tue, 26 Feb 2013 00:27:37 GMT
Modified: hive/trunk/ql/src/test/results/clientpositive/groupby_sort_1.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/groupby_sort_1.q.out?rev=1449981&r1=1449980&r2=1449981&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/groupby_sort_1.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/groupby_sort_1.q.out Tue Feb 26 00:27:32 2013
@@ -83,42 +83,35 @@ STAGE PLANS:
                 outputColumnNames: _col0, _col1
                 Select Operator
                   expressions:
-                        expr: _col0
-                        type: string
-                        expr: _col1
-                        type: bigint
+                        expr: UDFToInteger(_col0)
+                        type: int
+                        expr: UDFToInteger(_col1)
+                        type: int
                   outputColumnNames: _col0, _col1
-                  Select Operator
-                    expressions:
-                          expr: UDFToInteger(_col0)
-                          type: int
-                          expr: UDFToInteger(_col1)
-                          type: int
-                    outputColumnNames: _col0, _col1
-                    File Output Operator
-                      compressed: false
-                      GlobalTableId: 1
-#### A masked pattern was here ####
-                      NumFilesPerFileSink: 1
-#### A masked pattern was here ####
-                      table:
-                          input format: org.apache.hadoop.mapred.TextInputFormat
-                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                          properties:
-                            bucket_count -1
-                            columns key,cnt
-                            columns.types int:int
-#### A masked pattern was here ####
-                            name default.outputtbl1
-                            serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
-                            serialization.format 1
-                            serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                          name: default.outputtbl1
-                      TotalFiles: 1
-                      GatherStats: true
-                      MultiFileSpray: false
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 1
+#### A masked pattern was here ####
+                    NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        properties:
+                          bucket_count -1
+                          columns key,cnt
+                          columns.types int:int
+#### A masked pattern was here ####
+                          name default.outputtbl1
+                          serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
+                          serialization.format 1
+                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+#### A masked pattern was here ####
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: default.outputtbl1
+                    TotalFiles: 1
+                    GatherStats: true
+                    MultiFileSpray: false
       Needs Tagging: false
       Path -> Alias:
 #### A masked pattern was here ####
@@ -514,46 +507,37 @@ STAGE PLANS:
           outputColumnNames: _col0, _col1, _col2
           Select Operator
             expressions:
-                  expr: _col0
-                  type: string
+                  expr: UDFToInteger(_col0)
+                  type: int
                   expr: _col1
                   type: string
-                  expr: _col2
-                  type: bigint
+                  expr: UDFToInteger(_col2)
+                  type: int
             outputColumnNames: _col0, _col1, _col2
-            Select Operator
-              expressions:
-                    expr: UDFToInteger(_col0)
-                    type: int
-                    expr: _col1
-                    type: string
-                    expr: UDFToInteger(_col2)
-                    type: int
-              outputColumnNames: _col0, _col1, _col2
-              File Output Operator
-                compressed: false
-                GlobalTableId: 1
-#### A masked pattern was here ####
-                NumFilesPerFileSink: 1
-#### A masked pattern was here ####
-                table:
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns key1,key2,cnt
-                      columns.types int:string:int
-#### A masked pattern was here ####
-                      name default.outputtbl2
-                      serialization.ddl struct outputtbl2 { i32 key1, string key2, i32 cnt}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.outputtbl2
-                TotalFiles: 1
-                GatherStats: true
-                MultiFileSpray: false
+            File Output Operator
+              compressed: false
+              GlobalTableId: 1
+#### A masked pattern was here ####
+              NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+              table:
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  properties:
+                    bucket_count -1
+                    columns key1,key2,cnt
+                    columns.types int:string:int
+#### A masked pattern was here ####
+                    name default.outputtbl2
+                    serialization.ddl struct outputtbl2 { i32 key1, string key2, i32 cnt}
+                    serialization.format 1
+                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+#### A masked pattern was here ####
+                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  name: default.outputtbl2
+              TotalFiles: 1
+              GatherStats: true
+              MultiFileSpray: false
       Truncated Path -> Alias:
         /t1 [t1]
 
@@ -665,63 +649,51 @@ STAGE PLANS:
                     expr: key
                     type: string
               outputColumnNames: _col0
-              Select Operator
-                expressions:
+              Group By Operator
+                aggregations:
+                      expr: count(1)
+                bucketGroup: false
+                keys:
                       expr: _col0
                       type: string
-                outputColumnNames: _col0
-                Group By Operator
-                  aggregations:
-                        expr: count(1)
-                  bucketGroup: false
-                  keys:
-                        expr: _col0
-                        type: string
-                  mode: final
+                mode: final
+                outputColumnNames: _col0, _col1
+                Select Operator
+                  expressions:
+                        expr: UDFToInteger(_col0)
+                        type: int
+                        expr: UDFToInteger(_col1)
+                        type: int
                   outputColumnNames: _col0, _col1
-                  Select Operator
-                    expressions:
-                          expr: _col0
-                          type: string
-                          expr: _col1
-                          type: bigint
-                    outputColumnNames: _col0, _col1
-                    Select Operator
-                      expressions:
-                            expr: UDFToInteger(_col0)
-                            type: int
-                            expr: UDFToInteger(_col1)
-                            type: int
-                      outputColumnNames: _col0, _col1
-                      File Output Operator
-                        compressed: false
-                        GlobalTableId: 1
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 1
 #### A masked pattern was here ####
-                        NumFilesPerFileSink: 1
+                    NumFilesPerFileSink: 1
 #### A masked pattern was here ####
-                        table:
-                            input format: org.apache.hadoop.mapred.TextInputFormat
-                            output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                            properties:
-                              bucket_count -1
-                              columns key,cnt
-                              columns.types int:int
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        properties:
+                          bucket_count -1
+                          columns key,cnt
+                          columns.types int:int
 #### A masked pattern was here ####
-                              name default.outputtbl1
-                              numFiles 1
-                              numPartitions 0
-                              numRows 5
-                              rawDataSize 15
-                              serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
-                              serialization.format 1
-                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                              totalSize 20
+                          name default.outputtbl1
+                          numFiles 1
+                          numPartitions 0
+                          numRows 5
+                          rawDataSize 15
+                          serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
+                          serialization.format 1
+                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                          totalSize 20
 #### A masked pattern was here ####
-                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                            name: default.outputtbl1
-                        TotalFiles: 1
-                        GatherStats: true
-                        MultiFileSpray: false
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: default.outputtbl1
+                    TotalFiles: 1
+                    GatherStats: true
+                    MultiFileSpray: false
       Needs Tagging: false
       Path -> Alias:
 #### A masked pattern was here ####
@@ -1070,63 +1042,51 @@ STAGE PLANS:
                     expr: key
                     type: string
               outputColumnNames: _col0
-              Select Operator
-                expressions:
+              Group By Operator
+                aggregations:
+                      expr: count(1)
+                bucketGroup: false
+                keys:
                       expr: _col0
                       type: string
-                outputColumnNames: _col0
-                Group By Operator
-                  aggregations:
-                        expr: count(1)
-                  bucketGroup: false
-                  keys:
-                        expr: _col0
-                        type: string
-                  mode: final
+                mode: final
+                outputColumnNames: _col0, _col1
+                Select Operator
+                  expressions:
+                        expr: UDFToInteger(_col0)
+                        type: int
+                        expr: UDFToInteger(_col1)
+                        type: int
                   outputColumnNames: _col0, _col1
-                  Select Operator
-                    expressions:
-                          expr: _col0
-                          type: string
-                          expr: _col1
-                          type: bigint
-                    outputColumnNames: _col0, _col1
-                    Select Operator
-                      expressions:
-                            expr: UDFToInteger(_col0)
-                            type: int
-                            expr: UDFToInteger(_col1)
-                            type: int
-                      outputColumnNames: _col0, _col1
-                      File Output Operator
-                        compressed: false
-                        GlobalTableId: 1
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 1
 #### A masked pattern was here ####
-                        NumFilesPerFileSink: 1
+                    NumFilesPerFileSink: 1
 #### A masked pattern was here ####
-                        table:
-                            input format: org.apache.hadoop.mapred.TextInputFormat
-                            output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                            properties:
-                              bucket_count -1
-                              columns key,cnt
-                              columns.types int:int
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        properties:
+                          bucket_count -1
+                          columns key,cnt
+                          columns.types int:int
 #### A masked pattern was here ####
-                              name default.outputtbl1
-                              numFiles 1
-                              numPartitions 0
-                              numRows 5
-                              rawDataSize 15
-                              serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
-                              serialization.format 1
-                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                              totalSize 20
+                          name default.outputtbl1
+                          numFiles 1
+                          numPartitions 0
+                          numRows 5
+                          rawDataSize 15
+                          serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
+                          serialization.format 1
+                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                          totalSize 20
 #### A masked pattern was here ####
-                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                            name: default.outputtbl1
-                        TotalFiles: 1
-                        GatherStats: true
-                        MultiFileSpray: false
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: default.outputtbl1
+                    TotalFiles: 1
+                    GatherStats: true
+                    MultiFileSpray: false
       Needs Tagging: false
       Path -> Alias:
 #### A masked pattern was here ####
@@ -1514,44 +1474,35 @@ STAGE PLANS:
                   expressions:
                         expr: _col0
                         type: int
-                        expr: _col1
-                        type: string
-                        expr: _col2
-                        type: bigint
+                        expr: UDFToInteger(_col1)
+                        type: int
+                        expr: UDFToInteger(_col2)
+                        type: int
                   outputColumnNames: _col0, _col1, _col2
-                  Select Operator
-                    expressions:
-                          expr: _col0
-                          type: int
-                          expr: UDFToInteger(_col1)
-                          type: int
-                          expr: UDFToInteger(_col2)
-                          type: int
-                    outputColumnNames: _col0, _col1, _col2
-                    File Output Operator
-                      compressed: false
-                      GlobalTableId: 1
-#### A masked pattern was here ####
-                      NumFilesPerFileSink: 1
-#### A masked pattern was here ####
-                      table:
-                          input format: org.apache.hadoop.mapred.TextInputFormat
-                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                          properties:
-                            bucket_count -1
-                            columns key1,key2,cnt
-                            columns.types int:int:int
-#### A masked pattern was here ####
-                            name default.outputtbl3
-                            serialization.ddl struct outputtbl3 { i32 key1, i32 key2, i32 cnt}
-                            serialization.format 1
-                            serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                          name: default.outputtbl3
-                      TotalFiles: 1
-                      GatherStats: true
-                      MultiFileSpray: false
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 1
+#### A masked pattern was here ####
+                    NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        properties:
+                          bucket_count -1
+                          columns key1,key2,cnt
+                          columns.types int:int:int
+#### A masked pattern was here ####
+                          name default.outputtbl3
+                          serialization.ddl struct outputtbl3 { i32 key1, i32 key2, i32 cnt}
+                          serialization.format 1
+                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+#### A masked pattern was here ####
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: default.outputtbl3
+                    TotalFiles: 1
+                    GatherStats: true
+                    MultiFileSpray: false
       Needs Tagging: false
       Path -> Alias:
 #### A masked pattern was here ####
@@ -1995,50 +1946,39 @@ STAGE PLANS:
           outputColumnNames: _col0, _col1, _col2, _col3
           Select Operator
             expressions:
-                  expr: _col0
-                  type: string
+                  expr: UDFToInteger(_col0)
+                  type: int
                   expr: _col1
                   type: int
                   expr: _col2
                   type: string
-                  expr: _col3
-                  type: bigint
+                  expr: UDFToInteger(_col3)
+                  type: int
             outputColumnNames: _col0, _col1, _col2, _col3
-            Select Operator
-              expressions:
-                    expr: UDFToInteger(_col0)
-                    type: int
-                    expr: _col1
-                    type: int
-                    expr: _col2
-                    type: string
-                    expr: UDFToInteger(_col3)
-                    type: int
-              outputColumnNames: _col0, _col1, _col2, _col3
-              File Output Operator
-                compressed: false
-                GlobalTableId: 1
-#### A masked pattern was here ####
-                NumFilesPerFileSink: 1
-#### A masked pattern was here ####
-                table:
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns key1,key2,key3,cnt
-                      columns.types int:int:string:int
-#### A masked pattern was here ####
-                      name default.outputtbl4
-                      serialization.ddl struct outputtbl4 { i32 key1, i32 key2, string key3, i32 cnt}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.outputtbl4
-                TotalFiles: 1
-                GatherStats: true
-                MultiFileSpray: false
+            File Output Operator
+              compressed: false
+              GlobalTableId: 1
+#### A masked pattern was here ####
+              NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+              table:
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  properties:
+                    bucket_count -1
+                    columns key1,key2,key3,cnt
+                    columns.types int:int:string:int
+#### A masked pattern was here ####
+                    name default.outputtbl4
+                    serialization.ddl struct outputtbl4 { i32 key1, i32 key2, string key3, i32 cnt}
+                    serialization.format 1
+                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+#### A masked pattern was here ####
+                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  name: default.outputtbl4
+              TotalFiles: 1
+              GatherStats: true
+              MultiFileSpray: false
       Truncated Path -> Alias:
         /t1 [t1]
 
@@ -2269,51 +2209,42 @@ STAGE PLANS:
           outputColumnNames: _col0, _col1, _col2
           Select Operator
             expressions:
-                  expr: _col0
-                  type: string
-                  expr: _col1
-                  type: double
-                  expr: _col2
-                  type: bigint
+                  expr: UDFToInteger(_col0)
+                  type: int
+                  expr: UDFToInteger(_col1)
+                  type: int
+                  expr: UDFToInteger(_col2)
+                  type: int
             outputColumnNames: _col0, _col1, _col2
-            Select Operator
-              expressions:
-                    expr: UDFToInteger(_col0)
-                    type: int
-                    expr: UDFToInteger(_col1)
-                    type: int
-                    expr: UDFToInteger(_col2)
-                    type: int
-              outputColumnNames: _col0, _col1, _col2
-              File Output Operator
-                compressed: false
-                GlobalTableId: 1
-#### A masked pattern was here ####
-                NumFilesPerFileSink: 1
-#### A masked pattern was here ####
-                table:
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns key1,key2,cnt
-                      columns.types int:int:int
-#### A masked pattern was here ####
-                      name default.outputtbl3
-                      numFiles 1
-                      numPartitions 0
-                      numRows 5
-                      rawDataSize 25
-                      serialization.ddl struct outputtbl3 { i32 key1, i32 key2, i32 cnt}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      totalSize 30
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.outputtbl3
-                TotalFiles: 1
-                GatherStats: true
-                MultiFileSpray: false
+            File Output Operator
+              compressed: false
+              GlobalTableId: 1
+#### A masked pattern was here ####
+              NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+              table:
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  properties:
+                    bucket_count -1
+                    columns key1,key2,cnt
+                    columns.types int:int:int
+#### A masked pattern was here ####
+                    name default.outputtbl3
+                    numFiles 1
+                    numPartitions 0
+                    numRows 5
+                    rawDataSize 25
+                    serialization.ddl struct outputtbl3 { i32 key1, i32 key2, i32 cnt}
+                    serialization.format 1
+                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                    totalSize 30
+#### A masked pattern was here ####
+                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  name: default.outputtbl3
+              TotalFiles: 1
+              GatherStats: true
+              MultiFileSpray: false
       Truncated Path -> Alias:
         /t1 [t1]
 
@@ -2491,35 +2422,28 @@ STAGE PLANS:
                         type: string
                         expr: _col1
                         type: bigint
-                  outputColumnNames: _col0, _col1
-                  Select Operator
-                    expressions:
-                          expr: _col0
-                          type: string
-                          expr: _col1
-                          type: bigint
+                  outputColumnNames: _col0, _col1
+                  Group By Operator
+                    aggregations:
+                          expr: sum(_col1)
+                    bucketGroup: false
+                    keys:
+                          expr: (_col0 + _col0)
+                          type: double
+                    mode: hash
                     outputColumnNames: _col0, _col1
-                    Group By Operator
-                      aggregations:
-                            expr: sum(_col1)
-                      bucketGroup: false
-                      keys:
-                            expr: (_col0 + _col0)
+                    Reduce Output Operator
+                      key expressions:
+                            expr: _col0
                             type: double
-                      mode: hash
-                      outputColumnNames: _col0, _col1
-                      Reduce Output Operator
-                        key expressions:
-                              expr: _col0
-                              type: double
-                        sort order: +
-                        Map-reduce partition columns:
-                              expr: _col0
-                              type: double
-                        tag: -1
-                        value expressions:
-                              expr: _col1
-                              type: bigint
+                      sort order: +
+                      Map-reduce partition columns:
+                            expr: _col0
+                            type: double
+                      tag: -1
+                      value expressions:
+                            expr: _col1
+                            type: bigint
       Needs Tagging: false
       Path -> Alias:
 #### A masked pattern was here ####
@@ -2582,47 +2506,40 @@ STAGE PLANS:
           outputColumnNames: _col0, _col1
           Select Operator
             expressions:
-                  expr: _col0
-                  type: double
-                  expr: _col1
-                  type: bigint
+                  expr: UDFToInteger(_col0)
+                  type: int
+                  expr: UDFToInteger(_col1)
+                  type: int
             outputColumnNames: _col0, _col1
-            Select Operator
-              expressions:
-                    expr: UDFToInteger(_col0)
-                    type: int
-                    expr: UDFToInteger(_col1)
-                    type: int
-              outputColumnNames: _col0, _col1
-              File Output Operator
-                compressed: false
-                GlobalTableId: 1
-#### A masked pattern was here ####
-                NumFilesPerFileSink: 1
-#### A masked pattern was here ####
-                table:
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns key,cnt
-                      columns.types int:int
-#### A masked pattern was here ####
-                      name default.outputtbl1
-                      numFiles 1
-                      numPartitions 0
-                      numRows 5
-                      rawDataSize 15
-                      serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      totalSize 20
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.outputtbl1
-                TotalFiles: 1
-                GatherStats: true
-                MultiFileSpray: false
+            File Output Operator
+              compressed: false
+              GlobalTableId: 1
+#### A masked pattern was here ####
+              NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+              table:
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  properties:
+                    bucket_count -1
+                    columns key,cnt
+                    columns.types int:int
+#### A masked pattern was here ####
+                    name default.outputtbl1
+                    numFiles 1
+                    numPartitions 0
+                    numRows 5
+                    rawDataSize 15
+                    serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
+                    serialization.format 1
+                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                    totalSize 20
+#### A masked pattern was here ####
+                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  name: default.outputtbl1
+              TotalFiles: 1
+              GatherStats: true
+              MultiFileSpray: false
       Truncated Path -> Alias:
         /t1 [subq1:t1]
 
@@ -2817,47 +2734,40 @@ STAGE PLANS:
                   Union
                     Select Operator
                       expressions:
-                            expr: _col0
-                            type: string
-                            expr: _col1
-                            type: bigint
+                            expr: UDFToInteger(_col0)
+                            type: int
+                            expr: UDFToInteger(_col1)
+                            type: int
                       outputColumnNames: _col0, _col1
-                      Select Operator
-                        expressions:
-                              expr: UDFToInteger(_col0)
-                              type: int
-                              expr: UDFToInteger(_col1)
-                              type: int
-                        outputColumnNames: _col0, _col1
-                        File Output Operator
-                          compressed: false
-                          GlobalTableId: 1
-#### A masked pattern was here ####
-                          NumFilesPerFileSink: 1
-#### A masked pattern was here ####
-                          table:
-                              input format: org.apache.hadoop.mapred.TextInputFormat
-                              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                              properties:
-                                bucket_count -1
-                                columns key,cnt
-                                columns.types int:int
-#### A masked pattern was here ####
-                                name default.outputtbl1
-                                numFiles 1
-                                numPartitions 0
-                                numRows 5
-                                rawDataSize 17
-                                serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
-                                serialization.format 1
-                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                                totalSize 22
-#### A masked pattern was here ####
-                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                              name: default.outputtbl1
-                          TotalFiles: 1
-                          GatherStats: true
-                          MultiFileSpray: false
+                      File Output Operator
+                        compressed: false
+                        GlobalTableId: 1
+#### A masked pattern was here ####
+                        NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+                        table:
+                            input format: org.apache.hadoop.mapred.TextInputFormat
+                            output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                            properties:
+                              bucket_count -1
+                              columns key,cnt
+                              columns.types int:int
+#### A masked pattern was here ####
+                              name default.outputtbl1
+                              numFiles 1
+                              numPartitions 0
+                              numRows 5
+                              rawDataSize 17
+                              serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
+                              serialization.format 1
+                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                              totalSize 22
+#### A masked pattern was here ####
+                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                            name: default.outputtbl1
+                        TotalFiles: 1
+                        GatherStats: true
+                        MultiFileSpray: false
         null-subquery2:subq1-subquery2:t1 
           TableScan
             alias: t1
@@ -2886,47 +2796,40 @@ STAGE PLANS:
                   Union
                     Select Operator
                       expressions:
-                            expr: _col0
-                            type: string
-                            expr: _col1
-                            type: bigint
+                            expr: UDFToInteger(_col0)
+                            type: int
+                            expr: UDFToInteger(_col1)
+                            type: int
                       outputColumnNames: _col0, _col1
-                      Select Operator
-                        expressions:
-                              expr: UDFToInteger(_col0)
-                              type: int
-                              expr: UDFToInteger(_col1)
-                              type: int
-                        outputColumnNames: _col0, _col1
-                        File Output Operator
-                          compressed: false
-                          GlobalTableId: 1
-#### A masked pattern was here ####
-                          NumFilesPerFileSink: 1
-#### A masked pattern was here ####
-                          table:
-                              input format: org.apache.hadoop.mapred.TextInputFormat
-                              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                              properties:
-                                bucket_count -1
-                                columns key,cnt
-                                columns.types int:int
-#### A masked pattern was here ####
-                                name default.outputtbl1
-                                numFiles 1
-                                numPartitions 0
-                                numRows 5
-                                rawDataSize 17
-                                serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
-                                serialization.format 1
-                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                                totalSize 22
-#### A masked pattern was here ####
-                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                              name: default.outputtbl1
-                          TotalFiles: 1
-                          GatherStats: true
-                          MultiFileSpray: false
+                      File Output Operator
+                        compressed: false
+                        GlobalTableId: 1
+#### A masked pattern was here ####
+                        NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+                        table:
+                            input format: org.apache.hadoop.mapred.TextInputFormat
+                            output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                            properties:
+                              bucket_count -1
+                              columns key,cnt
+                              columns.types int:int
+#### A masked pattern was here ####
+                              name default.outputtbl1
+                              numFiles 1
+                              numPartitions 0
+                              numRows 5
+                              rawDataSize 17
+                              serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
+                              serialization.format 1
+                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                              totalSize 22
+#### A masked pattern was here ####
+                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                            name: default.outputtbl1
+                        TotalFiles: 1
+                        GatherStats: true
+                        MultiFileSpray: false
       Needs Tagging: false
       Path -> Alias:
 #### A masked pattern was here ####
@@ -3460,47 +3363,40 @@ STAGE PLANS:
             Union
               Select Operator
                 expressions:
-                      expr: _col0
-                      type: double
-                      expr: _col1
-                      type: bigint
+                      expr: UDFToInteger(_col0)
+                      type: int
+                      expr: UDFToInteger(_col1)
+                      type: int
                 outputColumnNames: _col0, _col1
-                Select Operator
-                  expressions:
-                        expr: UDFToInteger(_col0)
-                        type: int
-                        expr: UDFToInteger(_col1)
-                        type: int
-                  outputColumnNames: _col0, _col1
-                  File Output Operator
-                    compressed: false
-                    GlobalTableId: 1
-#### A masked pattern was here ####
-                    NumFilesPerFileSink: 1
-#### A masked pattern was here ####
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                        properties:
-                          bucket_count -1
-                          columns key,cnt
-                          columns.types int:int
-#### A masked pattern was here ####
-                          name default.outputtbl1
-                          numFiles 1
-                          numPartitions 0
-                          numRows 10
-                          rawDataSize 30
-                          serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
-                          serialization.format 1
-                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                          totalSize 40
-#### A masked pattern was here ####
-                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                        name: default.outputtbl1
-                    TotalFiles: 1
-                    GatherStats: true
-                    MultiFileSpray: false
+                File Output Operator
+                  compressed: false
+                  GlobalTableId: 1
+#### A masked pattern was here ####
+                  NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+                  table:
+                      input format: org.apache.hadoop.mapred.TextInputFormat
+                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                      properties:
+                        bucket_count -1
+                        columns key,cnt
+                        columns.types int:int
+#### A masked pattern was here ####
+                        name default.outputtbl1
+                        numFiles 1
+                        numPartitions 0
+                        numRows 10
+                        rawDataSize 30
+                        serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
+                        serialization.format 1
+                        serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        totalSize 40
+#### A masked pattern was here ####
+                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                      name: default.outputtbl1
+                  TotalFiles: 1
+                  GatherStats: true
+                  MultiFileSpray: false
         null-subquery1:subq1-subquery1:t1 
           TableScan
             alias: t1
@@ -3521,62 +3417,48 @@ STAGE PLANS:
                 outputColumnNames: _col0, _col1
                 Select Operator
                   expressions:
-                        expr: _col0
-                        type: string
+                        expr: UDFToDouble(_col0)
+                        type: double
                         expr: _col1
                         type: bigint
                   outputColumnNames: _col0, _col1
-                  Select Operator
-                    expressions:
-                          expr: UDFToDouble(_col0)
-                          type: double
-                          expr: _col1
-                          type: bigint
-                    outputColumnNames: _col0, _col1
-                    Union
-                      Select Operator
-                        expressions:
-                              expr: _col0
-                              type: double
-                              expr: _col1
-                              type: bigint
-                        outputColumnNames: _col0, _col1
-                        Select Operator
-                          expressions:
-                                expr: UDFToInteger(_col0)
-                                type: int
-                                expr: UDFToInteger(_col1)
-                                type: int
-                          outputColumnNames: _col0, _col1
-                          File Output Operator
-                            compressed: false
-                            GlobalTableId: 1
-#### A masked pattern was here ####
-                            NumFilesPerFileSink: 1
-#### A masked pattern was here ####
-                            table:
-                                input format: org.apache.hadoop.mapred.TextInputFormat
-                                output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                                properties:
-                                  bucket_count -1
-                                  columns key,cnt
-                                  columns.types int:int
-#### A masked pattern was here ####
-                                  name default.outputtbl1
-                                  numFiles 1
-                                  numPartitions 0
-                                  numRows 10
-                                  rawDataSize 30
-                                  serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
-                                  serialization.format 1
-                                  serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                                  totalSize 40
-#### A masked pattern was here ####
-                                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                                name: default.outputtbl1
-                            TotalFiles: 1
-                            GatherStats: true
-                            MultiFileSpray: false
+                  Union
+                    Select Operator
+                      expressions:
+                            expr: UDFToInteger(_col0)
+                            type: int
+                            expr: UDFToInteger(_col1)
+                            type: int
+                      outputColumnNames: _col0, _col1
+                      File Output Operator
+                        compressed: false
+                        GlobalTableId: 1
+#### A masked pattern was here ####
+                        NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+                        table:
+                            input format: org.apache.hadoop.mapred.TextInputFormat
+                            output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                            properties:
+                              bucket_count -1
+                              columns key,cnt
+                              columns.types int:int
+#### A masked pattern was here ####
+                              name default.outputtbl1
+                              numFiles 1
+                              numPartitions 0
+                              numRows 10
+                              rawDataSize 30
+                              serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
+                              serialization.format 1
+                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                              totalSize 40
+#### A masked pattern was here ####
+                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                            name: default.outputtbl1
+                        TotalFiles: 1
+                        GatherStats: true
+                        MultiFileSpray: false
       Needs Tagging: false
       Path -> Alias:
 #### A masked pattern was here ####
@@ -4140,47 +4022,40 @@ STAGE PLANS:
           outputColumnNames: _col0, _col1, _col3
           Select Operator
             expressions:
-                  expr: _col0
-                  type: string
-                  expr: (_col1 + _col3)
-                  type: bigint
+                  expr: UDFToInteger(_col0)
+                  type: int
+                  expr: UDFToInteger((_col1 + _col3))
+                  type: int
             outputColumnNames: _col0, _col1
-            Select Operator
-              expressions:
-                    expr: UDFToInteger(_col0)
-                    type: int
-                    expr: UDFToInteger(_col1)
-                    type: int
-              outputColumnNames: _col0, _col1
-              File Output Operator
-                compressed: false
-                GlobalTableId: 1
-#### A masked pattern was here ####
-                NumFilesPerFileSink: 1
-#### A masked pattern was here ####
-                table:
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns key,cnt
-                      columns.types int:int
-#### A masked pattern was here ####
-                      name default.outputtbl1
-                      numFiles 1
-                      numPartitions 0
-                      numRows 10
-                      rawDataSize 32
-                      serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      totalSize 42
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.outputtbl1
-                TotalFiles: 1
-                GatherStats: true
-                MultiFileSpray: false
+            File Output Operator
+              compressed: false
+              GlobalTableId: 1
+#### A masked pattern was here ####
+              NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+              table:
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  properties:
+                    bucket_count -1
+                    columns key,cnt
+                    columns.types int:int
+#### A masked pattern was here ####
+                    name default.outputtbl1
+                    numFiles 1
+                    numPartitions 0
+                    numRows 10
+                    rawDataSize 32
+                    serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
+                    serialization.format 1
+                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                    totalSize 42
+#### A masked pattern was here ####
+                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  name: default.outputtbl1
+              TotalFiles: 1
+              GatherStats: true
+              MultiFileSpray: false
       Truncated Path -> Alias:
         /t1 [subq1:t1, subq2:t1]
 
@@ -4884,47 +4759,40 @@ STAGE PLANS:
           outputColumnNames: _col0, _col1
           Select Operator
             expressions:
-                  expr: _col0
-                  type: string
-                  expr: _col1
-                  type: bigint
+                  expr: UDFToInteger(_col0)
+                  type: int
+                  expr: UDFToInteger(_col1)
+                  type: int
             outputColumnNames: _col0, _col1
-            Select Operator
-              expressions:
-                    expr: UDFToInteger(_col0)
-                    type: int
-                    expr: UDFToInteger(_col1)
-                    type: int
-              outputColumnNames: _col0, _col1
-              File Output Operator
-                compressed: false
-                GlobalTableId: 1
-#### A masked pattern was here ####
-                NumFilesPerFileSink: 1
-#### A masked pattern was here ####
-                table:
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      bucket_count -1
-                      columns key,cnt
-                      columns.types int:int
-#### A masked pattern was here ####
-                      name default.outputtbl1
-                      numFiles 1
-                      numPartitions 0
-                      numRows 5
-                      rawDataSize 15
-                      serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
-                      serialization.format 1
-                      serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      totalSize 20
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.outputtbl1
-                TotalFiles: 1
-                GatherStats: true
-                MultiFileSpray: false
+            File Output Operator
+              compressed: false
+              GlobalTableId: 1
+#### A masked pattern was here ####
+              NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+              table:
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  properties:
+                    bucket_count -1
+                    columns key,cnt
+                    columns.types int:int
+#### A masked pattern was here ####
+                    name default.outputtbl1
+                    numFiles 1
+                    numPartitions 0
+                    numRows 5
+                    rawDataSize 15
+                    serialization.ddl struct outputtbl1 { i32 key, i32 cnt}
+                    serialization.format 1
+                    serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                    totalSize 20
+#### A masked pattern was here ####
+                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  name: default.outputtbl1
+              TotalFiles: 1
+              GatherStats: true
+              MultiFileSpray: false
       Truncated Path -> Alias:
         /t2 [t2]
 
@@ -5137,55 +5005,44 @@ STAGE PLANS:
                 outputColumnNames: _col0, _col1, _col2, _col3
                 Select Operator
                   expressions:
-                        expr: _col0
-                        type: string
+                        expr: UDFToInteger(_col0)
+                        type: int
                         expr: _col1
                         type: int
                         expr: _col2
                         type: string
-                        expr: _col3
-                        type: bigint
+                        expr: UDFToInteger(_col3)
+                        type: int
                   outputColumnNames: _col0, _col1, _col2, _col3
-                  Select Operator
-                    expressions:
-                          expr: UDFToInteger(_col0)
-                          type: int
-                          expr: _col1
-                          type: int
-                          expr: _col2
-                          type: string
-                          expr: UDFToInteger(_col3)
-                          type: int
-                    outputColumnNames: _col0, _col1, _col2, _col3
-                    File Output Operator
-                      compressed: false
-                      GlobalTableId: 1
-#### A masked pattern was here ####
-                      NumFilesPerFileSink: 1
-#### A masked pattern was here ####
-                      table:
-                          input format: org.apache.hadoop.mapred.TextInputFormat
-                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                          properties:
-                            bucket_count -1
-                            columns key1,key2,key3,cnt
-                            columns.types int:int:string:int
-#### A masked pattern was here ####
-                            name default.outputtbl4
-                            numFiles 1
-                            numPartitions 0
-                            numRows 6
-                            rawDataSize 48
-                            serialization.ddl struct outputtbl4 { i32 key1, i32 key2, string key3, i32 cnt}
-                            serialization.format 1
-                            serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                            totalSize 54
-#### A masked pattern was here ####
-                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                          name: default.outputtbl4
-                      TotalFiles: 1
-                      GatherStats: true
-                      MultiFileSpray: false
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 1
+#### A masked pattern was here ####
+                    NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        properties:
+                          bucket_count -1
+                          columns key1,key2,key3,cnt
+                          columns.types int:int:string:int
+#### A masked pattern was here ####
+                          name default.outputtbl4
+                          numFiles 1
+                          numPartitions 0
+                          numRows 6
+                          rawDataSize 48
+                          serialization.ddl struct outputtbl4 { i32 key1, i32 key2, string key3, i32 cnt}
+                          serialization.format 1
+                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                          totalSize 54
+#### A masked pattern was here ####
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: default.outputtbl4
+                    TotalFiles: 1
+                    GatherStats: true
+                    MultiFileSpray: false
       Needs Tagging: false
       Path -> Alias:
 #### A masked pattern was here ####
@@ -5682,54 +5539,41 @@ STAGE PLANS:
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4
                 Select Operator
                   expressions:
-                        expr: _col0
-                        type: string
+                        expr: UDFToInteger(_col0)
+                        type: int
                         expr: _col1
                         type: int
                         expr: _col2
                         type: string
                         expr: _col3
                         type: int
-                        expr: _col4
-                        type: bigint
+                        expr: UDFToInteger(_col4)
+                        type: int
                   outputColumnNames: _col0, _col1, _col2, _col3, _col4
-                  Select Operator
-                    expressions:
-                          expr: UDFToInteger(_col0)
-                          type: int
-                          expr: _col1
-                          type: int
-                          expr: _col2
-                          type: string
-                          expr: _col3
-                          type: int
-                          expr: UDFToInteger(_col4)
-                          type: int
-                    outputColumnNames: _col0, _col1, _col2, _col3, _col4
-                    File Output Operator
-                      compressed: false
-                      GlobalTableId: 1
-#### A masked pattern was here ####
-                      NumFilesPerFileSink: 1
-#### A masked pattern was here ####
-                      table:
-                          input format: org.apache.hadoop.mapred.TextInputFormat
-                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                          properties:
-                            bucket_count -1
-                            columns key1,key2,key3,key4,cnt
-                            columns.types int:int:string:int:int
-#### A masked pattern was here ####
-                            name default.outputtbl5
-                            serialization.ddl struct outputtbl5 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt}
-                            serialization.format 1
-                            serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-#### A masked pattern was here ####
-                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                          name: default.outputtbl5
-                      TotalFiles: 1
-                      GatherStats: true
-                      MultiFileSpray: false
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 1
+#### A masked pattern was here ####
+                    NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        properties:
+                          bucket_count -1
+                          columns key1,key2,key3,key4,cnt
+                          columns.types int:int:string:int:int
+#### A masked pattern was here ####
+                          name default.outputtbl5
+                          serialization.ddl struct outputtbl5 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt}
+                          serialization.format 1
+                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+#### A masked pattern was here ####
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: default.outputtbl5
+                    TotalFiles: 1
+                    GatherStats: true
+                    MultiFileSpray: false
       Needs Tagging: false
       Path -> Alias:
 #### A masked pattern was here ####
@@ -6153,79 +5997,59 @@ STAGE PLANS:
                     expr: val
                     type: string
               outputColumnNames: _col0, _col1, _col2
-              Select Operator
-                expressions:
+              Group By Operator
+                aggregations:
+                      expr: count(1)
+                bucketGroup: false
+                keys:
                       expr: _col0
                       type: string
                       expr: _col1
                       type: int
                       expr: _col2
                       type: string
-                outputColumnNames: _col0, _col1, _col2
-                Group By Operator
-                  aggregations:
-                        expr: count(1)
-                  bucketGroup: false
-                  keys:
-                        expr: _col0
-                        type: string
+                mode: final
+                outputColumnNames: _col0, _col1, _col2, _col3
+                Select Operator
+                  expressions:
+                        expr: UDFToInteger(_col0)
+                        type: int
                         expr: _col1
                         type: int
                         expr: _col2
                         type: string
-                  mode: final
+                        expr: UDFToInteger(_col3)
+                        type: int
                   outputColumnNames: _col0, _col1, _col2, _col3
-                  Select Operator
-                    expressions:
-                          expr: _col0
-                          type: string
-                          expr: _col1
-                          type: int
-                          expr: _col2
-                          type: string
-                          expr: _col3
-                          type: bigint
-                    outputColumnNames: _col0, _col1, _col2, _col3
-                    Select Operator
-                      expressions:
-                            expr: UDFToInteger(_col0)
-                            type: int
-                            expr: _col1
-                            type: int
-                            expr: _col2
-                            type: string
-                            expr: UDFToInteger(_col3)
-                            type: int
-                      outputColumnNames: _col0, _col1, _col2, _col3
-                      File Output Operator
-                        compressed: false
-                        GlobalTableId: 1
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 1
 #### A masked pattern was here ####
-                        NumFilesPerFileSink: 1
+                    NumFilesPerFileSink: 1
 #### A masked pattern was here ####
-                        table:
-                            input format: org.apache.hadoop.mapred.TextInputFormat
-                            output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                            properties:
-                              bucket_count -1
-                              columns key1,key2,key3,cnt
-                              columns.types int:int:string:int
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        properties:
+                          bucket_count -1
+                          columns key1,key2,key3,cnt
+                          columns.types int:int:string:int
 #### A masked pattern was here ####
-                              name default.outputtbl4
-                              numFiles 1
-                              numPartitions 0
-                              numRows 6
-                              rawDataSize 48
-                              serialization.ddl struct outputtbl4 { i32 key1, i32 key2, string key3, i32 cnt}
-                              serialization.format 1
-                              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                              totalSize 54
+                          name default.outputtbl4
+                          numFiles 1
+                          numPartitions 0
+                          numRows 6
+                          rawDataSize 48
+                          serialization.ddl struct outputtbl4 { i32 key1, i32 key2, string key3, i32 cnt}
+                          serialization.format 1
+                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                          totalSize 54
 #### A masked pattern was here ####
-                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                            name: default.outputtbl4
-                        TotalFiles: 1
-                        GatherStats: true
-                        MultiFileSpray: false
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: default.outputtbl4
+                    TotalFiles: 1
+                    GatherStats: true
+                    MultiFileSpray: false
       Needs Tagging: false
       Path -> Alias:
 #### A masked pattern was here ####
@@ -6699,91 +6523,64 @@ STAGE PLANS:
               expressions:
                     expr: key
                     type: string
+                    expr: 2
+                    type: int
                     expr: val
                     type: string
-              outputColumnNames: _col0, _col2
-              Select Operator
-                expressions:
+              outputColumnNames: _col0, _col3, _col2
+              Group By Operator
+                aggregations:
+                      expr: count(1)
+                bucketGroup: false
+                keys:
                       expr: _col0
                       type: string
+                      expr: _col3
+                      type: int
                       expr: _col2
                       type: string
-                      expr: 2
-                      type: int
-                outputColumnNames: _col0, _col2, _col3
+                mode: final
+                outputColumnNames: _col0, _col1, _col2, _col3
                 Select Operator
                   expressions:
-                        expr: _col0
-                        type: string
-                        expr: _col3
+                        expr: UDFToInteger(_col0)
+                        type: int
+                        expr: _col1
                         type: int
                         expr: _col2
                         type: string
-                  outputColumnNames: _col0, _col3, _col2
-                  Group By Operator
-                    aggregations:
-                          expr: count(1)
-                    bucketGroup: false
-                    keys:
-                          expr: _col0
-                          type: string
-                          expr: _col3
-                          type: int
-                          expr: _col2
-                          type: string
-                    mode: final
-                    outputColumnNames: _col0, _col1, _col2, _col3
-                    Select Operator
-                      expressions:
-                            expr: _col0
-                            type: string
-                            expr: _col1
-                            type: int
-                            expr: _col2
-                            type: string
-                            expr: _col3
-                            type: bigint
-                      outputColumnNames: _col0, _col1, _col2, _col3
-                      Select Operator
-                        expressions:
-                              expr: UDFToInteger(_col0)
-                              type: int
-                              expr: _col1
-                              type: int
-                              expr: _col2
-                              type: string
-                              expr: UDFToInteger(_col3)
-                              type: int
-                        outputColumnNames: _col0, _col1, _col2, _col3
-                        File Output Operator
-                          compressed: false
-                          GlobalTableId: 1
-#### A masked pattern was here ####
-                          NumFilesPerFileSink: 1
-#### A masked pattern was here ####
-                          table:
-                              input format: org.apache.hadoop.mapred.TextInputFormat
-                              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                              properties:
-                                bucket_count -1
-                                columns key1,key2,key3,cnt
-                                columns.types int:int:string:int
-#### A masked pattern was here ####
-                                name default.outputtbl4
-                                numFiles 1
-                                numPartitions 0
-                                numRows 6
-                                rawDataSize 48
-                                serialization.ddl struct outputtbl4 { i32 key1, i32 key2, string key3, i32 cnt}
-                                serialization.format 1
-                                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                                totalSize 54
-#### A masked pattern was here ####
-                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                              name: default.outputtbl4
-                          TotalFiles: 1
-                          GatherStats: true
-                          MultiFileSpray: false
+                        expr: UDFToInteger(_col3)
+                        type: int
+                  outputColumnNames: _col0, _col1, _col2, _col3
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 1
+#### A masked pattern was here ####
+                    NumFilesPerFileSink: 1
+#### A masked pattern was here ####
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        properties:
+                          bucket_count -1
+                          columns key1,key2,key3,cnt
+                          columns.types int:int:string:int
+#### A masked pattern was here ####
+                          name default.outputtbl4
+                          numFiles 1
+                          numPartitions 0
+                          numRows 6
+                          rawDataSize 48
+                          serialization.ddl struct outputtbl4 { i32 key1, i32 key2, string key3, i32 cnt}
+                          serialization.format 1
+                          serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                          totalSize 54
+#### A masked pattern was here ####
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: default.outputtbl4
+                    TotalFiles: 1
+                    GatherStats: true
+                    MultiFileSpray: false
       Needs Tagging: false
       Path -> Alias:
 #### A masked pattern was here ####
@@ -7413,30 +7210,21 @@ STAGE PLANS:
                 outputColumnNames: _col0, _col1, _col2
                 Select Operator
                   expressions:
-                        expr: _col0
-                        type: string
+                        expr: UDFToInteger(_col0)
+                        type: int
                         expr: _col1
                         type: string
-                        expr: _col2
-                        type: bigint
+                        expr: UDFToInteger(_col2)
+                        type: int
                   outputColumnNames: _col0, _col1, _col2
-                  Select Operator
-                    expressions:
-                          expr: UDFToInteger(_col0)
-                          type: int
-                          expr: _col1
-                          type: string
-                          expr: UDFToInteger(_col2)
-                          type: int
-                    outputColumnNames: _col0, _col1, _col2
-                    File Output Operator
-                      compressed: true
-                      GlobalTableId: 2
-                      table:
-                          input format: org.apache.hadoop.mapred.TextInputFormat
-                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                          name: default.dest2
+                  File Output Operator
+                    compressed: true
+                    GlobalTableId: 2
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: default.dest2
       Reduce Operator Tree:
         Group By Operator
           aggregations:
@@ -7449,26 +7237,19 @@ STAGE PLANS:
           outputColumnNames: _col0, _col1
           Select Operator
             expressions:
-                  expr: _col0
-                  type: string
-                  expr: _col1
-                  type: bigint
+                  expr: UDFToInteger(_col0)
+                  type: int
+                  expr: UDFToInteger(_col1)
+                  type: int
             outputColumnNames: _col0, _col1
-            Select Operator
-              expressions:
-                    expr: UDFToInteger(_col0)
-                    type: int
-                    expr: UDFToInteger(_col1)
-                    type: int
-              outputColumnNames: _col0, _col1
-              File Output Operator
-                compressed: true
-                GlobalTableId: 1
-                table:
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.dest1
+            File Output Operator
+              compressed: true
+              GlobalTableId: 1
+              table:
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  name: default.dest1
 
   Stage: Stage-0
     Move Operator
@@ -7844,30 +7625,21 @@ STAGE PLANS:
                     outputColumnNames: _col0, _col1, _col2
                     Select Operator
                       expressions:
-                            expr: _col0
-                            type: string
+                            expr: UDFToInteger(_col0)
+                            type: int
                             expr: _col1
                             type: string
-                            expr: _col2
-                            type: bigint
+                            expr: UDFToInteger(_col2)
+                            type: int
                       outputColumnNames: _col0, _col1, _col2
-                      Select Operator
-                        expressions:
-                              expr: UDFToInteger(_col0)
-                              type: int
-                              expr: _col1
-                              type: string
-                              expr: UDFToInteger(_col2)
-                              type: int
-                        outputColumnNames: _col0, _col1, _col2
-                        File Output Operator
-                          compressed: true
-                          GlobalTableId: 2
-                          table:
-                              input format: org.apache.hadoop.mapred.TextInputFormat
-                              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                              name: default.dest2
+                      File Output Operator
+                        compressed: true
+                        GlobalTableId: 2
+                        table:
+                            input format: org.apache.hadoop.mapred.TextInputFormat
+                            output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                            name: default.dest2
       Reduce Operator Tree:
         Group By Operator
           aggregations:
@@ -7880,26 +7652,19 @@ STAGE PLANS:
           outputColumnNames: _col0, _col1
           Select Operator
             expressions:
-                  expr: _col0
-                  type: string
-                  expr: _col1
-                  type: bigint
+                  expr: UDFToInteger(_col0)
+                  type: int
+                  expr: UDFToInteger(_col1)
+                  type: int
             outputColumnNames: _col0, _col1
-            Select Operator
-              expressions:
-                    expr: UDFToInteger(_col0)
-                    type: int
-                    expr: UDFToInteger(_col1)
-                    type: int
-              outputColumnNames: _col0, _col1
-              File Output Operator
-                compressed: true
-                GlobalTableId: 1
-                table:
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.dest1
+            File Output Operator
+              compressed: true
+              GlobalTableId: 1
+              table:
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  name: default.dest1
 
   Stage: Stage-0
     Move Operator

Modified: hive/trunk/ql/src/test/results/clientpositive/groupby_sort_2.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/groupby_sort_2.q.out?rev=1449981&r1=1449980&r2=1449981&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/groupby_sort_2.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/groupby_sort_2.q.out Tue Feb 26 00:27:32 2013
@@ -99,24 +99,17 @@ STAGE PLANS:
             expressions:
                   expr: _col0
                   type: string
-                  expr: _col1
-                  type: bigint
+                  expr: UDFToInteger(_col1)
+                  type: int
             outputColumnNames: _col0, _col1
-            Select Operator
-              expressions:
-                    expr: _col0
-                    type: string
-                    expr: UDFToInteger(_col1)
-                    type: int
-              outputColumnNames: _col0, _col1
-              File Output Operator
-                compressed: false
-                GlobalTableId: 1
-                table:
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.outputtbl1
+            File Output Operator
+              compressed: false
+              GlobalTableId: 1
+              table:
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  name: default.outputtbl1
 
   Stage: Stage-0
     Move Operator

Modified: hive/trunk/ql/src/test/results/clientpositive/groupby_sort_3.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/groupby_sort_3.q.out?rev=1449981&r1=1449980&r2=1449981&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/groupby_sort_3.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/groupby_sort_3.q.out Tue Feb 26 00:27:32 2013
@@ -86,26 +86,17 @@ STAGE PLANS:
                         type: string
                         expr: _col1
                         type: string
-                        expr: _col2
-                        type: bigint
+                        expr: UDFToInteger(_col2)
+                        type: int
                   outputColumnNames: _col0, _col1, _col2
-                  Select Operator
-                    expressions:
-                          expr: _col0
-                          type: string
-                          expr: _col1
-                          type: string
-                          expr: UDFToInteger(_col2)
-                          type: int
-                    outputColumnNames: _col0, _col1, _col2
-                    File Output Operator
-                      compressed: false
-                      GlobalTableId: 1
-                      table:
-                          input format: org.apache.hadoop.mapred.TextInputFormat
-                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                          name: default.outputtbl1
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 1
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: default.outputtbl1
 
   Stage: Stage-7
     Conditional Operator
@@ -259,24 +250,17 @@ STAGE PLANS:
                   expressions:
                         expr: _col0
                         type: string
-                        expr: _col1
-                        type: bigint
+                        expr: UDFToInteger(_col1)
+                        type: int
                   outputColumnNames: _col0, _col1
-                  Select Operator
-                    expressions:
-                          expr: _col0
-                          type: string
-                          expr: UDFToInteger(_col1)
-                          type: int
-                    outputColumnNames: _col0, _col1
-                    File Output Operator
-                      compressed: false
-                      GlobalTableId: 1
-                      table:
-                          input format: org.apache.hadoop.mapred.TextInputFormat
-                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                          name: default.outputtbl2
+                  File Output Operator
+                    compressed: false
+                    GlobalTableId: 1
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: default.outputtbl2
 
   Stage: Stage-7
     Conditional Operator



Mime
View raw message