hadoop-hive-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From na...@apache.org
Subject svn commit: r887057 [2/6] - in /hadoop/hive/trunk: ./ common/src/java/org/apache/hadoop/hive/conf/ conf/ contrib/src/test/results/clientpositive/ ql/src/java/org/apache/hadoop/hive/ql/exec/ ql/src/java/org/apache/hadoop/hive/ql/metadata/ ql/src/java/or...
Date Fri, 04 Dec 2009 04:39:52 GMT
Added: hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket_groupby.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket_groupby.q.out?rev=887057&view=auto
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket_groupby.q.out (added)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/bucket_groupby.q.out Fri Dec  4 04:39:37 2009
@@ -0,0 +1,1728 @@
+PREHOOK: query: drop table clustergroupby
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table clustergroupby
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: create table clustergroupby(key string, value string) partitioned by(ds string)
+PREHOOK: type: CREATETABLE
+POSTHOOK: query: create table clustergroupby(key string, value string) partitioned by(ds string)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: default@clustergroupby
+PREHOOK: query: describe extended clustergroupby
+PREHOOK: type: DESCTABLE
+POSTHOOK: query: describe extended clustergroupby
+POSTHOOK: type: DESCTABLE
+key	string	
+value	string	
+ds	string	
+	 	 
+Detailed Table Information	Table(tableName:clustergroupby, dbName:default, owner:heyongqiang, createTime:1259868796, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/test/data/warehouse/clustergroupby, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null)], parameters:{transient_lastDdlTime=1259868796})	
+PREHOOK: query: insert overwrite table clustergroupby partition (ds='100') select key, value from src sort by key
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@clustergroupby@ds=100
+POSTHOOK: query: insert overwrite table clustergroupby partition (ds='100') select key, value from src sort by key
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@clustergroupby@ds=100
+PREHOOK: query: explain
+select key, count(1) from clustergroupby where ds='100' group by key limit 10
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+select key, count(1) from clustergroupby where ds='100' group by key limit 10
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+  (TOK_QUERY (TOK_FROM (TOK_TABREF clustergroupby)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_FUNCTION count 1))) (TOK_WHERE (= (TOK_TABLE_OR_COL ds) '100')) (TOK_GROUPBY (TOK_TABLE_OR_COL key)) (TOK_LIMIT 10)))
+
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Alias -> Map Operator Tree:
+        clustergroupby 
+          TableScan
+            alias: clustergroupby
+            Filter Operator
+              predicate:
+                  expr: (ds = '100')
+                  type: boolean
+              Filter Operator
+                predicate:
+                    expr: (ds = '100')
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: key
+                        type: string
+                  outputColumnNames: key
+                  Group By Operator
+                    aggregations:
+                          expr: count(1)
+                    bucketGroup: false
+                    keys:
+                          expr: key
+                          type: string
+                    mode: hash
+                    outputColumnNames: _col0, _col1
+                    Reduce Output Operator
+                      key expressions:
+                            expr: _col0
+                            type: string
+                      sort order: +
+                      Map-reduce partition columns:
+                            expr: _col0
+                            type: string
+                      tag: -1
+                      value expressions:
+                            expr: _col1
+                            type: bigint
+      Reduce Operator Tree:
+        Group By Operator
+          aggregations:
+                expr: count(VALUE._col0)
+          bucketGroup: false
+          keys:
+                expr: KEY._col0
+                type: string
+          mode: mergepartial
+          outputColumnNames: _col0, _col1
+          Select Operator
+            expressions:
+                  expr: _col0
+                  type: string
+                  expr: _col1
+                  type: bigint
+            outputColumnNames: _col0, _col1
+            Limit
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: 10
+
+
+PREHOOK: query: select key, count(1) from clustergroupby where ds='100' group by key limit 10
+PREHOOK: type: QUERY
+PREHOOK: Input: default@clustergroupby@ds=100
+PREHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/1923424614/10000
+POSTHOOK: query: select key, count(1) from clustergroupby where ds='100' group by key limit 10
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@clustergroupby@ds=100
+POSTHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/1923424614/10000
+0	3
+10	1
+100	2
+103	2
+104	2
+105	1
+11	1
+111	1
+113	2
+114	1
+PREHOOK: query: alter table clustergroupby clustered by (key) into 1 buckets
+PREHOOK: type: null
+POSTHOOK: query: alter table clustergroupby clustered by (key) into 1 buckets
+POSTHOOK: type: null
+POSTHOOK: Input: default@clustergroupby
+POSTHOOK: Output: default@clustergroupby
+PREHOOK: query: describe extended clustergroupby
+PREHOOK: type: DESCTABLE
+POSTHOOK: query: describe extended clustergroupby
+POSTHOOK: type: DESCTABLE
+key	string	
+value	string	
+ds	string	
+	 	 
+Detailed Table Information	Table(tableName:clustergroupby, dbName:default, owner:heyongqiang, createTime:1259868796, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/test/data/warehouse/clustergroupby, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[key], sortCols:[], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null)], parameters:{last_modified_by=heyongqiang,last_modified_time=1259868817,transient_lastDdlTime=1259868817})	
+PREHOOK: query: insert overwrite table clustergroupby partition (ds='101') select key, value from src distribute by key
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@clustergroupby@ds=101
+POSTHOOK: query: insert overwrite table clustergroupby partition (ds='101') select key, value from src distribute by key
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@clustergroupby@ds=101
+PREHOOK: query: --normal--
+explain
+select key, count(1) from clustergroupby  where ds='101'  group by key limit 10
+PREHOOK: type: QUERY
+POSTHOOK: query: --normal--
+explain
+select key, count(1) from clustergroupby  where ds='101'  group by key limit 10
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+  (TOK_QUERY (TOK_FROM (TOK_TABREF clustergroupby)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_FUNCTION count 1))) (TOK_WHERE (= (TOK_TABLE_OR_COL ds) '101')) (TOK_GROUPBY (TOK_TABLE_OR_COL key)) (TOK_LIMIT 10)))
+
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Alias -> Map Operator Tree:
+        clustergroupby 
+          TableScan
+            alias: clustergroupby
+            Filter Operator
+              predicate:
+                  expr: (ds = '101')
+                  type: boolean
+              Filter Operator
+                predicate:
+                    expr: (ds = '101')
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: key
+                        type: string
+                  outputColumnNames: key
+                  Group By Operator
+                    aggregations:
+                          expr: count(1)
+                    bucketGroup: true
+                    keys:
+                          expr: key
+                          type: string
+                    mode: hash
+                    outputColumnNames: _col0, _col1
+                    Reduce Output Operator
+                      key expressions:
+                            expr: _col0
+                            type: string
+                      sort order: +
+                      Map-reduce partition columns:
+                            expr: _col0
+                            type: string
+                      tag: -1
+                      value expressions:
+                            expr: _col1
+                            type: bigint
+      Reduce Operator Tree:
+        Group By Operator
+          aggregations:
+                expr: count(VALUE._col0)
+          bucketGroup: false
+          keys:
+                expr: KEY._col0
+                type: string
+          mode: mergepartial
+          outputColumnNames: _col0, _col1
+          Select Operator
+            expressions:
+                  expr: _col0
+                  type: string
+                  expr: _col1
+                  type: bigint
+            outputColumnNames: _col0, _col1
+            Limit
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: 10
+
+
+PREHOOK: query: select key, count(1) from clustergroupby  where ds='101' group by key limit 10
+PREHOOK: type: QUERY
+PREHOOK: Input: default@clustergroupby@ds=101
+PREHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/1926882001/10000
+POSTHOOK: query: select key, count(1) from clustergroupby  where ds='101' group by key limit 10
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@clustergroupby@ds=101
+POSTHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/1926882001/10000
+0	3
+10	1
+100	2
+103	2
+104	2
+105	1
+11	1
+111	1
+113	2
+114	1
+PREHOOK: query: --function--
+explain
+select length(key), count(1) from clustergroupby  where ds='101'  group by length(key) limit 10
+PREHOOK: type: QUERY
+POSTHOOK: query: --function--
+explain
+select length(key), count(1) from clustergroupby  where ds='101'  group by length(key) limit 10
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+  (TOK_QUERY (TOK_FROM (TOK_TABREF clustergroupby)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION length (TOK_TABLE_OR_COL key))) (TOK_SELEXPR (TOK_FUNCTION count 1))) (TOK_WHERE (= (TOK_TABLE_OR_COL ds) '101')) (TOK_GROUPBY (TOK_FUNCTION length (TOK_TABLE_OR_COL key))) (TOK_LIMIT 10)))
+
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Alias -> Map Operator Tree:
+        clustergroupby 
+          TableScan
+            alias: clustergroupby
+            Filter Operator
+              predicate:
+                  expr: (ds = '101')
+                  type: boolean
+              Filter Operator
+                predicate:
+                    expr: (ds = '101')
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: key
+                        type: string
+                  outputColumnNames: key
+                  Group By Operator
+                    aggregations:
+                          expr: count(1)
+                    bucketGroup: true
+                    keys:
+                          expr: length(key)
+                          type: int
+                    mode: hash
+                    outputColumnNames: _col0, _col1
+                    Reduce Output Operator
+                      key expressions:
+                            expr: _col0
+                            type: int
+                      sort order: +
+                      Map-reduce partition columns:
+                            expr: _col0
+                            type: int
+                      tag: -1
+                      value expressions:
+                            expr: _col1
+                            type: bigint
+      Reduce Operator Tree:
+        Group By Operator
+          aggregations:
+                expr: count(VALUE._col0)
+          bucketGroup: false
+          keys:
+                expr: KEY._col0
+                type: int
+          mode: mergepartial
+          outputColumnNames: _col0, _col1
+          Select Operator
+            expressions:
+                  expr: _col0
+                  type: int
+                  expr: _col1
+                  type: bigint
+            outputColumnNames: _col0, _col1
+            Limit
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: 10
+
+
+PREHOOK: query: select length(key), count(1) from clustergroupby  where ds='101' group by length(key) limit 10
+PREHOOK: type: QUERY
+PREHOOK: Input: default@clustergroupby@ds=101
+PREHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/695955681/10000
+POSTHOOK: query: select length(key), count(1) from clustergroupby  where ds='101' group by length(key) limit 10
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@clustergroupby@ds=101
+POSTHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/695955681/10000
+1	10
+2	74
+3	416
+PREHOOK: query: explain
+select abs(length(key)), count(1) from clustergroupby  where ds='101'  group by abs(length(key)) limit 10
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+select abs(length(key)), count(1) from clustergroupby  where ds='101'  group by abs(length(key)) limit 10
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+  (TOK_QUERY (TOK_FROM (TOK_TABREF clustergroupby)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION abs (TOK_FUNCTION length (TOK_TABLE_OR_COL key)))) (TOK_SELEXPR (TOK_FUNCTION count 1))) (TOK_WHERE (= (TOK_TABLE_OR_COL ds) '101')) (TOK_GROUPBY (TOK_FUNCTION abs (TOK_FUNCTION length (TOK_TABLE_OR_COL key)))) (TOK_LIMIT 10)))
+
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Alias -> Map Operator Tree:
+        clustergroupby 
+          TableScan
+            alias: clustergroupby
+            Filter Operator
+              predicate:
+                  expr: (ds = '101')
+                  type: boolean
+              Filter Operator
+                predicate:
+                    expr: (ds = '101')
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: key
+                        type: string
+                  outputColumnNames: key
+                  Group By Operator
+                    aggregations:
+                          expr: count(1)
+                    bucketGroup: true
+                    keys:
+                          expr: abs(length(key))
+                          type: int
+                    mode: hash
+                    outputColumnNames: _col0, _col1
+                    Reduce Output Operator
+                      key expressions:
+                            expr: _col0
+                            type: int
+                      sort order: +
+                      Map-reduce partition columns:
+                            expr: _col0
+                            type: int
+                      tag: -1
+                      value expressions:
+                            expr: _col1
+                            type: bigint
+      Reduce Operator Tree:
+        Group By Operator
+          aggregations:
+                expr: count(VALUE._col0)
+          bucketGroup: false
+          keys:
+                expr: KEY._col0
+                type: int
+          mode: mergepartial
+          outputColumnNames: _col0, _col1
+          Select Operator
+            expressions:
+                  expr: _col0
+                  type: int
+                  expr: _col1
+                  type: bigint
+            outputColumnNames: _col0, _col1
+            Limit
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: 10
+
+
+PREHOOK: query: select abs(length(key)), count(1) from clustergroupby  where ds='101' group by abs(length(key)) limit 10
+PREHOOK: type: QUERY
+PREHOOK: Input: default@clustergroupby@ds=101
+PREHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/1888922671/10000
+POSTHOOK: query: select abs(length(key)), count(1) from clustergroupby  where ds='101' group by abs(length(key)) limit 10
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@clustergroupby@ds=101
+POSTHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/1888922671/10000
+1	10
+2	74
+3	416
+PREHOOK: query: --constant--
+explain
+select key, count(1) from clustergroupby  where ds='101'  group by key,3 limit 10
+PREHOOK: type: QUERY
+POSTHOOK: query: --constant--
+explain
+select key, count(1) from clustergroupby  where ds='101'  group by key,3 limit 10
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+  (TOK_QUERY (TOK_FROM (TOK_TABREF clustergroupby)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_FUNCTION count 1))) (TOK_WHERE (= (TOK_TABLE_OR_COL ds) '101')) (TOK_GROUPBY (TOK_TABLE_OR_COL key) 3) (TOK_LIMIT 10)))
+
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Alias -> Map Operator Tree:
+        clustergroupby 
+          TableScan
+            alias: clustergroupby
+            Filter Operator
+              predicate:
+                  expr: (ds = '101')
+                  type: boolean
+              Filter Operator
+                predicate:
+                    expr: (ds = '101')
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: key
+                        type: string
+                  outputColumnNames: key
+                  Group By Operator
+                    aggregations:
+                          expr: count(1)
+                    bucketGroup: true
+                    keys:
+                          expr: key
+                          type: string
+                          expr: 3
+                          type: int
+                    mode: hash
+                    outputColumnNames: _col0, _col1, _col2
+                    Reduce Output Operator
+                      key expressions:
+                            expr: _col0
+                            type: string
+                            expr: _col1
+                            type: int
+                      sort order: ++
+                      Map-reduce partition columns:
+                            expr: _col0
+                            type: string
+                            expr: _col1
+                            type: int
+                      tag: -1
+                      value expressions:
+                            expr: _col2
+                            type: bigint
+      Reduce Operator Tree:
+        Group By Operator
+          aggregations:
+                expr: count(VALUE._col0)
+          bucketGroup: false
+          keys:
+                expr: KEY._col0
+                type: string
+                expr: KEY._col1
+                type: int
+          mode: mergepartial
+          outputColumnNames: _col0, _col1, _col2
+          Select Operator
+            expressions:
+                  expr: _col0
+                  type: string
+                  expr: _col2
+                  type: bigint
+            outputColumnNames: _col0, _col1
+            Limit
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: 10
+
+
+PREHOOK: query: select key, count(1) from clustergroupby  where ds='101' group by key,3 limit 10
+PREHOOK: type: QUERY
+PREHOOK: Input: default@clustergroupby@ds=101
+PREHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/297590608/10000
+POSTHOOK: query: select key, count(1) from clustergroupby  where ds='101' group by key,3 limit 10
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@clustergroupby@ds=101
+POSTHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/297590608/10000
+0	3
+10	1
+100	2
+103	2
+104	2
+105	1
+11	1
+111	1
+113	2
+114	1
+PREHOOK: query: --subquery--
+explain
+select key, count(1) from (select value as key, key as value from clustergroupby where ds='101')subq group by key limit 10
+PREHOOK: type: QUERY
+POSTHOOK: query: --subquery--
+explain
+select key, count(1) from (select value as key, key as value from clustergroupby where ds='101')subq group by key limit 10
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+  (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF clustergroupby)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL value) key) (TOK_SELEXPR (TOK_TABLE_OR_COL key) value)) (TOK_WHERE (= (TOK_TABLE_OR_COL ds) '101')))) subq)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_FUNCTION count 1))) (TOK_GROUPBY (TOK_TABLE_OR_COL key)) (TOK_LIMIT 10)))
+
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Alias -> Map Operator Tree:
+        subq:clustergroupby 
+          TableScan
+            alias: clustergroupby
+            Filter Operator
+              predicate:
+                  expr: (ds = '101')
+                  type: boolean
+              Filter Operator
+                predicate:
+                    expr: (ds = '101')
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: value
+                        type: string
+                  outputColumnNames: _col0
+                  Select Operator
+                    expressions:
+                          expr: _col0
+                          type: string
+                    outputColumnNames: _col0
+                    Group By Operator
+                      aggregations:
+                            expr: count(1)
+                      bucketGroup: false
+                      keys:
+                            expr: _col0
+                            type: string
+                      mode: hash
+                      outputColumnNames: _col0, _col1
+                      Reduce Output Operator
+                        key expressions:
+                              expr: _col0
+                              type: string
+                        sort order: +
+                        Map-reduce partition columns:
+                              expr: _col0
+                              type: string
+                        tag: -1
+                        value expressions:
+                              expr: _col1
+                              type: bigint
+      Reduce Operator Tree:
+        Group By Operator
+          aggregations:
+                expr: count(VALUE._col0)
+          bucketGroup: false
+          keys:
+                expr: KEY._col0
+                type: string
+          mode: mergepartial
+          outputColumnNames: _col0, _col1
+          Select Operator
+            expressions:
+                  expr: _col0
+                  type: string
+                  expr: _col1
+                  type: bigint
+            outputColumnNames: _col0, _col1
+            Limit
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: 10
+
+
+PREHOOK: query: select key, count(1) from (select value as key, key as value from clustergroupby where ds='101')subq group by key limit 10
+PREHOOK: type: QUERY
+PREHOOK: Input: default@clustergroupby@ds=101
+PREHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/1523288366/10000
+POSTHOOK: query: select key, count(1) from (select value as key, key as value from clustergroupby where ds='101')subq group by key limit 10
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@clustergroupby@ds=101
+POSTHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/1523288366/10000
+val_0	3
+val_10	1
+val_100	2
+val_103	2
+val_104	2
+val_105	1
+val_11	1
+val_111	1
+val_113	2
+val_114	1
+PREHOOK: query: explain
+select key, count(1) from clustergroupby  group by key
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+select key, count(1) from clustergroupby  group by key
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+  (TOK_QUERY (TOK_FROM (TOK_TABREF clustergroupby)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_FUNCTION count 1))) (TOK_GROUPBY (TOK_TABLE_OR_COL key))))
+
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Alias -> Map Operator Tree:
+        clustergroupby 
+          TableScan
+            alias: clustergroupby
+            Select Operator
+              expressions:
+                    expr: key
+                    type: string
+              outputColumnNames: key
+              Group By Operator
+                aggregations:
+                      expr: count(1)
+                bucketGroup: false
+                keys:
+                      expr: key
+                      type: string
+                mode: hash
+                outputColumnNames: _col0, _col1
+                Reduce Output Operator
+                  key expressions:
+                        expr: _col0
+                        type: string
+                  sort order: +
+                  Map-reduce partition columns:
+                        expr: _col0
+                        type: string
+                  tag: -1
+                  value expressions:
+                        expr: _col1
+                        type: bigint
+      Reduce Operator Tree:
+        Group By Operator
+          aggregations:
+                expr: count(VALUE._col0)
+          bucketGroup: false
+          keys:
+                expr: KEY._col0
+                type: string
+          mode: mergepartial
+          outputColumnNames: _col0, _col1
+          Select Operator
+            expressions:
+                  expr: _col0
+                  type: string
+                  expr: _col1
+                  type: bigint
+            outputColumnNames: _col0, _col1
+            File Output Operator
+              compressed: false
+              GlobalTableId: 0
+              table:
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+
+
+PREHOOK: query: select key, count(1) from clustergroupby  group by key
+PREHOOK: type: QUERY
+PREHOOK: Input: default@clustergroupby@ds=100
+PREHOOK: Input: default@clustergroupby@ds=101
+PREHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/1435163727/10000
+POSTHOOK: query: select key, count(1) from clustergroupby  group by key
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@clustergroupby@ds=100
+POSTHOOK: Input: default@clustergroupby@ds=101
+POSTHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/1435163727/10000
+0	6
+10	2
+100	4
+103	4
+104	4
+105	2
+11	2
+111	2
+113	4
+114	2
+116	2
+118	4
+119	6
+12	4
+120	4
+125	4
+126	2
+128	6
+129	4
+131	2
+133	2
+134	4
+136	2
+137	4
+138	8
+143	2
+145	2
+146	4
+149	4
+15	4
+150	2
+152	4
+153	2
+155	2
+156	2
+157	2
+158	2
+160	2
+162	2
+163	2
+164	4
+165	4
+166	2
+167	6
+168	2
+169	8
+17	2
+170	2
+172	4
+174	4
+175	4
+176	4
+177	2
+178	2
+179	4
+18	4
+180	2
+181	2
+183	2
+186	2
+187	6
+189	2
+19	2
+190	2
+191	4
+192	2
+193	6
+194	2
+195	4
+196	2
+197	4
+199	6
+2	2
+20	2
+200	4
+201	2
+202	2
+203	4
+205	4
+207	4
+208	6
+209	4
+213	4
+214	2
+216	4
+217	4
+218	2
+219	4
+221	4
+222	2
+223	4
+224	4
+226	2
+228	2
+229	4
+230	10
+233	4
+235	2
+237	4
+238	4
+239	4
+24	4
+241	2
+242	4
+244	2
+247	2
+248	2
+249	2
+252	2
+255	4
+256	4
+257	2
+258	2
+26	4
+260	2
+262	2
+263	2
+265	4
+266	2
+27	2
+272	4
+273	6
+274	2
+275	2
+277	8
+278	4
+28	2
+280	4
+281	4
+282	4
+283	2
+284	2
+285	2
+286	2
+287	2
+288	4
+289	2
+291	2
+292	2
+296	2
+298	6
+30	2
+302	2
+305	2
+306	2
+307	4
+308	2
+309	4
+310	2
+311	6
+315	2
+316	6
+317	4
+318	6
+321	4
+322	4
+323	2
+325	4
+327	6
+33	2
+331	4
+332	2
+333	4
+335	2
+336	2
+338	2
+339	2
+34	2
+341	2
+342	4
+344	4
+345	2
+348	10
+35	6
+351	2
+353	4
+356	2
+360	2
+362	2
+364	2
+365	2
+366	2
+367	4
+368	2
+369	6
+37	4
+373	2
+374	2
+375	2
+377	2
+378	2
+379	2
+382	4
+384	6
+386	2
+389	2
+392	2
+393	2
+394	2
+395	4
+396	6
+397	4
+399	4
+4	2
+400	2
+401	10
+402	2
+403	6
+404	4
+406	8
+407	2
+409	6
+41	2
+411	2
+413	4
+414	4
+417	6
+418	2
+419	2
+42	4
+421	2
+424	4
+427	2
+429	4
+43	2
+430	6
+431	6
+432	2
+435	2
+436	2
+437	2
+438	6
+439	4
+44	2
+443	2
+444	2
+446	2
+448	2
+449	2
+452	2
+453	2
+454	6
+455	2
+457	2
+458	4
+459	4
+460	2
+462	4
+463	4
+466	6
+467	2
+468	8
+469	10
+47	2
+470	2
+472	2
+475	2
+477	2
+478	4
+479	2
+480	6
+481	2
+482	2
+483	2
+484	2
+485	2
+487	2
+489	8
+490	2
+491	2
+492	4
+493	2
+494	2
+495	2
+496	2
+497	2
+498	6
+5	6
+51	4
+53	2
+54	2
+57	2
+58	4
+64	2
+65	2
+66	2
+67	4
+69	2
+70	6
+72	4
+74	2
+76	4
+77	2
+78	2
+8	2
+80	2
+82	2
+83	4
+84	4
+85	2
+86	2
+87	2
+9	2
+90	6
+92	2
+95	4
+96	2
+97	4
+98	4
+PREHOOK: query: explain
+select key, count(1) from clustergroupby  group by key, 3
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+select key, count(1) from clustergroupby  group by key, 3
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+  (TOK_QUERY (TOK_FROM (TOK_TABREF clustergroupby)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_FUNCTION count 1))) (TOK_GROUPBY (TOK_TABLE_OR_COL key) 3)))
+
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Alias -> Map Operator Tree:
+        clustergroupby 
+          TableScan
+            alias: clustergroupby
+            Select Operator
+              expressions:
+                    expr: key
+                    type: string
+              outputColumnNames: key
+              Group By Operator
+                aggregations:
+                      expr: count(1)
+                bucketGroup: false
+                keys:
+                      expr: key
+                      type: string
+                      expr: 3
+                      type: int
+                mode: hash
+                outputColumnNames: _col0, _col1, _col2
+                Reduce Output Operator
+                  key expressions:
+                        expr: _col0
+                        type: string
+                        expr: _col1
+                        type: int
+                  sort order: ++
+                  Map-reduce partition columns:
+                        expr: _col0
+                        type: string
+                        expr: _col1
+                        type: int
+                  tag: -1
+                  value expressions:
+                        expr: _col2
+                        type: bigint
+      Reduce Operator Tree:
+        Group By Operator
+          aggregations:
+                expr: count(VALUE._col0)
+          bucketGroup: false
+          keys:
+                expr: KEY._col0
+                type: string
+                expr: KEY._col1
+                type: int
+          mode: mergepartial
+          outputColumnNames: _col0, _col1, _col2
+          Select Operator
+            expressions:
+                  expr: _col0
+                  type: string
+                  expr: _col2
+                  type: bigint
+            outputColumnNames: _col0, _col1
+            File Output Operator
+              compressed: false
+              GlobalTableId: 0
+              table:
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+
+
+PREHOOK: query: --sort columns--
+alter table clustergroupby clustered by (value) sorted by (key, value) into 1 buckets
+PREHOOK: type: null
+POSTHOOK: query: --sort columns--
+alter table clustergroupby clustered by (value) sorted by (key, value) into 1 buckets
+POSTHOOK: type: null
+POSTHOOK: Input: default@clustergroupby
+POSTHOOK: Output: default@clustergroupby
+PREHOOK: query: describe extended clustergroupby
+PREHOOK: type: DESCTABLE
+POSTHOOK: query: describe extended clustergroupby
+POSTHOOK: type: DESCTABLE
+key	string	
+value	string	
+ds	string	
+	 	 
+Detailed Table Information	Table(tableName:clustergroupby, dbName:default, owner:heyongqiang, createTime:1259868796, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/test/data/warehouse/clustergroupby, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[value], sortCols:[Order(col:key, order:1), Order(col:value, order:1)], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null)], parameters:{last_modified_by=heyongqiang,last_modified_time=1259868863,transient_lastDdlTime=1259868863})	
+PREHOOK: query: insert overwrite table clustergroupby partition (ds='102') select key, value from src distribute by value sort by key, value
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@clustergroupby@ds=102
+POSTHOOK: query: insert overwrite table clustergroupby partition (ds='102') select key, value from src distribute by value sort by key, value
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@clustergroupby@ds=102
+PREHOOK: query: explain
+select key, count(1) from clustergroupby  where ds='102'  group by key limit 10
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+select key, count(1) from clustergroupby  where ds='102'  group by key limit 10
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+  (TOK_QUERY (TOK_FROM (TOK_TABREF clustergroupby)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_FUNCTION count 1))) (TOK_WHERE (= (TOK_TABLE_OR_COL ds) '102')) (TOK_GROUPBY (TOK_TABLE_OR_COL key)) (TOK_LIMIT 10)))
+
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Alias -> Map Operator Tree:
+        clustergroupby 
+          TableScan
+            alias: clustergroupby
+            Filter Operator
+              predicate:
+                  expr: (ds = '102')
+                  type: boolean
+              Filter Operator
+                predicate:
+                    expr: (ds = '102')
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: key
+                        type: string
+                  outputColumnNames: key
+                  Group By Operator
+                    aggregations:
+                          expr: count(1)
+                    bucketGroup: true
+                    keys:
+                          expr: key
+                          type: string
+                    mode: hash
+                    outputColumnNames: _col0, _col1
+                    Reduce Output Operator
+                      key expressions:
+                            expr: _col0
+                            type: string
+                      sort order: +
+                      Map-reduce partition columns:
+                            expr: _col0
+                            type: string
+                      tag: -1
+                      value expressions:
+                            expr: _col1
+                            type: bigint
+      Reduce Operator Tree:
+        Group By Operator
+          aggregations:
+                expr: count(VALUE._col0)
+          bucketGroup: false
+          keys:
+                expr: KEY._col0
+                type: string
+          mode: mergepartial
+          outputColumnNames: _col0, _col1
+          Select Operator
+            expressions:
+                  expr: _col0
+                  type: string
+                  expr: _col1
+                  type: bigint
+            outputColumnNames: _col0, _col1
+            Limit
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: 10
+
+
+PREHOOK: query: select key, count(1) from clustergroupby  where ds='102' group by key limit 10
+PREHOOK: type: QUERY
+PREHOOK: Input: default@clustergroupby@ds=102
+PREHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/1143276036/10000
+POSTHOOK: query: select key, count(1) from clustergroupby  where ds='102' group by key limit 10
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@clustergroupby@ds=102
+POSTHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/1143276036/10000
+0	3
+10	1
+100	2
+103	2
+104	2
+105	1
+11	1
+111	1
+113	2
+114	1
+PREHOOK: query: explain
+select value, count(1) from clustergroupby  where ds='102'  group by value limit 10
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+select value, count(1) from clustergroupby  where ds='102'  group by value limit 10
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+  (TOK_QUERY (TOK_FROM (TOK_TABREF clustergroupby)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL value)) (TOK_SELEXPR (TOK_FUNCTION count 1))) (TOK_WHERE (= (TOK_TABLE_OR_COL ds) '102')) (TOK_GROUPBY (TOK_TABLE_OR_COL value)) (TOK_LIMIT 10)))
+
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Alias -> Map Operator Tree:
+        clustergroupby 
+          TableScan
+            alias: clustergroupby
+            Filter Operator
+              predicate:
+                  expr: (ds = '102')
+                  type: boolean
+              Filter Operator
+                predicate:
+                    expr: (ds = '102')
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: value
+                        type: string
+                  outputColumnNames: value
+                  Group By Operator
+                    aggregations:
+                          expr: count(1)
+                    bucketGroup: false
+                    keys:
+                          expr: value
+                          type: string
+                    mode: hash
+                    outputColumnNames: _col0, _col1
+                    Reduce Output Operator
+                      key expressions:
+                            expr: _col0
+                            type: string
+                      sort order: +
+                      Map-reduce partition columns:
+                            expr: _col0
+                            type: string
+                      tag: -1
+                      value expressions:
+                            expr: _col1
+                            type: bigint
+      Reduce Operator Tree:
+        Group By Operator
+          aggregations:
+                expr: count(VALUE._col0)
+          bucketGroup: false
+          keys:
+                expr: KEY._col0
+                type: string
+          mode: mergepartial
+          outputColumnNames: _col0, _col1
+          Select Operator
+            expressions:
+                  expr: _col0
+                  type: string
+                  expr: _col1
+                  type: bigint
+            outputColumnNames: _col0, _col1
+            Limit
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: 10
+
+
+PREHOOK: query: select value, count(1) from clustergroupby  where ds='102'  group by value limit 10
+PREHOOK: type: QUERY
+PREHOOK: Input: default@clustergroupby@ds=102
+PREHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/177035338/10000
+POSTHOOK: query: select value, count(1) from clustergroupby  where ds='102'  group by value limit 10
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@clustergroupby@ds=102
+POSTHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/177035338/10000
+val_0	3
+val_10	1
+val_100	2
+val_103	2
+val_104	2
+val_105	1
+val_11	1
+val_111	1
+val_113	2
+val_114	1
+PREHOOK: query: explain
+select key, count(1) from clustergroupby  where ds='102'  group by key, value limit 10
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+select key, count(1) from clustergroupby  where ds='102'  group by key, value limit 10
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+  (TOK_QUERY (TOK_FROM (TOK_TABREF clustergroupby)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_FUNCTION count 1))) (TOK_WHERE (= (TOK_TABLE_OR_COL ds) '102')) (TOK_GROUPBY (TOK_TABLE_OR_COL key) (TOK_TABLE_OR_COL value)) (TOK_LIMIT 10)))
+
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Alias -> Map Operator Tree:
+        clustergroupby 
+          TableScan
+            alias: clustergroupby
+            Filter Operator
+              predicate:
+                  expr: (ds = '102')
+                  type: boolean
+              Filter Operator
+                predicate:
+                    expr: (ds = '102')
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: key
+                        type: string
+                        expr: value
+                        type: string
+                  outputColumnNames: key, value
+                  Group By Operator
+                    aggregations:
+                          expr: count(1)
+                    bucketGroup: true
+                    keys:
+                          expr: key
+                          type: string
+                          expr: value
+                          type: string
+                    mode: hash
+                    outputColumnNames: _col0, _col1, _col2
+                    Reduce Output Operator
+                      key expressions:
+                            expr: _col0
+                            type: string
+                            expr: _col1
+                            type: string
+                      sort order: ++
+                      Map-reduce partition columns:
+                            expr: _col0
+                            type: string
+                            expr: _col1
+                            type: string
+                      tag: -1
+                      value expressions:
+                            expr: _col2
+                            type: bigint
+      Reduce Operator Tree:
+        Group By Operator
+          aggregations:
+                expr: count(VALUE._col0)
+          bucketGroup: false
+          keys:
+                expr: KEY._col0
+                type: string
+                expr: KEY._col1
+                type: string
+          mode: mergepartial
+          outputColumnNames: _col0, _col1, _col2
+          Select Operator
+            expressions:
+                  expr: _col0
+                  type: string
+                  expr: _col2
+                  type: bigint
+            outputColumnNames: _col0, _col1
+            Limit
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: 10
+
+
+PREHOOK: query: select key, count(1) from clustergroupby  where ds='102'  group by key, value limit 10
+PREHOOK: type: QUERY
+PREHOOK: Input: default@clustergroupby@ds=102
+PREHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/971850477/10000
+POSTHOOK: query: select key, count(1) from clustergroupby  where ds='102'  group by key, value limit 10
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@clustergroupby@ds=102
+POSTHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/971850477/10000
+0	3
+10	1
+100	2
+103	2
+104	2
+105	1
+11	1
+111	1
+113	2
+114	1
+PREHOOK: query: alter table clustergroupby clustered by (value, key) sorted by (key) into 1 buckets
+PREHOOK: type: null
+POSTHOOK: query: alter table clustergroupby clustered by (value, key) sorted by (key) into 1 buckets
+POSTHOOK: type: null
+POSTHOOK: Input: default@clustergroupby
+POSTHOOK: Output: default@clustergroupby
+PREHOOK: query: describe extended clustergroupby
+PREHOOK: type: DESCTABLE
+POSTHOOK: query: describe extended clustergroupby
+POSTHOOK: type: DESCTABLE
+key	string	
+value	string	
+ds	string	
+	 	 
+Detailed Table Information	Table(tableName:clustergroupby, dbName:default, owner:heyongqiang, createTime:1259868796, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/test/data/warehouse/clustergroupby, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[value, key], sortCols:[Order(col:key, order:1)], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null)], parameters:{last_modified_by=heyongqiang,last_modified_time=1259868885,transient_lastDdlTime=1259868885})	
+PREHOOK: query: insert overwrite table clustergroupby partition (ds='103') select key, value from src distribute by value, key sort by key
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@clustergroupby@ds=103
+POSTHOOK: query: insert overwrite table clustergroupby partition (ds='103') select key, value from src distribute by value, key sort by key
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@clustergroupby@ds=103
+PREHOOK: query: explain
+select key, count(1) from clustergroupby  where ds='103'  group by key limit 10
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+select key, count(1) from clustergroupby  where ds='103'  group by key limit 10
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+  (TOK_QUERY (TOK_FROM (TOK_TABREF clustergroupby)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_FUNCTION count 1))) (TOK_WHERE (= (TOK_TABLE_OR_COL ds) '103')) (TOK_GROUPBY (TOK_TABLE_OR_COL key)) (TOK_LIMIT 10)))
+
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Alias -> Map Operator Tree:
+        clustergroupby 
+          TableScan
+            alias: clustergroupby
+            Filter Operator
+              predicate:
+                  expr: (ds = '103')
+                  type: boolean
+              Filter Operator
+                predicate:
+                    expr: (ds = '103')
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: key
+                        type: string
+                  outputColumnNames: key
+                  Group By Operator
+                    aggregations:
+                          expr: count(1)
+                    bucketGroup: true
+                    keys:
+                          expr: key
+                          type: string
+                    mode: hash
+                    outputColumnNames: _col0, _col1
+                    Reduce Output Operator
+                      key expressions:
+                            expr: _col0
+                            type: string
+                      sort order: +
+                      Map-reduce partition columns:
+                            expr: _col0
+                            type: string
+                      tag: -1
+                      value expressions:
+                            expr: _col1
+                            type: bigint
+      Reduce Operator Tree:
+        Group By Operator
+          aggregations:
+                expr: count(VALUE._col0)
+          bucketGroup: false
+          keys:
+                expr: KEY._col0
+                type: string
+          mode: mergepartial
+          outputColumnNames: _col0, _col1
+          Select Operator
+            expressions:
+                  expr: _col0
+                  type: string
+                  expr: _col1
+                  type: bigint
+            outputColumnNames: _col0, _col1
+            Limit
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: 10
+
+
+PREHOOK: query: select key, count(1) from clustergroupby  where ds='103' group by key limit 10
+PREHOOK: type: QUERY
+PREHOOK: Input: default@clustergroupby@ds=103
+PREHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/116849772/10000
+POSTHOOK: query: select key, count(1) from clustergroupby  where ds='103' group by key limit 10
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@clustergroupby@ds=103
+POSTHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/116849772/10000
+0	3
+10	1
+100	2
+103	2
+104	2
+105	1
+11	1
+111	1
+113	2
+114	1
+PREHOOK: query: explain
+select key, count(1) from clustergroupby  where ds='103'  group by value, key limit 10
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+select key, count(1) from clustergroupby  where ds='103'  group by value, key limit 10
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+  (TOK_QUERY (TOK_FROM (TOK_TABREF clustergroupby)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_FUNCTION count 1))) (TOK_WHERE (= (TOK_TABLE_OR_COL ds) '103')) (TOK_GROUPBY (TOK_TABLE_OR_COL value) (TOK_TABLE_OR_COL key)) (TOK_LIMIT 10)))
+
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Alias -> Map Operator Tree:
+        clustergroupby 
+          TableScan
+            alias: clustergroupby
+            Filter Operator
+              predicate:
+                  expr: (ds = '103')
+                  type: boolean
+              Filter Operator
+                predicate:
+                    expr: (ds = '103')
+                    type: boolean
+                Select Operator
+                  expressions:
+                        expr: value
+                        type: string
+                        expr: key
+                        type: string
+                  outputColumnNames: value, key
+                  Group By Operator
+                    aggregations:
+                          expr: count(1)
+                    bucketGroup: false
+                    keys:
+                          expr: value
+                          type: string
+                          expr: key
+                          type: string
+                    mode: hash
+                    outputColumnNames: _col0, _col1, _col2
+                    Reduce Output Operator
+                      key expressions:
+                            expr: _col0
+                            type: string
+                            expr: _col1
+                            type: string
+                      sort order: ++
+                      Map-reduce partition columns:
+                            expr: _col0
+                            type: string
+                            expr: _col1
+                            type: string
+                      tag: -1
+                      value expressions:
+                            expr: _col2
+                            type: bigint
+      Reduce Operator Tree:
+        Group By Operator
+          aggregations:
+                expr: count(VALUE._col0)
+          bucketGroup: false
+          keys:
+                expr: KEY._col0
+                type: string
+                expr: KEY._col1
+                type: string
+          mode: mergepartial
+          outputColumnNames: _col0, _col1, _col2
+          Select Operator
+            expressions:
+                  expr: _col1
+                  type: string
+                  expr: _col2
+                  type: bigint
+            outputColumnNames: _col0, _col1
+            Limit
+              File Output Operator
+                compressed: false
+                GlobalTableId: 0
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: 10
+
+
+PREHOOK: query: select key, count(1) from clustergroupby  where ds='103' group by  value, key limit 10
+PREHOOK: type: QUERY
+PREHOOK: Input: default@clustergroupby@ds=103
+PREHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/881916878/10000
+POSTHOOK: query: select key, count(1) from clustergroupby  where ds='103' group by  value, key limit 10
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@clustergroupby@ds=103
+POSTHOOK: Output: file:/Users/heyongqiang/Documents/workspace/Hive-trunk/build/ql/tmp/881916878/10000
+0	3
+10	1
+100	2
+103	2
+104	2
+105	1
+11	1
+111	1
+113	2
+114	1
+PREHOOK: query: drop table clustergroupby
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table clustergroupby
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Output: default@clustergroupby

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/create_genericudaf.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/create_genericudaf.q.out?rev=887057&r1=887056&r2=887057&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/create_genericudaf.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/create_genericudaf.q.out Fri Dec  4 04:39:37 2009
@@ -53,6 +53,7 @@
                 aggregations:
                       expr: test_avg(1)
                       expr: test_avg(substr(value, 5))
+                bucketGroup: false
                 mode: hash
                 outputColumnNames: _col0, _col1
                 Reduce Output Operator
@@ -68,6 +69,7 @@
           aggregations:
                 expr: test_avg(VALUE._col0)
                 expr: test_avg(VALUE._col1)
+          bucketGroup: false
           mode: mergepartial
           outputColumnNames: _col0, _col1
           Select Operator
@@ -95,14 +97,14 @@
 FROM src
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1951411616/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/942592102/10000
 POSTHOOK: query: SELECT
     test_avg(1),
     test_avg(substr(value,5))
 FROM src
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1951411616/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/942592102/10000
 1.0	260.182
 PREHOOK: query: DROP TEMPORARY FUNCTIOn test_avg
 PREHOOK: type: DROPFUNCTION

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1.q.out?rev=887057&r1=887056&r2=887057&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1.q.out Fri Dec  4 04:39:37 2009
@@ -47,6 +47,7 @@
         Group By Operator
           aggregations:
                 expr: sum(VALUE._col0)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -62,7 +63,7 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        invalidscheme:/data/users/njain/hive5/hive5/.ptest_3/build/ql/tmp/1275589613/10002 
+        invalidscheme:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1515720962/10002 
             Reduce Output Operator
               key expressions:
                     expr: _col0
@@ -79,6 +80,7 @@
         Group By Operator
           aggregations:
                 expr: sum(VALUE._col0)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -129,11 +131,11 @@
 PREHOOK: query: SELECT dest_g1.* FROM dest_g1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest_g1
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/.ptest_3/build/ql/tmp/1595205675/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1461207051/10000
 POSTHOOK: query: SELECT dest_g1.* FROM dest_g1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest_g1
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/.ptest_3/build/ql/tmp/1595205675/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1461207051/10000
 0	0.0
 10	10.0
 100	200.0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby10.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby10.q.out?rev=887057&r1=887056&r2=887057&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby10.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby10.q.out Fri Dec  4 04:39:37 2009
@@ -75,6 +75,7 @@
             aggregations:
                   expr: count(KEY._col0)
                   expr: count(DISTINCT KEY._col0)
+            bucketGroup: false
             keys:
                   expr: VALUE._col0
                   type: int
@@ -90,6 +91,7 @@
             aggregations:
                   expr: sum(KEY._col0)
                   expr: sum(DISTINCT KEY._col0)
+            bucketGroup: false
             keys:
                   expr: VALUE._col0
                   type: int
@@ -105,7 +107,7 @@
   Stage: Stage-3
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/njain/hive5/hive5/build/ql/tmp/1745933773/10004 
+        file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1034721639/10004 
             Reduce Output Operator
               key expressions:
                     expr: _col0
@@ -125,6 +127,7 @@
           aggregations:
                 expr: count(VALUE._col0)
                 expr: count(VALUE._col1)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: int
@@ -170,7 +173,7 @@
   Stage: Stage-4
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/njain/hive5/hive5/build/ql/tmp/1745933773/10005 
+        file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1034721639/10005 
             Reduce Output Operator
               key expressions:
                     expr: _col0
@@ -190,6 +193,7 @@
           aggregations:
                 expr: sum(VALUE._col0)
                 expr: sum(VALUE._col1)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: int
@@ -250,11 +254,11 @@
 PREHOOK: query: SELECT * from dest1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/810841334/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1454167877/10000
 POSTHOOK: query: SELECT * from dest1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/810841334/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1454167877/10000
 27	1	1
 66	1	1
 86	1	1
@@ -278,11 +282,11 @@
 PREHOOK: query: SELECT * from dest2
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest2
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/566919289/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/750550239/10000
 POSTHOOK: query: SELECT * from dest2
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest2
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/566919289/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/750550239/10000
 27	27	27
 66	66	66
 86	86	86

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby11.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby11.q.out?rev=887057&r1=887056&r2=887057&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby11.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby11.q.out Fri Dec  4 04:39:37 2009
@@ -67,6 +67,7 @@
             aggregations:
                   expr: count(KEY._col0)
                   expr: count(DISTINCT KEY._col0)
+            bucketGroup: false
             keys:
                   expr: VALUE._col0
                   type: string
@@ -82,6 +83,7 @@
             aggregations:
                   expr: count(KEY._col0)
                   expr: count(DISTINCT KEY._col0)
+            bucketGroup: false
             keys:
                   expr: VALUE._col1
                   type: string
@@ -97,7 +99,7 @@
   Stage: Stage-3
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/njain/hive5/hive5/build/ql/tmp/799679609/10004 
+        file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1187956750/10004 
             Reduce Output Operator
               key expressions:
                     expr: _col0
@@ -117,6 +119,7 @@
           aggregations:
                 expr: count(VALUE._col0)
                 expr: count(VALUE._col1)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -164,7 +167,7 @@
   Stage: Stage-4
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/njain/hive5/hive5/build/ql/tmp/799679609/10005 
+        file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1187956750/10005 
             Reduce Output Operator
               key expressions:
                     expr: _col0
@@ -184,6 +187,7 @@
           aggregations:
                 expr: count(VALUE._col0)
                 expr: count(VALUE._col1)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -250,11 +254,11 @@
 PREHOOK: query: SELECT * from dest1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest1@ds=111
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1044642512/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1884622458/10000
 POSTHOOK: query: SELECT * from dest1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest1@ds=111
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1044642512/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1884622458/10000
 val_0	3	1	111
 val_10	1	1	111
 val_100	2	1	111
@@ -567,11 +571,11 @@
 PREHOOK: query: SELECT * from dest2
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest2@ds=111
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1431820508/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1919394559/10000
 POSTHOOK: query: SELECT * from dest2
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest2@ds=111
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1431820508/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1919394559/10000
 0	3	1	111
 10	1	1	111
 100	2	1	111

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_limit.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_limit.q.out?rev=887057&r1=887056&r2=887057&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_limit.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_limit.q.out Fri Dec  4 04:39:37 2009
@@ -34,6 +34,7 @@
               Group By Operator
                 aggregations:
                       expr: sum(substr(value, 5))
+                bucketGroup: false
                 keys:
                       expr: key
                       type: string
@@ -55,6 +56,7 @@
         Group By Operator
           aggregations:
                 expr: sum(VALUE._col0)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -78,7 +80,7 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/njain/hive5/hive5/build/ql/tmp/1010104085/10002 
+        file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/780128058/10002 
             Reduce Output Operator
               sort order: 
               tag: -1
@@ -128,11 +130,11 @@
 PREHOOK: query: SELECT dest1.* FROM dest1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/776994640/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1678285663/10000
 POSTHOOK: query: SELECT dest1.* FROM dest1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/776994640/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1678285663/10000
 0	0.0
 10	10.0
 100	200.0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_map.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_map.q.out?rev=887057&r1=887056&r2=887057&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_map.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_map.q.out Fri Dec  4 04:39:37 2009
@@ -33,6 +33,7 @@
               Group By Operator
                 aggregations:
                       expr: sum(substr(value, 5))
+                bucketGroup: false
                 keys:
                       expr: key
                       type: string
@@ -54,6 +55,7 @@
         Group By Operator
           aggregations:
                 expr: sum(VALUE._col0)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -104,11 +106,11 @@
 PREHOOK: query: SELECT dest1.* FROM dest1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/759116141/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/233901356/10000
 POSTHOOK: query: SELECT dest1.* FROM dest1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/759116141/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/233901356/10000
 0	0.0
 10	10.0
 100	200.0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_map_nomap.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_map_nomap.q.out?rev=887057&r1=887056&r2=887057&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_map_nomap.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_map_nomap.q.out Fri Dec  4 04:39:37 2009
@@ -33,6 +33,7 @@
               Group By Operator
                 aggregations:
                       expr: sum(substr(value, 5))
+                bucketGroup: false
                 keys:
                       expr: key
                       type: string
@@ -54,6 +55,7 @@
         Group By Operator
           aggregations:
                 expr: sum(VALUE._col0)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -104,11 +106,11 @@
 PREHOOK: query: SELECT dest1.* FROM dest1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/2053916777/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/156800012/10000
 POSTHOOK: query: SELECT dest1.* FROM dest1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/2053916777/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/156800012/10000
 0	0.0
 10	10.0
 100	200.0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_map_skew.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_map_skew.q.out?rev=887057&r1=887056&r2=887057&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_map_skew.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_map_skew.q.out Fri Dec  4 04:39:37 2009
@@ -34,6 +34,7 @@
               Group By Operator
                 aggregations:
                       expr: sum(substr(value, 5))
+                bucketGroup: false
                 keys:
                       expr: key
                       type: string
@@ -55,6 +56,7 @@
         Group By Operator
           aggregations:
                 expr: sum(VALUE._col0)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -70,7 +72,7 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/njain/hive5/hive5/build/ql/tmp/1252089764/10002 
+        file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/858199914/10002 
             Reduce Output Operator
               key expressions:
                     expr: _col0
@@ -87,6 +89,7 @@
         Group By Operator
           aggregations:
                 expr: sum(VALUE._col0)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -137,11 +140,11 @@
 PREHOOK: query: SELECT dest1.* FROM dest1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1734846725/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1273565267/10000
 POSTHOOK: query: SELECT dest1.* FROM dest1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1734846725/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1273565267/10000
 0	0.0
 10	10.0
 100	200.0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_noskew.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_noskew.q.out?rev=887057&r1=887056&r2=887057&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_noskew.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_noskew.q.out Fri Dec  4 04:39:37 2009
@@ -46,6 +46,7 @@
         Group By Operator
           aggregations:
                 expr: sum(VALUE._col0)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -96,11 +97,11 @@
 PREHOOK: query: SELECT dest_g1.* FROM dest_g1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest_g1
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/537324321/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1084528559/10000
 POSTHOOK: query: SELECT dest_g1.* FROM dest_g1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest_g1
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/537324321/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1084528559/10000
 0	0.0
 10	10.0
 100	200.0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2.q.out?rev=887057&r1=887056&r2=887057&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2.q.out Fri Dec  4 04:39:37 2009
@@ -51,6 +51,7 @@
           aggregations:
                 expr: count(DISTINCT KEY._col1)
                 expr: sum(KEY._col1)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -66,7 +67,7 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/njain/hive5/hive5/.ptest_3/build/ql/tmp/1120967939/10002 
+        file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/2060806095/10002 
             Reduce Output Operator
               key expressions:
                     expr: _col0
@@ -86,6 +87,7 @@
           aggregations:
                 expr: count(VALUE._col0)
                 expr: sum(VALUE._col1)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -142,11 +144,11 @@
 PREHOOK: query: SELECT dest_g2.* FROM dest_g2
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest_g2
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/.ptest_3/build/ql/tmp/923386772/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1250974216/10000
 POSTHOOK: query: SELECT dest_g2.* FROM dest_g2
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest_g2
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/.ptest_3/build/ql/tmp/923386772/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1250974216/10000
 0	1	00.0
 1	71	116414.0
 2	69	225571.0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_limit.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_limit.q.out?rev=887057&r1=887056&r2=887057&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_limit.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_limit.q.out Fri Dec  4 04:39:37 2009
@@ -28,6 +28,7 @@
               Group By Operator
                 aggregations:
                       expr: sum(substr(value, 5))
+                bucketGroup: false
                 keys:
                       expr: key
                       type: string
@@ -49,6 +50,7 @@
         Group By Operator
           aggregations:
                 expr: sum(VALUE._col0)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -77,11 +79,11 @@
 PREHOOK: query: SELECT src.key, sum(substr(src.value,5)) FROM src GROUP BY src.key LIMIT 5
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/2031440533/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/184969854/10000
 POSTHOOK: query: SELECT src.key, sum(substr(src.value,5)) FROM src GROUP BY src.key LIMIT 5
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/2031440533/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/184969854/10000
 0	0.0
 10	10.0
 100	200.0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_map.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_map.q.out?rev=887057&r1=887056&r2=887057&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_map.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_map.q.out Fri Dec  4 04:39:37 2009
@@ -36,6 +36,7 @@
                 aggregations:
                       expr: count(DISTINCT substr(value, 5))
                       expr: sum(substr(value, 5))
+                bucketGroup: false
                 keys:
                       expr: substr(key, 1, 1)
                       type: string
@@ -64,6 +65,7 @@
           aggregations:
                 expr: count(DISTINCT KEY._col1)
                 expr: sum(VALUE._col1)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -120,11 +122,11 @@
 PREHOOK: query: SELECT dest1.* FROM dest1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1046632747/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/615915015/10000
 POSTHOOK: query: SELECT dest1.* FROM dest1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1046632747/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/615915015/10000
 0	1	00.0
 1	71	116414.0
 2	69	225571.0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_map_skew.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_map_skew.q.out?rev=887057&r1=887056&r2=887057&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_map_skew.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_map_skew.q.out Fri Dec  4 04:39:37 2009
@@ -37,6 +37,7 @@
                 aggregations:
                       expr: count(DISTINCT substr(value, 5))
                       expr: sum(substr(value, 5))
+                bucketGroup: false
                 keys:
                       expr: substr(key, 1, 1)
                       type: string
@@ -67,6 +68,7 @@
           aggregations:
                 expr: count(DISTINCT KEY._col1)
                 expr: sum(VALUE._col1)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -82,7 +84,7 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/njain/hive5/hive5/build/ql/tmp/365164854/10002 
+        file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/715055378/10002 
             Reduce Output Operator
               key expressions:
                     expr: _col0
@@ -102,6 +104,7 @@
           aggregations:
                 expr: count(VALUE._col0)
                 expr: sum(VALUE._col1)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -158,11 +161,11 @@
 PREHOOK: query: SELECT dest1.* FROM dest1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/927058236/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1863077019/10000
 POSTHOOK: query: SELECT dest1.* FROM dest1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/927058236/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1863077019/10000
 0	1	00.0
 1	71	116414.0
 2	69	225571.0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_noskew.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_noskew.q.out?rev=887057&r1=887056&r2=887057&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_noskew.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_noskew.q.out Fri Dec  4 04:39:37 2009
@@ -48,6 +48,7 @@
           aggregations:
                 expr: count(DISTINCT KEY._col1)
                 expr: sum(KEY._col1)
+          bucketGroup: false
           keys:
                 expr: KEY._col0
                 type: string
@@ -104,11 +105,11 @@
 PREHOOK: query: SELECT dest_g2.* FROM dest_g2
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest_g2
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1459670670/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1769464573/10000
 POSTHOOK: query: SELECT dest_g2.* FROM dest_g2
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest_g2
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1459670670/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1769464573/10000
 0	1	00.0
 1	71	116414.0
 2	69	225571.0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby3.q.out?rev=887057&r1=887056&r2=887057&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby3.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby3.q.out Fri Dec  4 04:39:37 2009
@@ -70,6 +70,7 @@
                 expr: stddev_samp(KEY._col0)
                 expr: variance(KEY._col0)
                 expr: var_samp(KEY._col0)
+          bucketGroup: false
           mode: partial1
           outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8
           File Output Operator
@@ -82,7 +83,7 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/njain/hive5/hive5/build/ql/tmp/837344184/10002 
+        file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/925660734/10002 
             Reduce Output Operator
               sort order: 
               tag: -1
@@ -117,6 +118,7 @@
                 expr: stddev_samp(VALUE._col6)
                 expr: variance(VALUE._col7)
                 expr: var_samp(VALUE._col8)
+          bucketGroup: false
           mode: final
           outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8
           Select Operator
@@ -212,11 +214,11 @@
 PREHOOK: query: SELECT dest1.* FROM dest1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/837906913/10000
+PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/110467196/10000
 POSTHOOK: query: SELECT dest1.* FROM dest1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/837906913/10000
+POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/110467196/10000
 130091.0	260.182	256.10355987055016	98.0	0.0	142.92680950752379	143.06995106518903	20428.072875999995	20469.010897795586
 PREHOOK: query: DROP TABLE dest1
 PREHOOK: type: DROPTABLE



Mime
View raw message