hadoop-hive-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From nzh...@apache.org
Subject svn commit: r909965 [6/15] - in /hadoop/hive/trunk: ./ contrib/src/test/results/clientnegative/ contrib/src/test/results/clientpositive/ ql/src/java/org/apache/hadoop/hive/ql/exec/ ql/src/test/results/clientpositive/
Date Sun, 14 Feb 2010 02:31:44 GMT
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join_map_ppr.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join_map_ppr.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join_map_ppr.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join_map_ppr.q.out Sun Feb 14 02:31:40 2010
@@ -84,37 +84,37 @@
                       File Output Operator
                         compressed: false
                         GlobalTableId: 1
-                        directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1886349005/10002
+                        directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10002
                         table:
                             input format: org.apache.hadoop.mapred.TextInputFormat
                             output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                             properties:
-                              name dest_j1
+                              bucket_count -1
+                              columns key,value,val2
                               columns.types string:string:string
+                              file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                              location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+                              name dest_j1
                               serialization.ddl struct dest_j1 { string key, string value, string val2}
                               serialization.format 1
-                              columns key,value,val2
-                              bucket_count -1
                               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                              file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                              location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
-                              transient_lastDdlTime 1263595165
+                              transient_lastDdlTime 1266042780
                             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                             name: dest_j1
       Local Work:
         Map Reduce Local Work
           Alias -> Map Local Tables:
-            y 
+            x 
               Fetch Operator
                 limit: -1
-            x 
+            y 
               Fetch Operator
                 limit: -1
           Alias -> Map Local Operator Tree:
-            y 
+            x 
               TableScan
-                alias: y
+                alias: x
                 Common Join Operator
                   condition map:
                        Inner Join 0 to 1
@@ -160,27 +160,27 @@
                         File Output Operator
                           compressed: false
                           GlobalTableId: 1
-                          directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1886349005/10002
+                          directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10002
                           table:
                               input format: org.apache.hadoop.mapred.TextInputFormat
                               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                               properties:
-                                name dest_j1
+                                bucket_count -1
+                                columns key,value,val2
                                 columns.types string:string:string
+                                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+                                name dest_j1
                                 serialization.ddl struct dest_j1 { string key, string value, string val2}
                                 serialization.format 1
-                                columns key,value,val2
-                                bucket_count -1
                                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                                location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
-                                transient_lastDdlTime 1263595165
+                                transient_lastDdlTime 1266042780
                               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                               name: dest_j1
-            x 
+            y 
               TableScan
-                alias: x
+                alias: y
                 Common Join Operator
                   condition map:
                        Inner Join 0 to 1
@@ -226,29 +226,29 @@
                         File Output Operator
                           compressed: false
                           GlobalTableId: 1
-                          directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1886349005/10002
+                          directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10002
                           table:
                               input format: org.apache.hadoop.mapred.TextInputFormat
                               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                               properties:
-                                name dest_j1
+                                bucket_count -1
+                                columns key,value,val2
                                 columns.types string:string:string
+                                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+                                name dest_j1
                                 serialization.ddl struct dest_j1 { string key, string value, string val2}
                                 serialization.format 1
-                                columns key,value,val2
-                                bucket_count -1
                                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                                location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
-                                transient_lastDdlTime 1263595165
+                                transient_lastDdlTime 1266042780
                               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                               name: dest_j1
       Needs Tagging: false
       Path -> Alias:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z]
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z]
       Path -> Partition:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
           Partition
             base file name: hr=11
             input format: org.apache.hadoop.mapred.TextInputFormat
@@ -257,35 +257,35 @@
               ds 2008-04-08
               hr 11
             properties:
-              name srcpart
+              bucket_count -1
+              columns key,value
               columns.types string:string
+              file.inputformat org.apache.hadoop.mapred.TextInputFormat
+              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+              name srcpart
+              partition_columns ds/hr
               serialization.ddl struct srcpart { string key, string value}
               serialization.format 1
-              columns key,value
-              partition_columns ds/hr
-              bucket_count -1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              file.inputformat org.apache.hadoop.mapred.TextInputFormat
-              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/srcpart
-              transient_lastDdlTime 1263595164
+              transient_lastDdlTime 1266042778
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
-                name srcpart
+                bucket_count -1
+                columns key,value
                 columns.types string:string
+                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+                name srcpart
+                partition_columns ds/hr
                 serialization.ddl struct srcpart { string key, string value}
                 serialization.format 1
-                columns key,value
-                partition_columns ds/hr
-                bucket_count -1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/srcpart
-                transient_lastDdlTime 1263595164
+                transient_lastDdlTime 1266042778
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
             name: srcpart
@@ -297,37 +297,37 @@
     Move Operator
       files:
           hdfs directory: true
-          source: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1886349005/10002
-          destination: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/141026855/10000
+          source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10002
+          destination: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10000
 
   Stage: Stage-0
     Move Operator
       tables:
           replace: true
-          source: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/141026855/10000
+          source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10000
           table:
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
-                name dest_j1
+                bucket_count -1
+                columns key,value,val2
                 columns.types string:string:string
+                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+                name dest_j1
                 serialization.ddl struct dest_j1 { string key, string value, string val2}
                 serialization.format 1
-                columns key,value,val2
-                bucket_count -1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
-                transient_lastDdlTime 1263595165
+                transient_lastDdlTime 1266042780
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: dest_j1
-          tmp directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/141026855/10001
+          tmp directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10001
 
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1886349005/10002 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10002 
             Reduce Output Operator
               sort order: 
               Map-reduce partition columns:
@@ -343,41 +343,41 @@
                     type: string
       Needs Tagging: false
       Path -> Alias:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1886349005/10002 [file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1886349005/10002]
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10002 [file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10002]
       Path -> Partition:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1886349005/10002 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10002 
           Partition
             base file name: 10002
             input format: org.apache.hadoop.mapred.TextInputFormat
             output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
             properties:
-              name dest_j1
+              bucket_count -1
+              columns key,value,val2
               columns.types string:string:string
+              file.inputformat org.apache.hadoop.mapred.TextInputFormat
+              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+              name dest_j1
               serialization.ddl struct dest_j1 { string key, string value, string val2}
               serialization.format 1
-              columns key,value,val2
-              bucket_count -1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              file.inputformat org.apache.hadoop.mapred.TextInputFormat
-              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
-              transient_lastDdlTime 1263595165
+              transient_lastDdlTime 1266042780
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
-                name dest_j1
+                bucket_count -1
+                columns key,value,val2
                 columns.types string:string:string
+                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+                name dest_j1
                 serialization.ddl struct dest_j1 { string key, string value, string val2}
                 serialization.format 1
-                columns key,value,val2
-                bucket_count -1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
-                transient_lastDdlTime 1263595165
+                transient_lastDdlTime 1266042780
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: dest_j1
             name: dest_j1
@@ -386,22 +386,22 @@
           File Output Operator
             compressed: false
             GlobalTableId: 0
-            directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/141026855/10000
+            directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-00_399_352290245960327717/10000
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                 properties:
-                  name dest_j1
+                  bucket_count -1
+                  columns key,value,val2
                   columns.types string:string:string
+                  file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                  file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+                  name dest_j1
                   serialization.ddl struct dest_j1 { string key, string value, string val2}
                   serialization.format 1
-                  columns key,value,val2
-                  bucket_count -1
                   serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                  file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                  location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
-                  file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  transient_lastDdlTime 1263595165
+                  transient_lastDdlTime 1266042780
                 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 name: dest_j1
 
@@ -429,11 +429,11 @@
 PREHOOK: query: select * from dest_j1 x order by x.key
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest_j1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/886762969/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-06_516_3338852843532728387/10000
 POSTHOOK: query: select * from dest_j1 x order by x.key
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest_j1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/886762969/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-06_516_3338852843532728387/10000
 128	val_128	val_128
 128	val_128	val_128
 128	val_128	val_128
@@ -648,37 +648,37 @@
                       File Output Operator
                         compressed: false
                         GlobalTableId: 1
-                        directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1775434766/10002
+                        directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10002
                         table:
                             input format: org.apache.hadoop.mapred.TextInputFormat
                             output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                             properties:
-                              name dest_j1
+                              bucket_count -1
+                              columns key,value,val2
                               columns.types string:string:string
+                              file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                              location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+                              name dest_j1
                               serialization.ddl struct dest_j1 { string key, string value, string val2}
                               serialization.format 1
-                              columns key,value,val2
-                              bucket_count -1
                               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                              file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                              location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
-                              transient_lastDdlTime 1263595165
+                              transient_lastDdlTime 1266042780
                             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                             name: dest_j1
       Local Work:
         Map Reduce Local Work
           Alias -> Map Local Tables:
-            y 
+            x 
               Fetch Operator
                 limit: -1
-            x 
+            y 
               Fetch Operator
                 limit: -1
           Alias -> Map Local Operator Tree:
-            y 
+            x 
               TableScan
-                alias: y
+                alias: x
                 Common Join Operator
                   condition map:
                        Inner Join 0 to 1
@@ -724,27 +724,27 @@
                         File Output Operator
                           compressed: false
                           GlobalTableId: 1
-                          directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1775434766/10002
+                          directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10002
                           table:
                               input format: org.apache.hadoop.mapred.TextInputFormat
                               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                               properties:
-                                name dest_j1
+                                bucket_count -1
+                                columns key,value,val2
                                 columns.types string:string:string
+                                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+                                name dest_j1
                                 serialization.ddl struct dest_j1 { string key, string value, string val2}
                                 serialization.format 1
-                                columns key,value,val2
-                                bucket_count -1
                                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                                location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
-                                transient_lastDdlTime 1263595165
+                                transient_lastDdlTime 1266042780
                               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                               name: dest_j1
-            x 
+            y 
               TableScan
-                alias: x
+                alias: y
                 Common Join Operator
                   condition map:
                        Inner Join 0 to 1
@@ -790,29 +790,29 @@
                         File Output Operator
                           compressed: false
                           GlobalTableId: 1
-                          directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1775434766/10002
+                          directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10002
                           table:
                               input format: org.apache.hadoop.mapred.TextInputFormat
                               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                               properties:
-                                name dest_j1
+                                bucket_count -1
+                                columns key,value,val2
                                 columns.types string:string:string
+                                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+                                name dest_j1
                                 serialization.ddl struct dest_j1 { string key, string value, string val2}
                                 serialization.format 1
-                                columns key,value,val2
-                                bucket_count -1
                                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                                location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
-                                transient_lastDdlTime 1263595165
+                                transient_lastDdlTime 1266042780
                               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                               name: dest_j1
       Needs Tagging: false
       Path -> Alias:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z]
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z]
       Path -> Partition:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
           Partition
             base file name: hr=11
             input format: org.apache.hadoop.mapred.TextInputFormat
@@ -821,35 +821,35 @@
               ds 2008-04-08
               hr 11
             properties:
-              name srcpart
+              bucket_count -1
+              columns key,value
               columns.types string:string
+              file.inputformat org.apache.hadoop.mapred.TextInputFormat
+              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+              name srcpart
+              partition_columns ds/hr
               serialization.ddl struct srcpart { string key, string value}
               serialization.format 1
-              columns key,value
-              partition_columns ds/hr
-              bucket_count -1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              file.inputformat org.apache.hadoop.mapred.TextInputFormat
-              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/srcpart
-              transient_lastDdlTime 1263595164
+              transient_lastDdlTime 1266042778
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
-                name srcpart
+                bucket_count -1
+                columns key,value
                 columns.types string:string
+                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/srcpart
+                name srcpart
+                partition_columns ds/hr
                 serialization.ddl struct srcpart { string key, string value}
                 serialization.format 1
-                columns key,value
-                partition_columns ds/hr
-                bucket_count -1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/srcpart
-                transient_lastDdlTime 1263595164
+                transient_lastDdlTime 1266042778
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
             name: srcpart
@@ -861,37 +861,37 @@
     Move Operator
       files:
           hdfs directory: true
-          source: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1775434766/10002
-          destination: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1505671688/10000
+          source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10002
+          destination: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10000
 
   Stage: Stage-0
     Move Operator
       tables:
           replace: true
-          source: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1505671688/10000
+          source: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10000
           table:
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
-                name dest_j1
+                bucket_count -1
+                columns key,value,val2
                 columns.types string:string:string
+                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+                name dest_j1
                 serialization.ddl struct dest_j1 { string key, string value, string val2}
                 serialization.format 1
-                columns key,value,val2
-                bucket_count -1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
-                transient_lastDdlTime 1263595165
+                transient_lastDdlTime 1266042780
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: dest_j1
-          tmp directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1505671688/10001
+          tmp directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10001
 
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1775434766/10002 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10002 
             Reduce Output Operator
               sort order: 
               Map-reduce partition columns:
@@ -907,41 +907,41 @@
                     type: string
       Needs Tagging: false
       Path -> Alias:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1775434766/10002 [file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1775434766/10002]
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10002 [file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10002]
       Path -> Partition:
-        file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1775434766/10002 
+        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10002 
           Partition
             base file name: 10002
             input format: org.apache.hadoop.mapred.TextInputFormat
             output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
             properties:
-              name dest_j1
+              bucket_count -1
+              columns key,value,val2
               columns.types string:string:string
+              file.inputformat org.apache.hadoop.mapred.TextInputFormat
+              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+              location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+              name dest_j1
               serialization.ddl struct dest_j1 { string key, string value, string val2}
               serialization.format 1
-              columns key,value,val2
-              bucket_count -1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              file.inputformat org.apache.hadoop.mapred.TextInputFormat
-              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
-              transient_lastDdlTime 1263595165
+              transient_lastDdlTime 1266042780
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
               properties:
-                name dest_j1
+                bucket_count -1
+                columns key,value,val2
                 columns.types string:string:string
+                file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+                name dest_j1
                 serialization.ddl struct dest_j1 { string key, string value, string val2}
                 serialization.format 1
-                columns key,value,val2
-                bucket_count -1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
-                transient_lastDdlTime 1263595165
+                transient_lastDdlTime 1266042780
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: dest_j1
             name: dest_j1
@@ -950,22 +950,22 @@
           File Output Operator
             compressed: false
             GlobalTableId: 0
-            directory: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/1505671688/10000
+            directory: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-20_130_2089622276113662506/10000
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                 properties:
-                  name dest_j1
+                  bucket_count -1
+                  columns key,value,val2
                   columns.types string:string:string
+                  file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                  file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  location file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/test/data/warehouse/dest_j1
+                  name dest_j1
                   serialization.ddl struct dest_j1 { string key, string value, string val2}
                   serialization.format 1
-                  columns key,value,val2
-                  bucket_count -1
                   serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                  file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                  location file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/test/data/warehouse/dest_j1
-                  file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  transient_lastDdlTime 1263595165
+                  transient_lastDdlTime 1266042780
                 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 name: dest_j1
 
@@ -993,11 +993,11 @@
 PREHOOK: query: select * from dest_j1 x order by x.key
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest_j1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/877753191/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-26_697_820182467019293566/10000
 POSTHOOK: query: select * from dest_j1 x order by x.key
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest_j1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/877753191/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-26_697_820182467019293566/10000
 128	val_128	val_128
 128	val_128	val_128
 128	val_128	val_128

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join_rc.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join_rc.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join_rc.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join_rc.q.out Sun Feb 14 02:31:40 2010
@@ -51,9 +51,9 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        join_rc2 
+        join_rc1 
           TableScan
-            alias: join_rc2
+            alias: join_rc1
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -62,13 +62,13 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 1
+              tag: 0
               value expressions:
-                    expr: value
+                    expr: key
                     type: string
-        join_rc1 
+        join_rc2 
           TableScan
-            alias: join_rc1
+            alias: join_rc2
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -77,9 +77,9 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 0
+              tag: 1
               value expressions:
-                    expr: key
+                    expr: value
                     type: string
       Reduce Operator Tree:
         Join Operator
@@ -114,13 +114,13 @@
 PREHOOK: type: QUERY
 PREHOOK: Input: default@join_rc2
 PREHOOK: Input: default@join_rc1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1138840055/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-27-47_640_1441550174396831657/10000
 POSTHOOK: query: select join_rc1.key, join_rc2.value
 FROM join_rc1 JOIN join_rc2 ON join_rc1.key = join_rc2.key
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@join_rc2
 POSTHOOK: Input: default@join_rc1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1138840055/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-27-47_640_1441550174396831657/10000
 0	val_0
 0	val_0
 0	val_0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder.q.out Sun Feb 14 02:31:40 2010
@@ -57,38 +57,38 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        c 
+        a 
           TableScan
-            alias: c
+            alias: a
             Reduce Output Operator
               key expressions:
-                    expr: (key + 1)
+                    expr: UDFToDouble(key)
                     type: double
               sort order: +
               Map-reduce partition columns:
-                    expr: (key + 1)
+                    expr: UDFToDouble(key)
                     type: double
-              tag: 1
+              tag: 0
               value expressions:
                     expr: key
                     type: string
-        a 
+                    expr: val
+                    type: string
+        c 
           TableScan
-            alias: a
+            alias: c
             Reduce Output Operator
               key expressions:
-                    expr: UDFToDouble(key)
+                    expr: (key + 1)
                     type: double
               sort order: +
               Map-reduce partition columns:
-                    expr: UDFToDouble(key)
+                    expr: (key + 1)
                     type: double
-              tag: 0
+              tag: 1
               value expressions:
                     expr: key
                     type: string
-                    expr: val
-                    type: string
       Reduce Operator Tree:
         Join Operator
           condition map:
@@ -136,38 +136,38 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        c 
+        a 
           TableScan
-            alias: c
+            alias: a
             Reduce Output Operator
               key expressions:
-                    expr: (key + 1)
+                    expr: UDFToDouble(key)
                     type: double
               sort order: +
               Map-reduce partition columns:
-                    expr: (key + 1)
+                    expr: UDFToDouble(key)
                     type: double
-              tag: 0
+              tag: 1
               value expressions:
                     expr: key
                     type: string
-        a 
+                    expr: val
+                    type: string
+        c 
           TableScan
-            alias: a
+            alias: c
             Reduce Output Operator
               key expressions:
-                    expr: UDFToDouble(key)
+                    expr: (key + 1)
                     type: double
               sort order: +
               Map-reduce partition columns:
-                    expr: UDFToDouble(key)
+                    expr: (key + 1)
                     type: double
-              tag: 1
+              tag: 0
               value expressions:
                     expr: key
                     type: string
-                    expr: val
-                    type: string
       Reduce Operator Tree:
         Join Operator
           condition map:
@@ -203,13 +203,13 @@
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
 PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1966956520/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-27-56_522_2427479678690495417/10000
 POSTHOOK: query: FROM T1 a JOIN src c ON c.key+1=a.key
 SELECT a.key, a.val, c.key
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
 POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1966956520/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-27-56_522_2427479678690495417/10000
 1	11	0
 1	11	0
 1	11	0
@@ -219,13 +219,13 @@
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
 PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/2092123428/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-01_499_4645186819885871483/10000
 POSTHOOK: query: FROM T1 a JOIN src c ON c.key+1=a.key
 SELECT /*+ STREAMTABLE(a) */ a.key, a.val, c.key
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
 POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/2092123428/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-01_499_4645186819885871483/10000
 1	11	0
 1	11	0
 1	11	0
@@ -252,9 +252,9 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        b 
+        a 
           TableScan
-            alias: b
+            alias: a
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -263,13 +263,15 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 1
+              tag: 0
               value expressions:
                     expr: key
                     type: string
-        a 
+                    expr: val
+                    type: string
+        b 
           TableScan
-            alias: a
+            alias: b
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -278,12 +280,10 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 0
+              tag: 1
               value expressions:
                     expr: key
                     type: string
-                    expr: val
-                    type: string
       Reduce Operator Tree:
         Join Operator
           condition map:
@@ -389,9 +389,9 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        b 
+        a 
           TableScan
-            alias: b
+            alias: a
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -400,13 +400,15 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 0
+              tag: 1
               value expressions:
                     expr: key
                     type: string
-        a 
+                    expr: val
+                    type: string
+        b 
           TableScan
-            alias: a
+            alias: b
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -415,12 +417,10 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 1
+              tag: 0
               value expressions:
                     expr: key
                     type: string
-                    expr: val
-                    type: string
       Reduce Operator Tree:
         Join Operator
           condition map:
@@ -512,7 +512,7 @@
 PREHOOK: Input: default@t2
 PREHOOK: Input: default@t3
 PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1966669209/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-07_922_2294786151582488765/10000
 POSTHOOK: query: FROM T1 a
   LEFT OUTER JOIN T2 b ON (b.key=a.key)
   RIGHT OUTER JOIN T3 c ON (c.val = a.val)
@@ -521,7 +521,7 @@
 POSTHOOK: Input: default@t2
 POSTHOOK: Input: default@t3
 POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1966669209/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-07_922_2294786151582488765/10000
 2	2	12	12
 NULL	NULL	NULL	14
 NULL	NULL	NULL	16
@@ -534,7 +534,7 @@
 PREHOOK: Input: default@t2
 PREHOOK: Input: default@t3
 PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/2040128444/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-20_301_8638454003102682278/10000
 POSTHOOK: query: FROM T1 a
   LEFT OUTER JOIN T2 b ON (b.key=a.key)
   RIGHT OUTER JOIN T3 c ON (c.val = a.val)
@@ -543,7 +543,7 @@
 POSTHOOK: Input: default@t2
 POSTHOOK: Input: default@t3
 POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/2040128444/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-20_301_8638454003102682278/10000
 2	2	12	12
 NULL	NULL	NULL	14
 NULL	NULL	NULL	16
@@ -571,9 +571,9 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        b 
+        a 
           TableScan
-            alias: b
+            alias: a
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -586,13 +586,13 @@
                     type: string
                     expr: val
                     type: string
-              tag: 1
+              tag: 0
               value expressions:
                     expr: key
                     type: string
-        c 
+        b 
           TableScan
-            alias: c
+            alias: b
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -605,13 +605,13 @@
                     type: string
                     expr: val
                     type: string
-              tag: 2
+              tag: 1
               value expressions:
                     expr: key
                     type: string
-        a 
+        c 
           TableScan
-            alias: a
+            alias: c
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -624,7 +624,7 @@
                     type: string
                     expr: val
                     type: string
-              tag: 0
+              tag: 2
               value expressions:
                     expr: key
                     type: string
@@ -683,9 +683,9 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        b 
+        a 
           TableScan
-            alias: b
+            alias: a
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -698,13 +698,13 @@
                     type: string
                     expr: val
                     type: string
-              tag: 2
+              tag: 0
               value expressions:
                     expr: key
                     type: string
-        c 
+        b 
           TableScan
-            alias: c
+            alias: b
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -717,13 +717,13 @@
                     type: string
                     expr: val
                     type: string
-              tag: 1
+              tag: 2
               value expressions:
                     expr: key
                     type: string
-        a 
+        c 
           TableScan
-            alias: a
+            alias: c
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -736,7 +736,7 @@
                     type: string
                     expr: val
                     type: string
-              tag: 0
+              tag: 1
               value expressions:
                     expr: key
                     type: string
@@ -781,7 +781,7 @@
 PREHOOK: Input: default@t2
 PREHOOK: Input: default@t3
 PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/685649783/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-31_833_1632343405751503469/10000
 POSTHOOK: query: FROM UNIQUEJOIN
   PRESERVE T1 a (a.key, a.val),
   PRESERVE T2 b (b.key, b.val),
@@ -791,7 +791,7 @@
 POSTHOOK: Input: default@t2
 POSTHOOK: Input: default@t3
 POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/685649783/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-31_833_1632343405751503469/10000
 1	NULL	NULL
 2	NULL	2
 NULL	2	NULL
@@ -812,7 +812,7 @@
 PREHOOK: Input: default@t2
 PREHOOK: Input: default@t3
 PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1834964794/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-38_626_3188343531104185080/10000
 POSTHOOK: query: FROM UNIQUEJOIN
   PRESERVE T1 a (a.key, a.val),
   PRESERVE T2 b (b.key, b.val),
@@ -822,7 +822,7 @@
 POSTHOOK: Input: default@t2
 POSTHOOK: Input: default@t3
 POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_0/build/ql/tmp/1834964794/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-28-38_626_3188343531104185080/10000
 1	NULL	NULL
 2	NULL	2
 NULL	2	NULL

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder2.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder2.q.out Sun Feb 14 02:31:40 2010
@@ -77,9 +77,9 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        d 
+        a 
           TableScan
-            alias: d
+            alias: a
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -88,7 +88,7 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 0
+              tag: 3
               value expressions:
                     expr: key
                     type: string
@@ -128,9 +128,9 @@
                     type: string
                     expr: val
                     type: string
-        a 
+        d 
           TableScan
-            alias: a
+            alias: d
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -139,7 +139,7 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 3
+              tag: 0
               value expressions:
                     expr: key
                     type: string
@@ -198,7 +198,7 @@
 PREHOOK: Input: default@t2
 PREHOOK: Input: default@t3
 PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1499291437/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-28-02_101_1898487455882506837/10000
 POSTHOOK: query: SELECT /*+ STREAMTABLE(a) */ *
 FROM T1 a JOIN T2 b ON a.key = b.key
           JOIN T3 c ON b.key = c.key
@@ -208,7 +208,7 @@
 POSTHOOK: Input: default@t2
 POSTHOOK: Input: default@t3
 POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1499291437/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-28-02_101_1898487455882506837/10000
 2	12	2	22	2	12	2	12
 PREHOOK: query: EXPLAIN
 SELECT /*+ STREAMTABLE(a) */ *
@@ -235,9 +235,9 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        b 
+        a 
           TableScan
-            alias: b
+            alias: a
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -246,15 +246,15 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 0
+              tag: 1
               value expressions:
                     expr: key
                     type: string
                     expr: val
                     type: string
-        a 
+        b 
           TableScan
-            alias: a
+            alias: b
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -263,7 +263,7 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 1
+              tag: 0
               value expressions:
                     expr: key
                     type: string
@@ -343,23 +343,6 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        d 
-          TableScan
-            alias: d
-            Reduce Output Operator
-              key expressions:
-                    expr: (key + 1)
-                    type: double
-              sort order: +
-              Map-reduce partition columns:
-                    expr: (key + 1)
-                    type: double
-              tag: 0
-              value expressions:
-                    expr: key
-                    type: string
-                    expr: val
-                    type: string
         $INTNAME 
             Reduce Output Operator
               key expressions:
@@ -383,6 +366,23 @@
                     type: string
                     expr: _col3
                     type: string
+        d 
+          TableScan
+            alias: d
+            Reduce Output Operator
+              key expressions:
+                    expr: (key + 1)
+                    type: double
+              sort order: +
+              Map-reduce partition columns:
+                    expr: (key + 1)
+                    type: double
+              tag: 0
+              value expressions:
+                    expr: key
+                    type: string
+                    expr: val
+                    type: string
       Reduce Operator Tree:
         Join Operator
           condition map:
@@ -431,7 +431,7 @@
 PREHOOK: Input: default@t2
 PREHOOK: Input: default@t3
 PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/347443381/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-28-09_022_5817489663400235377/10000
 POSTHOOK: query: SELECT /*+ STREAMTABLE(a) */ *
 FROM T1 a JOIN T2 b ON a.key = b.key
           JOIN T3 c ON a.val = c.val
@@ -441,7 +441,7 @@
 POSTHOOK: Input: default@t2
 POSTHOOK: Input: default@t3
 POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/347443381/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-28-09_022_5817489663400235377/10000
 2	22	2	12	2	12	2	12
 PREHOOK: query: DROP TABLE T1
 PREHOOK: type: DROPTABLE

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder3.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder3.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join_reorder3.q.out Sun Feb 14 02:31:40 2010
@@ -77,9 +77,9 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        d 
+        a 
           TableScan
-            alias: d
+            alias: a
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -88,7 +88,7 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 0
+              tag: 3
               value expressions:
                     expr: key
                     type: string
@@ -128,9 +128,9 @@
                     type: string
                     expr: val
                     type: string
-        a 
+        d 
           TableScan
-            alias: a
+            alias: d
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -139,7 +139,7 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 3
+              tag: 0
               value expressions:
                     expr: key
                     type: string
@@ -198,7 +198,7 @@
 PREHOOK: Input: default@t2
 PREHOOK: Input: default@t3
 PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/922482225/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-34_907_1935969450376975821/10000
 POSTHOOK: query: SELECT /*+ STREAMTABLE(a,c) */ *
 FROM T1 a JOIN T2 b ON a.key = b.key
           JOIN T3 c ON b.key = c.key
@@ -208,7 +208,7 @@
 POSTHOOK: Input: default@t2
 POSTHOOK: Input: default@t3
 POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/922482225/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-34_907_1935969450376975821/10000
 2	12	2	22	2	12	2	12
 PREHOOK: query: EXPLAIN
 SELECT /*+ STREAMTABLE(a,c) */ *
@@ -235,9 +235,9 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        b 
+        a 
           TableScan
-            alias: b
+            alias: a
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -246,15 +246,15 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 0
+              tag: 1
               value expressions:
                     expr: key
                     type: string
                     expr: val
                     type: string
-        a 
+        b 
           TableScan
-            alias: a
+            alias: b
             Reduce Output Operator
               key expressions:
                     expr: key
@@ -263,7 +263,7 @@
               Map-reduce partition columns:
                     expr: key
                     type: string
-              tag: 1
+              tag: 0
               value expressions:
                     expr: key
                     type: string
@@ -343,23 +343,6 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        d 
-          TableScan
-            alias: d
-            Reduce Output Operator
-              key expressions:
-                    expr: (key + 1)
-                    type: double
-              sort order: +
-              Map-reduce partition columns:
-                    expr: (key + 1)
-                    type: double
-              tag: 0
-              value expressions:
-                    expr: key
-                    type: string
-                    expr: val
-                    type: string
         $INTNAME 
             Reduce Output Operator
               key expressions:
@@ -383,6 +366,23 @@
                     type: string
                     expr: _col3
                     type: string
+        d 
+          TableScan
+            alias: d
+            Reduce Output Operator
+              key expressions:
+                    expr: (key + 1)
+                    type: double
+              sort order: +
+              Map-reduce partition columns:
+                    expr: (key + 1)
+                    type: double
+              tag: 0
+              value expressions:
+                    expr: key
+                    type: string
+                    expr: val
+                    type: string
       Reduce Operator Tree:
         Join Operator
           condition map:
@@ -431,7 +431,7 @@
 PREHOOK: Input: default@t2
 PREHOOK: Input: default@t3
 PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/831264775/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-41_586_5996486802122601404/10000
 POSTHOOK: query: SELECT /*+ STREAMTABLE(a,c) */ *
 FROM T1 a JOIN T2 b ON a.key = b.key
           JOIN T3 c ON a.val = c.val
@@ -441,7 +441,7 @@
 POSTHOOK: Input: default@t2
 POSTHOOK: Input: default@t3
 POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_2/build/ql/tmp/831264775/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-41_586_5996486802122601404/10000
 2	22	2	12	2	12	2	12
 PREHOOK: query: DROP TABLE T1
 PREHOOK: type: DROPTABLE

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join_thrift.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join_thrift.q.out?rev=909965&r1=909964&r2=909965&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join_thrift.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join_thrift.q.out Sun Feb 14 02:31:40 2010
@@ -31,9 +31,9 @@
   Stage: Stage-1
     Map Reduce
       Alias -> Map Operator Tree:
-        s2 
+        s1 
           TableScan
-            alias: s2
+            alias: s1
             Reduce Output Operator
               key expressions:
                     expr: aint
@@ -42,13 +42,13 @@
               Map-reduce partition columns:
                     expr: aint
                     type: int
-              tag: 1
+              tag: 0
               value expressions:
-                    expr: lintstring
-                    type: array<struct<myint:int,mystring:string,underscore_int:int>>
-        s1 
+                    expr: aint
+                    type: int
+        s2 
           TableScan
-            alias: s1
+            alias: s2
             Reduce Output Operator
               key expressions:
                     expr: aint
@@ -57,10 +57,10 @@
               Map-reduce partition columns:
                     expr: aint
                     type: int
-              tag: 0
+              tag: 1
               value expressions:
-                    expr: aint
-                    type: int
+                    expr: lintstring
+                    type: array<struct<myint:int,mystring:string,underscore_int:int>>
       Reduce Operator Tree:
         Join Operator
           condition map:
@@ -95,14 +95,14 @@
 ON s1.aint = s2.aint
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src_thrift
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1362622837/10000
+PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-28-28_139_5468299142078849499/10000
 POSTHOOK: query: SELECT s1.aint, s2.lintstring
 FROM src_thrift s1
 JOIN src_thrift s2
 ON s1.aint = s2.aint
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src_thrift
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk/.ptest_1/build/ql/tmp/1362622837/10000
+POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-28-28_139_5468299142078849499/10000
 -1952710710	[{"myint":25,"mystring":"125","underscore_int":5}]
 -1461153973	[{"myint":49,"mystring":"343","underscore_int":7}]
 -751827638	[{"myint":4,"mystring":"8","underscore_int":2}]



Mime
View raw message