hadoop-hive-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From j..@apache.org
Subject svn commit: r980659 [21/34] - in /hadoop/hive/trunk: ./ common/src/java/org/apache/hadoop/hive/common/ contrib/src/test/results/clientpositive/ metastore/if/ metastore/src/gen-cpp/ metastore/src/gen-javabean/org/apache/hadoop/hive/metastore/api/ metast...
Date Fri, 30 Jul 2010 06:40:11 GMT
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join37.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join37.q.out?rev=980659&r1=980658&r2=980659&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join37.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join37.q.out Fri Jul 30 06:40:04 2010
@@ -40,7 +40,7 @@ STAGE PLANS:
               keys:
                 0 [Column[key]]
                 1 [Column[key]]
-              outputColumnNames: _col0, _col1, _col3
+              outputColumnNames: _col0, _col1, _col5
               Position of Big Table: 1
               Select Operator
                 expressions:
@@ -48,16 +48,16 @@ STAGE PLANS:
                       type: string
                       expr: _col1
                       type: string
-                      expr: _col3
+                      expr: _col5
                       type: string
-                outputColumnNames: _col0, _col1, _col3
+                outputColumnNames: _col0, _col1, _col5
                 Select Operator
                   expressions:
                         expr: _col0
                         type: string
                         expr: _col1
                         type: string
-                        expr: _col3
+                        expr: _col5
                         type: string
                   outputColumnNames: _col0, _col1, _col2
                   Select Operator
@@ -97,7 +97,7 @@ STAGE PLANS:
                   keys:
                     0 [Column[key]]
                     1 [Column[key]]
-                  outputColumnNames: _col0, _col1, _col3
+                  outputColumnNames: _col0, _col1, _col5
                   Position of Big Table: 1
                   Select Operator
                     expressions:
@@ -105,16 +105,16 @@ STAGE PLANS:
                           type: string
                           expr: _col1
                           type: string
-                          expr: _col3
+                          expr: _col5
                           type: string
-                    outputColumnNames: _col0, _col1, _col3
+                    outputColumnNames: _col0, _col1, _col5
                     Select Operator
                       expressions:
                             expr: _col0
                             type: string
                             expr: _col1
                             type: string
-                            expr: _col3
+                            expr: _col5
                             type: string
                       outputColumnNames: _col0, _col1, _col2
                       Select Operator
@@ -142,7 +142,7 @@ STAGE PLANS:
     Move Operator
       files:
           hdfs directory: true
-          destination: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-34-21_008_268143953010712985/10000
+          destination: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_12-59-23_084_5025868117733478455/-ext-10000
 
   Stage: Stage-0
     Move Operator
@@ -157,7 +157,7 @@ STAGE PLANS:
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-34-21_008_268143953010712985/10002 
+        pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_12-59-23_084_5025868117733478455/-ext-10002 
             Reduce Output Operator
               sort order: 
               Map-reduce partition columns:
@@ -203,11 +203,11 @@ POSTHOOK: Lineage: dest_j1.value SIMPLE 
 PREHOOK: query: select * from dest_j1 x order by x.key
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest_j1
-PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-34-23_746_1310159606592895751/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_12-59-28_143_4579787448027407977/-mr-10000
 POSTHOOK: query: select * from dest_j1 x order by x.key
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest_j1
-POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-34-23_746_1310159606592895751/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_12-59-28_143_4579787448027407977/-mr-10000
 POSTHOOK: Lineage: dest_j1.key EXPRESSION [(src1)x.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: dest_j1.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ]
 POSTHOOK: Lineage: dest_j1.value SIMPLE [(src1)x.FieldSchema(name:value, type:string, comment:default), ]

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join38.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join38.q.out?rev=980659&r1=980658&r2=980659&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join38.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join38.q.out Fri Jul 30 06:40:04 2010
@@ -26,11 +26,11 @@ POSTHOOK: Lineage: tmp.col9 EXPRESSION [
 PREHOOK: query: select * from tmp
 PREHOOK: type: QUERY
 PREHOOK: Input: default@tmp
-PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-34-29_174_1856986099030897576/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_12-59-37_683_5319482143791387584/-mr-10000
 POSTHOOK: query: select * from tmp
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@tmp
-POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-34-29_174_1856986099030897576/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_12-59-37_683_5319482143791387584/-mr-10000
 POSTHOOK: Lineage: tmp.col0 SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: tmp.col1 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: tmp.col10 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -98,7 +98,7 @@ STAGE PLANS:
                 keys:
                   0 [Column[key]]
                   1 [Column[col11]]
-                outputColumnNames: _col1, _col7, _col13
+                outputColumnNames: _col1, _col9, _col15
                 Position of Big Table: 1
                 File Output Operator
                   compressed: false
@@ -126,7 +126,7 @@ STAGE PLANS:
                   keys:
                     0 [Column[key]]
                     1 [Column[col11]]
-                  outputColumnNames: _col1, _col7, _col13
+                  outputColumnNames: _col1, _col9, _col15
                   Position of Big Table: 1
                   File Output Operator
                     compressed: false
@@ -138,27 +138,27 @@ STAGE PLANS:
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/tmp/jssarma/hive_2010-07-21_11-34-29_228_6035667266137509751/10002 
+        file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_12-59-37_907_1369838299280728635/-mr-10002 
           Select Operator
             expressions:
                   expr: _col1
                   type: string
-                  expr: _col7
+                  expr: _col9
                   type: string
-                  expr: _col13
+                  expr: _col15
                   type: string
-            outputColumnNames: _col1, _col7, _col13
+            outputColumnNames: _col1, _col9, _col15
             Filter Operator
               predicate:
-                  expr: (_col13 = 111)
+                  expr: (_col15 = 111)
                   type: boolean
               Select Operator
                 expressions:
                       expr: _col1
                       type: string
-                      expr: _col7
+                      expr: _col9
                       type: string
-                outputColumnNames: _col1, _col7
+                outputColumnNames: _col1, _col9
                 Group By Operator
                   aggregations:
                         expr: count(1)
@@ -166,7 +166,7 @@ STAGE PLANS:
                   keys:
                         expr: _col1
                         type: string
-                        expr: _col7
+                        expr: _col9
                         type: string
                   mode: hash
                   outputColumnNames: _col0, _col1, _col2
@@ -226,7 +226,7 @@ group by a.value, b.col5
 PREHOOK: type: QUERY
 PREHOOK: Input: default@tmp
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-34-29_328_1884952246433236749/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_12-59-38_019_4750710911180862739/-mr-10000
 POSTHOOK: query: FROM src a JOIN tmp b ON (a.key = b.col11)
 SELECT /*+ MAPJOIN(a) */ a.value, b.col5, count(1) as count
 where b.col11 = 111
@@ -234,7 +234,7 @@ group by a.value, b.col5
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@tmp
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-34-29_328_1884952246433236749/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_12-59-38_019_4750710911180862739/-mr-10000
 POSTHOOK: Lineage: tmp.col0 SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: tmp.col1 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: tmp.col10 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join40.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join40.q.out?rev=980659&r1=980658&r2=980659&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join40.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join40.q.out Fri Jul 30 06:40:04 2010
@@ -72,16 +72,16 @@ STAGE PLANS:
             0 {VALUE._col0} {VALUE._col1}
             1 {VALUE._col0} {VALUE._col1}
           handleSkewJoin: false
-          outputColumnNames: _col0, _col1, _col2, _col3
+          outputColumnNames: _col0, _col1, _col4, _col5
           Select Operator
             expressions:
                   expr: _col0
                   type: string
                   expr: _col1
                   type: string
-                  expr: _col2
+                  expr: _col4
                   type: string
-                  expr: _col3
+                  expr: _col5
                   type: string
             outputColumnNames: _col0, _col1, _col2, _col3
             File Output Operator
@@ -100,12 +100,12 @@ PREHOOK: query: SELECT x.key, x.value, y
 FROM src x left outer JOIN (select * from src where key <= 100) y ON (x.key = y.key)
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-26-20_375_8739608897771687968/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-00-06_210_967311768247652763/-mr-10000
 POSTHOOK: query: SELECT x.key, x.value, y.key, y.value
 FROM src x left outer JOIN (select * from src where key <= 100) y ON (x.key = y.key)
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-26-20_375_8739608897771687968/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-00-06_210_967311768247652763/-mr-10000
 0	val_0	0	val_0
 0	val_0	0	val_0
 0	val_0	0	val_0
@@ -727,12 +727,12 @@ STAGE PLANS:
             0 {VALUE._col0}
             1 {VALUE._col1}
           handleSkewJoin: false
-          outputColumnNames: _col0, _col3
+          outputColumnNames: _col0, _col5
           Select Operator
             expressions:
                   expr: _col0
                   type: string
-                  expr: _col3
+                  expr: _col5
                   type: string
             outputColumnNames: _col0, _col1
             File Output Operator
@@ -751,12 +751,12 @@ PREHOOK: query: select src1.key, src2.va
 FROM src src1 JOIN src src2 ON (src1.key = src2.key)
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-26-28_099_8963486530334152093/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-00-17_142_417700842326185495/-mr-10000
 POSTHOOK: query: select src1.key, src2.value 
 FROM src src1 JOIN src src2 ON (src1.key = src2.key)
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-26-28_099_8963486530334152093/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-00-17_142_417700842326185495/-mr-10000
 0	val_0
 0	val_0
 0	val_0
@@ -1882,21 +1882,21 @@ STAGE PLANS:
             1 {VALUE._col0} {VALUE._col1}
             2 {VALUE._col0} {VALUE._col1}
           handleSkewJoin: false
-          outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
+          outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9
           Select Operator
             expressions:
                   expr: _col0
                   type: string
                   expr: _col1
                   type: string
-                  expr: _col2
-                  type: string
-                  expr: _col3
-                  type: string
                   expr: _col4
                   type: string
                   expr: _col5
                   type: string
+                  expr: _col8
+                  type: string
+                  expr: _col9
+                  type: string
             outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
             File Output Operator
               compressed: false
@@ -1908,7 +1908,7 @@ STAGE PLANS:
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-26-35_615_2133554865043548038/10002 
+        file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-00-31_857_1059875095406707508/-mr-10002 
             Reduce Output Operator
               key expressions:
                     expr: _col0
@@ -1956,12 +1956,12 @@ PREHOOK: query: SELECT * FROM src src1 J
 SORT BY src1.key, src1.value, src2.key, src2.value, src3.key, src3.value
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-26-35_711_5635391210022763694/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-00-31_970_2552113638239570524/-mr-10000
 POSTHOOK: query: SELECT * FROM src src1 JOIN src src2 ON (src1.key = src2.key AND src1.key < 10) RIGHT OUTER JOIN src src3 ON (src1.key = src3.key AND src3.key < 20)
 SORT BY src1.key, src1.value, src2.key, src2.value, src3.key, src3.value
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-26-35_711_5635391210022763694/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-00-31_970_2552113638239570524/-mr-10000
 NULL	NULL	NULL	NULL	10	val_10
 NULL	NULL	NULL	NULL	11	val_11
 NULL	NULL	NULL	NULL	12	val_12
@@ -2135,21 +2135,21 @@ STAGE PLANS:
             1 {VALUE._col0} {VALUE._col1}
             2 {VALUE._col0} {VALUE._col1}
           handleSkewJoin: false
-          outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
+          outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9
           Select Operator
             expressions:
                   expr: _col0
                   type: string
                   expr: _col1
                   type: string
-                  expr: _col2
-                  type: string
-                  expr: _col3
-                  type: string
                   expr: _col4
                   type: string
                   expr: _col5
                   type: string
+                  expr: _col8
+                  type: string
+                  expr: _col9
+                  type: string
             outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
             File Output Operator
               compressed: false
@@ -2161,7 +2161,7 @@ STAGE PLANS:
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-26-47_526_7396170302427973294/10002 
+        file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-00-48_236_188923787391871879/-mr-10002 
             Reduce Output Operator
               key expressions:
                     expr: _col0
@@ -2209,12 +2209,12 @@ PREHOOK: query: SELECT * FROM src src1 J
 SORT BY src1.key, src1.value, src2.key, src2.value, src3.key, src3.value
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-26-47_626_1855366263501057105/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-00-48_363_6899226935845466819/-mr-10000
 POSTHOOK: query: SELECT * FROM src src1 JOIN src src2 ON (src1.key = src2.key AND src1.key < 10 AND src2.key < 15) RIGHT OUTER JOIN src src3 ON (src1.key = src3.key AND src3.key < 20)
 SORT BY src1.key, src1.value, src2.key, src2.value, src3.key, src3.value
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-26-47_626_1855366263501057105/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-00-48_363_6899226935845466819/-mr-10000
 NULL	NULL	NULL	NULL	10	val_10
 NULL	NULL	NULL	NULL	11	val_11
 NULL	NULL	NULL	NULL	12	val_12
@@ -2315,7 +2315,7 @@ STAGE PLANS:
               keys:
                 0 [Column[key]]
                 1 [Column[_col0]]
-              outputColumnNames: _col0, _col1, _col2, _col3
+              outputColumnNames: _col0, _col1, _col4, _col5
               Position of Big Table: 0
               Select Operator
                 expressions:
@@ -2323,20 +2323,20 @@ STAGE PLANS:
                       type: string
                       expr: _col1
                       type: string
-                      expr: _col2
+                      expr: _col4
                       type: string
-                      expr: _col3
+                      expr: _col5
                       type: string
-                outputColumnNames: _col0, _col1, _col2, _col3
+                outputColumnNames: _col0, _col1, _col4, _col5
                 Select Operator
                   expressions:
                         expr: _col0
                         type: string
                         expr: _col1
                         type: string
-                        expr: _col2
+                        expr: _col4
                         type: string
-                        expr: _col3
+                        expr: _col5
                         type: string
                   outputColumnNames: _col0, _col1, _col2, _col3
                   File Output Operator
@@ -2380,7 +2380,7 @@ STAGE PLANS:
                         keys:
                           0 [Column[key]]
                           1 [Column[_col0]]
-                        outputColumnNames: _col0, _col1, _col2, _col3
+                        outputColumnNames: _col0, _col1, _col4, _col5
                         Position of Big Table: 0
                         Select Operator
                           expressions:
@@ -2388,20 +2388,20 @@ STAGE PLANS:
                                 type: string
                                 expr: _col1
                                 type: string
-                                expr: _col2
+                                expr: _col4
                                 type: string
-                                expr: _col3
+                                expr: _col5
                                 type: string
-                          outputColumnNames: _col0, _col1, _col2, _col3
+                          outputColumnNames: _col0, _col1, _col4, _col5
                           Select Operator
                             expressions:
                                   expr: _col0
                                   type: string
                                   expr: _col1
                                   type: string
-                                  expr: _col2
+                                  expr: _col4
                                   type: string
-                                  expr: _col3
+                                  expr: _col5
                                   type: string
                             outputColumnNames: _col0, _col1, _col2, _col3
                             File Output Operator
@@ -2420,12 +2420,12 @@ PREHOOK: query: SELECT /*+ MAPJOIN(y) */
 FROM src x left outer JOIN (select * from src where key <= 100) y ON (x.key = y.key)
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-26-58_663_5257751854137888714/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-01-00_497_7810479910052206762/-mr-10000
 POSTHOOK: query: SELECT /*+ MAPJOIN(y) */ x.key, x.value, y.key, y.value
 FROM src x left outer JOIN (select * from src where key <= 100) y ON (x.key = y.key)
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-26-58_663_5257751854137888714/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-01-00_497_7810479910052206762/-mr-10000
 238	val_238	NULL	NULL
 86	val_86	86	val_86
 311	val_311	NULL	NULL
@@ -3059,7 +3059,7 @@ STAGE PLANS:
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-27-07_552_3382271159548134599/10002 
+        file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-01-06_156_3096923421519776221/-mr-10002 
             Reduce Output Operator
               sort order: 
               tag: -1
@@ -3093,9 +3093,9 @@ STAGE PLANS:
 PREHOOK: query: SELECT COUNT(1) FROM SRC A JOIN SRC B ON (A.KEY=B.KEY)
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-27-07_618_4569452327751615405/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-01-06_237_3766500628681283400/-mr-10000
 POSTHOOK: query: SELECT COUNT(1) FROM SRC A JOIN SRC B ON (A.KEY=B.KEY)
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_1/build/ql/scratchdir/hive_2010-02-12_22-27-07_618_4569452327751615405/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-01-06_237_3766500628681283400/-mr-10000
 1028

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join9.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join9.q.out?rev=980659&r1=980658&r2=980659&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join9.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join9.q.out Fri Jul 30 06:40:04 2010
@@ -63,10 +63,10 @@ STAGE PLANS:
                     type: string
       Needs Tagging: true
       Path -> Alias:
-        file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/test/data/warehouse/src [src2]
-        file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [src1]
+        pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/src [src2]
+        pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [src1]
       Path -> Partition:
-        file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/test/data/warehouse/src 
+        pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/src 
           Partition
             base file name: src
             input format: org.apache.hadoop.mapred.TextInputFormat
@@ -77,12 +77,12 @@ STAGE PLANS:
               columns.types string:string
               file.inputformat org.apache.hadoop.mapred.TextInputFormat
               file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/test/data/warehouse/src
+              location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/src
               name src
               serialization.ddl struct src { string key, string value}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              transient_lastDdlTime 1270516625
+              transient_lastDdlTime 1280430369
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -93,16 +93,16 @@ STAGE PLANS:
                 columns.types string:string
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/test/data/warehouse/src
+                location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/src
                 name src
                 serialization.ddl struct src { string key, string value}
                 serialization.format 1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                transient_lastDdlTime 1270516625
+                transient_lastDdlTime 1280430369
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: src
             name: src
-        file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 
+        pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 
           Partition
             base file name: hr=12
             input format: org.apache.hadoop.mapred.TextInputFormat
@@ -116,13 +116,13 @@ STAGE PLANS:
               columns.types string:string
               file.inputformat org.apache.hadoop.mapred.TextInputFormat
               file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/test/data/warehouse/srcpart
+              location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/srcpart
               name srcpart
               partition_columns ds/hr
               serialization.ddl struct srcpart { string key, string value}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              transient_lastDdlTime 1270516624
+              transient_lastDdlTime 1280430361
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -133,13 +133,13 @@ STAGE PLANS:
                 columns.types string:string
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/test/data/warehouse/srcpart
+                location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/srcpart
                 name srcpart
                 partition_columns ds/hr
                 serialization.ddl struct srcpart { string key, string value}
                 serialization.format 1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                transient_lastDdlTime 1270516624
+                transient_lastDdlTime 1280430361
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
             name: srcpart
@@ -151,7 +151,7 @@ STAGE PLANS:
             0 {VALUE._col0} {VALUE._col2} {VALUE._col3}
             1 {VALUE._col1}
           handleSkewJoin: false
-          outputColumnNames: _col0, _col2, _col3, _col5
+          outputColumnNames: _col0, _col2, _col3, _col7
           Filter Operator
             isSamplingPred: false
             predicate:
@@ -161,7 +161,7 @@ STAGE PLANS:
               expressions:
                     expr: _col0
                     type: string
-                    expr: _col5
+                    expr: _col7
                     type: string
               outputColumnNames: _col0, _col1
               Select Operator
@@ -174,7 +174,7 @@ STAGE PLANS:
                 File Output Operator
                   compressed: false
                   GlobalTableId: 1
-                  directory: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-17-06_135_7326848729377231299/10000
+                  directory: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-01-48_646_887463072369912010/-ext-10000
                   NumFilesPerFileSink: 1
                   table:
                       input format: org.apache.hadoop.mapred.TextInputFormat
@@ -185,12 +185,12 @@ STAGE PLANS:
                         columns.types int:string
                         file.inputformat org.apache.hadoop.mapred.TextInputFormat
                         file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                        location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/test/data/warehouse/dest1
+                        location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/dest1
                         name dest1
                         serialization.ddl struct dest1 { i32 key, string value}
                         serialization.format 1
                         serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                        transient_lastDdlTime 1270516626
+                        transient_lastDdlTime 1280433708
                       serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                       name: dest1
                   TotalFiles: 1
@@ -200,7 +200,7 @@ STAGE PLANS:
     Move Operator
       tables:
           replace: true
-          source: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-17-06_135_7326848729377231299/10000
+          source: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-01-48_646_887463072369912010/-ext-10000
           table:
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -210,15 +210,15 @@ STAGE PLANS:
                 columns.types int:string
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/test/data/warehouse/dest1
+                location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/dest1
                 name dest1
                 serialization.ddl struct dest1 { i32 key, string value}
                 serialization.format 1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                transient_lastDdlTime 1270516626
+                transient_lastDdlTime 1280433708
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: dest1
-          tmp directory: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-17-06_135_7326848729377231299/10001
+          tmp directory: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-01-48_646_887463072369912010/-ext-10001
 
 
 PREHOOK: query: FROM srcpart src1 JOIN src src2 ON (src1.key = src2.key)
@@ -238,11 +238,11 @@ POSTHOOK: Lineage: dest1.value SIMPLE [(
 PREHOOK: query: SELECT dest1.* FROM dest1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-17-13_273_2600052382026656668/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-01-54_035_2958246273325433849/-mr-10000
 POSTHOOK: query: SELECT dest1.* FROM dest1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-17-13_273_2600052382026656668/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-01-54_035_2958246273325433849/-mr-10000
 POSTHOOK: Lineage: dest1.key EXPRESSION [(srcpart)src1.FieldSchema(name:ds, type:string, comment:null), ]
 POSTHOOK: Lineage: dest1.value SIMPLE [(src)src2.FieldSchema(name:value, type:string, comment:default), ]
 0	val_0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join_hive_626.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join_hive_626.q.out?rev=980659&r1=980658&r2=980659&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join_hive_626.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join_hive_626.q.out Fri Jul 30 06:40:04 2010
@@ -98,7 +98,7 @@ STAGE PLANS:
             0 {VALUE._col1}
             1 {VALUE._col0} {VALUE._col4}
           handleSkewJoin: false
-          outputColumnNames: _col1, _col6, _col10
+          outputColumnNames: _col1, _col8, _col12
           File Output Operator
             compressed: false
             GlobalTableId: 0
@@ -112,17 +112,17 @@ STAGE PLANS:
         $INTNAME 
             Reduce Output Operator
               key expressions:
-                    expr: _col6
+                    expr: _col8
                     type: int
               sort order: +
               Map-reduce partition columns:
-                    expr: _col6
+                    expr: _col8
                     type: int
               tag: 0
               value expressions:
                     expr: _col1
                     type: string
-                    expr: _col10
+                    expr: _col12
                     type: string
         hive_count 
           TableScan
@@ -144,17 +144,17 @@ STAGE PLANS:
           condition map:
                Inner Join 0 to 1
           condition expressions:
-            0 {VALUE._col1} {VALUE._col10}
+            0 {VALUE._col1} {VALUE._col12}
             1 {VALUE._col1}
           handleSkewJoin: false
-          outputColumnNames: _col1, _col10, _col16
+          outputColumnNames: _col1, _col12, _col20
           Select Operator
             expressions:
                   expr: _col1
                   type: string
-                  expr: _col10
+                  expr: _col12
                   type: string
-                  expr: _col16
+                  expr: _col20
                   type: int
             outputColumnNames: _col0, _col1, _col2
             File Output Operator
@@ -175,12 +175,12 @@ PREHOOK: type: QUERY
 PREHOOK: Input: default@hive_foo
 PREHOOK: Input: default@hive_count
 PREHOOK: Input: default@hive_bar
-PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-35-37_392_1859121185876168833/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-01-56_605_4228947637564753632/-mr-10000
 POSTHOOK: query: select hive_foo.foo_name, hive_bar.bar_name, n from hive_foo join hive_bar on hive_foo.foo_id =
 hive_bar.foo_id join hive_count on hive_count.bar_id = hive_bar.bar_id
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@hive_foo
 POSTHOOK: Input: default@hive_count
 POSTHOOK: Input: default@hive_bar
-POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-35-37_392_1859121185876168833/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-01-56_605_4228947637564753632/-mr-10000
 foo1	bar10	2

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join_map_ppr.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join_map_ppr.q.out?rev=980659&r1=980658&r2=980659&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join_map_ppr.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join_map_ppr.q.out Fri Jul 30 06:40:04 2010
@@ -52,39 +52,39 @@ STAGE PLANS:
                   0 [Column[key]]
                   1 [Column[key]]
                   2 [Column[key]]
-                outputColumnNames: _col0, _col3, _col5, _col6, _col7
+                outputColumnNames: _col0, _col5, _col9, _col10, _col11
                 Position of Big Table: 2
                 Select Operator
                   expressions:
                         expr: _col0
                         type: string
-                        expr: _col3
-                        type: string
                         expr: _col5
                         type: string
-                        expr: _col6
+                        expr: _col9
+                        type: string
+                        expr: _col10
                         type: string
-                        expr: _col7
+                        expr: _col11
                         type: string
-                  outputColumnNames: _col0, _col3, _col5, _col6, _col7
+                  outputColumnNames: _col0, _col5, _col9, _col10, _col11
                   Filter Operator
                     isSamplingPred: false
                     predicate:
-                        expr: ((_col6 = '2008-04-08') and (_col7 = 11))
+                        expr: ((_col10 = '2008-04-08') and (_col11 = 11))
                         type: boolean
                     Select Operator
                       expressions:
                             expr: _col0
                             type: string
-                            expr: _col5
+                            expr: _col9
                             type: string
-                            expr: _col3
+                            expr: _col5
                             type: string
                       outputColumnNames: _col0, _col1, _col2
                       File Output Operator
                         compressed: false
                         GlobalTableId: 1
-                        directory: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-32_178_1568785792517767702/-ext-10002
+                        directory: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-06_654_4872924397003947925/-ext-10002
                         NumFilesPerFileSink: 1
                         table:
                             input format: org.apache.hadoop.mapred.TextInputFormat
@@ -95,12 +95,12 @@ STAGE PLANS:
                               columns.types string:string:string
                               file.inputformat org.apache.hadoop.mapred.TextInputFormat
                               file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                              location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/dest_j1
+                              location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/dest_j1
                               name dest_j1
                               serialization.ddl struct dest_j1 { string key, string value, string val2}
                               serialization.format 1
                               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                              transient_lastDdlTime 1280084852
+                              transient_lastDdlTime 1280433726
                             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                             name: dest_j1
                         TotalFiles: 1
@@ -131,39 +131,39 @@ STAGE PLANS:
                     0 [Column[key]]
                     1 [Column[key]]
                     2 [Column[key]]
-                  outputColumnNames: _col0, _col3, _col5, _col6, _col7
+                  outputColumnNames: _col0, _col5, _col9, _col10, _col11
                   Position of Big Table: 2
                   Select Operator
                     expressions:
                           expr: _col0
                           type: string
-                          expr: _col3
-                          type: string
                           expr: _col5
                           type: string
-                          expr: _col6
+                          expr: _col9
+                          type: string
+                          expr: _col10
                           type: string
-                          expr: _col7
+                          expr: _col11
                           type: string
-                    outputColumnNames: _col0, _col3, _col5, _col6, _col7
+                    outputColumnNames: _col0, _col5, _col9, _col10, _col11
                     Filter Operator
                       isSamplingPred: false
                       predicate:
-                          expr: ((_col6 = '2008-04-08') and (_col7 = 11))
+                          expr: ((_col10 = '2008-04-08') and (_col11 = 11))
                           type: boolean
                       Select Operator
                         expressions:
                               expr: _col0
                               type: string
-                              expr: _col5
+                              expr: _col9
                               type: string
-                              expr: _col3
+                              expr: _col5
                               type: string
                         outputColumnNames: _col0, _col1, _col2
                         File Output Operator
                           compressed: false
                           GlobalTableId: 1
-                          directory: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-32_178_1568785792517767702/-ext-10002
+                          directory: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-06_654_4872924397003947925/-ext-10002
                           NumFilesPerFileSink: 1
                           table:
                               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -174,12 +174,12 @@ STAGE PLANS:
                                 columns.types string:string:string
                                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                                location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/dest_j1
+                                location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/dest_j1
                                 name dest_j1
                                 serialization.ddl struct dest_j1 { string key, string value, string val2}
                                 serialization.format 1
                                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                                transient_lastDdlTime 1280084852
+                                transient_lastDdlTime 1280433726
                               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                               name: dest_j1
                           TotalFiles: 1
@@ -200,39 +200,39 @@ STAGE PLANS:
                     0 [Column[key]]
                     1 [Column[key]]
                     2 [Column[key]]
-                  outputColumnNames: _col0, _col3, _col5, _col6, _col7
+                  outputColumnNames: _col0, _col5, _col9, _col10, _col11
                   Position of Big Table: 2
                   Select Operator
                     expressions:
                           expr: _col0
                           type: string
-                          expr: _col3
-                          type: string
                           expr: _col5
                           type: string
-                          expr: _col6
+                          expr: _col9
                           type: string
-                          expr: _col7
+                          expr: _col10
                           type: string
-                    outputColumnNames: _col0, _col3, _col5, _col6, _col7
+                          expr: _col11
+                          type: string
+                    outputColumnNames: _col0, _col5, _col9, _col10, _col11
                     Filter Operator
                       isSamplingPred: false
                       predicate:
-                          expr: ((_col6 = '2008-04-08') and (_col7 = 11))
+                          expr: ((_col10 = '2008-04-08') and (_col11 = 11))
                           type: boolean
                       Select Operator
                         expressions:
                               expr: _col0
                               type: string
-                              expr: _col5
+                              expr: _col9
                               type: string
-                              expr: _col3
+                              expr: _col5
                               type: string
                         outputColumnNames: _col0, _col1, _col2
                         File Output Operator
                           compressed: false
                           GlobalTableId: 1
-                          directory: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-32_178_1568785792517767702/-ext-10002
+                          directory: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-06_654_4872924397003947925/-ext-10002
                           NumFilesPerFileSink: 1
                           table:
                               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -243,21 +243,21 @@ STAGE PLANS:
                                 columns.types string:string:string
                                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                                location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/dest_j1
+                                location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/dest_j1
                                 name dest_j1
                                 serialization.ddl struct dest_j1 { string key, string value, string val2}
                                 serialization.format 1
                                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                                transient_lastDdlTime 1280084852
+                                transient_lastDdlTime 1280433726
                               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                               name: dest_j1
                           TotalFiles: 1
                           MultiFileSpray: false
       Needs Tagging: false
       Path -> Alias:
-        pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z]
+        pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z]
       Path -> Partition:
-        pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
+        pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
           Partition
             base file name: hr=11
             input format: org.apache.hadoop.mapred.TextInputFormat
@@ -271,13 +271,13 @@ STAGE PLANS:
               columns.types string:string
               file.inputformat org.apache.hadoop.mapred.TextInputFormat
               file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/srcpart
+              location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/srcpart
               name srcpart
               partition_columns ds/hr
               serialization.ddl struct srcpart { string key, string value}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              transient_lastDdlTime 1280082967
+              transient_lastDdlTime 1280430361
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -288,13 +288,13 @@ STAGE PLANS:
                 columns.types string:string
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/srcpart
+                location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/srcpart
                 name srcpart
                 partition_columns ds/hr
                 serialization.ddl struct srcpart { string key, string value}
                 serialization.format 1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                transient_lastDdlTime 1280082967
+                transient_lastDdlTime 1280430361
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
             name: srcpart
@@ -306,14 +306,14 @@ STAGE PLANS:
     Move Operator
       files:
           hdfs directory: true
-          source: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-32_178_1568785792517767702/-ext-10002
-          destination: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-32_178_1568785792517767702/-ext-10000
+          source: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-06_654_4872924397003947925/-ext-10002
+          destination: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-06_654_4872924397003947925/-ext-10000
 
   Stage: Stage-0
     Move Operator
       tables:
           replace: true
-          source: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-32_178_1568785792517767702/-ext-10000
+          source: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-06_654_4872924397003947925/-ext-10000
           table:
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -323,20 +323,20 @@ STAGE PLANS:
                 columns.types string:string:string
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/dest_j1
+                location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/dest_j1
                 name dest_j1
                 serialization.ddl struct dest_j1 { string key, string value, string val2}
                 serialization.format 1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                transient_lastDdlTime 1280084852
+                transient_lastDdlTime 1280433726
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: dest_j1
-          tmp directory: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-32_178_1568785792517767702/-ext-10001
+          tmp directory: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-06_654_4872924397003947925/-ext-10001
 
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-32_178_1568785792517767702/-ext-10002 
+        pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-06_654_4872924397003947925/-ext-10002 
             Reduce Output Operator
               sort order: 
               Map-reduce partition columns:
@@ -352,9 +352,9 @@ STAGE PLANS:
                     type: string
       Needs Tagging: false
       Path -> Alias:
-        pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-32_178_1568785792517767702/-ext-10002 [pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-32_178_1568785792517767702/-ext-10002]
+        pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-06_654_4872924397003947925/-ext-10002 [pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-06_654_4872924397003947925/-ext-10002]
       Path -> Partition:
-        pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-32_178_1568785792517767702/-ext-10002 
+        pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-06_654_4872924397003947925/-ext-10002 
           Partition
             base file name: -ext-10002
             input format: org.apache.hadoop.mapred.TextInputFormat
@@ -365,12 +365,12 @@ STAGE PLANS:
               columns.types string:string:string
               file.inputformat org.apache.hadoop.mapred.TextInputFormat
               file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/dest_j1
+              location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/dest_j1
               name dest_j1
               serialization.ddl struct dest_j1 { string key, string value, string val2}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              transient_lastDdlTime 1280084852
+              transient_lastDdlTime 1280433726
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -381,12 +381,12 @@ STAGE PLANS:
                 columns.types string:string:string
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/dest_j1
+                location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/dest_j1
                 name dest_j1
                 serialization.ddl struct dest_j1 { string key, string value, string val2}
                 serialization.format 1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                transient_lastDdlTime 1280084852
+                transient_lastDdlTime 1280433726
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: dest_j1
             name: dest_j1
@@ -395,7 +395,7 @@ STAGE PLANS:
           File Output Operator
             compressed: false
             GlobalTableId: 0
-            directory: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-32_178_1568785792517767702/-ext-10000
+            directory: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-06_654_4872924397003947925/-ext-10000
             NumFilesPerFileSink: 1
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
@@ -406,12 +406,12 @@ STAGE PLANS:
                   columns.types string:string:string
                   file.inputformat org.apache.hadoop.mapred.TextInputFormat
                   file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/dest_j1
+                  location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/dest_j1
                   name dest_j1
                   serialization.ddl struct dest_j1 { string key, string value, string val2}
                   serialization.format 1
                   serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                  transient_lastDdlTime 1280084852
+                  transient_lastDdlTime 1280433726
                 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 name: dest_j1
             TotalFiles: 1
@@ -443,11 +443,11 @@ POSTHOOK: Lineage: dest_j1.value SIMPLE 
 PREHOOK: query: select * from dest_j1 x order by x.key
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest_j1
-PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-25_12-07-35_322_6005529636073624222/-mr-10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-02-12_429_1549668140257729881/-mr-10000
 POSTHOOK: query: select * from dest_j1 x order by x.key
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest_j1
-POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-25_12-07-35_322_6005529636073624222/-mr-10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-02-12_429_1549668140257729881/-mr-10000
 POSTHOOK: Lineage: dest_j1.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: dest_j1.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ]
 POSTHOOK: Lineage: dest_j1.value SIMPLE [(srcpart)z.FieldSchema(name:hr, type:string, comment:null), ]
@@ -658,39 +658,39 @@ STAGE PLANS:
                   0 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[key]()]
                   1 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[key]()]
                   2 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[key]()]
-                outputColumnNames: _col0, _col3, _col5, _col6, _col7
+                outputColumnNames: _col0, _col5, _col9, _col10, _col11
                 Position of Big Table: 2
                 Select Operator
                   expressions:
                         expr: _col0
                         type: string
-                        expr: _col3
-                        type: string
                         expr: _col5
                         type: string
-                        expr: _col6
+                        expr: _col9
+                        type: string
+                        expr: _col10
                         type: string
-                        expr: _col7
+                        expr: _col11
                         type: string
-                  outputColumnNames: _col0, _col3, _col5, _col6, _col7
+                  outputColumnNames: _col0, _col5, _col9, _col10, _col11
                   Filter Operator
                     isSamplingPred: false
                     predicate:
-                        expr: ((_col6 = '2008-04-08') and (_col7 = 11))
+                        expr: ((_col10 = '2008-04-08') and (_col11 = 11))
                         type: boolean
                     Select Operator
                       expressions:
                             expr: _col0
                             type: string
-                            expr: _col5
+                            expr: _col9
                             type: string
-                            expr: _col3
+                            expr: _col5
                             type: string
                       outputColumnNames: _col0, _col1, _col2
                       File Output Operator
                         compressed: false
                         GlobalTableId: 1
-                        directory: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-43_638_9048807845471585670/-ext-10002
+                        directory: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-26_132_5292131418902400294/-ext-10002
                         NumFilesPerFileSink: 1
                         table:
                             input format: org.apache.hadoop.mapred.TextInputFormat
@@ -701,12 +701,12 @@ STAGE PLANS:
                               columns.types string:string:string
                               file.inputformat org.apache.hadoop.mapred.TextInputFormat
                               file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                              location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/dest_j1
+                              location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/dest_j1
                               name dest_j1
                               serialization.ddl struct dest_j1 { string key, string value, string val2}
                               serialization.format 1
                               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                              transient_lastDdlTime 1280084855
+                              transient_lastDdlTime 1280433732
                             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                             name: dest_j1
                         TotalFiles: 1
@@ -737,39 +737,39 @@ STAGE PLANS:
                     0 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[key]()]
                     1 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[key]()]
                     2 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[key]()]
-                  outputColumnNames: _col0, _col3, _col5, _col6, _col7
+                  outputColumnNames: _col0, _col5, _col9, _col10, _col11
                   Position of Big Table: 2
                   Select Operator
                     expressions:
                           expr: _col0
                           type: string
-                          expr: _col3
-                          type: string
                           expr: _col5
                           type: string
-                          expr: _col6
+                          expr: _col9
+                          type: string
+                          expr: _col10
                           type: string
-                          expr: _col7
+                          expr: _col11
                           type: string
-                    outputColumnNames: _col0, _col3, _col5, _col6, _col7
+                    outputColumnNames: _col0, _col5, _col9, _col10, _col11
                     Filter Operator
                       isSamplingPred: false
                       predicate:
-                          expr: ((_col6 = '2008-04-08') and (_col7 = 11))
+                          expr: ((_col10 = '2008-04-08') and (_col11 = 11))
                           type: boolean
                       Select Operator
                         expressions:
                               expr: _col0
                               type: string
-                              expr: _col5
+                              expr: _col9
                               type: string
-                              expr: _col3
+                              expr: _col5
                               type: string
                         outputColumnNames: _col0, _col1, _col2
                         File Output Operator
                           compressed: false
                           GlobalTableId: 1
-                          directory: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-43_638_9048807845471585670/-ext-10002
+                          directory: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-26_132_5292131418902400294/-ext-10002
                           NumFilesPerFileSink: 1
                           table:
                               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -780,12 +780,12 @@ STAGE PLANS:
                                 columns.types string:string:string
                                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                                location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/dest_j1
+                                location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/dest_j1
                                 name dest_j1
                                 serialization.ddl struct dest_j1 { string key, string value, string val2}
                                 serialization.format 1
                                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                                transient_lastDdlTime 1280084855
+                                transient_lastDdlTime 1280433732
                               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                               name: dest_j1
                           TotalFiles: 1
@@ -806,39 +806,39 @@ STAGE PLANS:
                     0 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[key]()]
                     1 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[key]()]
                     2 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[key]()]
-                  outputColumnNames: _col0, _col3, _col5, _col6, _col7
+                  outputColumnNames: _col0, _col5, _col9, _col10, _col11
                   Position of Big Table: 2
                   Select Operator
                     expressions:
                           expr: _col0
                           type: string
-                          expr: _col3
-                          type: string
                           expr: _col5
                           type: string
-                          expr: _col6
+                          expr: _col9
                           type: string
-                          expr: _col7
+                          expr: _col10
                           type: string
-                    outputColumnNames: _col0, _col3, _col5, _col6, _col7
+                          expr: _col11
+                          type: string
+                    outputColumnNames: _col0, _col5, _col9, _col10, _col11
                     Filter Operator
                       isSamplingPred: false
                       predicate:
-                          expr: ((_col6 = '2008-04-08') and (_col7 = 11))
+                          expr: ((_col10 = '2008-04-08') and (_col11 = 11))
                           type: boolean
                       Select Operator
                         expressions:
                               expr: _col0
                               type: string
-                              expr: _col5
+                              expr: _col9
                               type: string
-                              expr: _col3
+                              expr: _col5
                               type: string
                         outputColumnNames: _col0, _col1, _col2
                         File Output Operator
                           compressed: false
                           GlobalTableId: 1
-                          directory: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-43_638_9048807845471585670/-ext-10002
+                          directory: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-26_132_5292131418902400294/-ext-10002
                           NumFilesPerFileSink: 1
                           table:
                               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -849,21 +849,21 @@ STAGE PLANS:
                                 columns.types string:string:string
                                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                                location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/dest_j1
+                                location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/dest_j1
                                 name dest_j1
                                 serialization.ddl struct dest_j1 { string key, string value, string val2}
                                 serialization.format 1
                                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                                transient_lastDdlTime 1280084855
+                                transient_lastDdlTime 1280433732
                               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                               name: dest_j1
                           TotalFiles: 1
                           MultiFileSpray: false
       Needs Tagging: false
       Path -> Alias:
-        pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z]
+        pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z]
       Path -> Partition:
-        pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
+        pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
           Partition
             base file name: hr=11
             input format: org.apache.hadoop.mapred.TextInputFormat
@@ -877,13 +877,13 @@ STAGE PLANS:
               columns.types string:string
               file.inputformat org.apache.hadoop.mapred.TextInputFormat
               file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/srcpart
+              location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/srcpart
               name srcpart
               partition_columns ds/hr
               serialization.ddl struct srcpart { string key, string value}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              transient_lastDdlTime 1280082967
+              transient_lastDdlTime 1280430361
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -894,13 +894,13 @@ STAGE PLANS:
                 columns.types string:string
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/srcpart
+                location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/srcpart
                 name srcpart
                 partition_columns ds/hr
                 serialization.ddl struct srcpart { string key, string value}
                 serialization.format 1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                transient_lastDdlTime 1280082967
+                transient_lastDdlTime 1280430361
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: srcpart
             name: srcpart
@@ -912,14 +912,14 @@ STAGE PLANS:
     Move Operator
       files:
           hdfs directory: true
-          source: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-43_638_9048807845471585670/-ext-10002
-          destination: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-43_638_9048807845471585670/-ext-10000
+          source: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-26_132_5292131418902400294/-ext-10002
+          destination: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-26_132_5292131418902400294/-ext-10000
 
   Stage: Stage-0
     Move Operator
       tables:
           replace: true
-          source: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-43_638_9048807845471585670/-ext-10000
+          source: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-26_132_5292131418902400294/-ext-10000
           table:
               input format: org.apache.hadoop.mapred.TextInputFormat
               output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -929,20 +929,20 @@ STAGE PLANS:
                 columns.types string:string:string
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/dest_j1
+                location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/dest_j1
                 name dest_j1
                 serialization.ddl struct dest_j1 { string key, string value, string val2}
                 serialization.format 1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                transient_lastDdlTime 1280084855
+                transient_lastDdlTime 1280433732
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: dest_j1
-          tmp directory: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-43_638_9048807845471585670/-ext-10001
+          tmp directory: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-26_132_5292131418902400294/-ext-10001
 
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-43_638_9048807845471585670/-ext-10002 
+        pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-26_132_5292131418902400294/-ext-10002 
             Reduce Output Operator
               sort order: 
               Map-reduce partition columns:
@@ -958,9 +958,9 @@ STAGE PLANS:
                     type: string
       Needs Tagging: false
       Path -> Alias:
-        pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-43_638_9048807845471585670/-ext-10002 [pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-43_638_9048807845471585670/-ext-10002]
+        pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-26_132_5292131418902400294/-ext-10002 [pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-26_132_5292131418902400294/-ext-10002]
       Path -> Partition:
-        pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-43_638_9048807845471585670/-ext-10002 
+        pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-26_132_5292131418902400294/-ext-10002 
           Partition
             base file name: -ext-10002
             input format: org.apache.hadoop.mapred.TextInputFormat
@@ -971,12 +971,12 @@ STAGE PLANS:
               columns.types string:string:string
               file.inputformat org.apache.hadoop.mapred.TextInputFormat
               file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-              location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/dest_j1
+              location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/dest_j1
               name dest_j1
               serialization.ddl struct dest_j1 { string key, string value, string val2}
               serialization.format 1
               serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-              transient_lastDdlTime 1280084855
+              transient_lastDdlTime 1280433732
             serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
           
               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -987,12 +987,12 @@ STAGE PLANS:
                 columns.types string:string:string
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/dest_j1
+                location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/dest_j1
                 name dest_j1
                 serialization.ddl struct dest_j1 { string key, string value, string val2}
                 serialization.format 1
                 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                transient_lastDdlTime 1280084855
+                transient_lastDdlTime 1280433732
               serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
               name: dest_j1
             name: dest_j1
@@ -1001,7 +1001,7 @@ STAGE PLANS:
           File Output Operator
             compressed: false
             GlobalTableId: 0
-            directory: pfile:/data/users/jssarma/hive_trunk/build/ql/scratchdir/hive_2010-07-25_12-07-43_638_9048807845471585670/-ext-10000
+            directory: pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/scratchdir/hive_2010-07-29_13-02-26_132_5292131418902400294/-ext-10000
             NumFilesPerFileSink: 1
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
@@ -1012,12 +1012,12 @@ STAGE PLANS:
                   columns.types string:string:string
                   file.inputformat org.apache.hadoop.mapred.TextInputFormat
                   file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  location pfile:/data/users/jssarma/hive_trunk/build/ql/test/data/warehouse/dest_j1
+                  location pfile:/Users/heyongqiang/Documents/workspace/Hive-2/build/ql/test/data/warehouse/dest_j1
                   name dest_j1
                   serialization.ddl struct dest_j1 { string key, string value, string val2}
                   serialization.format 1
                   serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                  transient_lastDdlTime 1280084855
+                  transient_lastDdlTime 1280433732
                 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                 name: dest_j1
             TotalFiles: 1
@@ -1056,11 +1056,11 @@ POSTHOOK: Lineage: src_copy.value SIMPLE
 PREHOOK: query: select * from dest_j1 x order by x.key
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest_j1
-PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-25_12-07-46_911_6550146957730775353/-mr-10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-02-32_947_7414062586677410982/-mr-10000
 POSTHOOK: query: select * from dest_j1 x order by x.key
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest_j1
-POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-25_12-07-46_911_6550146957730775353/-mr-10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-02-32_947_7414062586677410982/-mr-10000
 POSTHOOK: Lineage: dest_j1.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: dest_j1.key SIMPLE [(src1_copy)x.FieldSchema(name:key, type:string, comment:null), ]
 POSTHOOK: Lineage: dest_j1.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ]

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/join_rc.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/join_rc.q.out?rev=980659&r1=980658&r2=980659&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/join_rc.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/join_rc.q.out Fri Jul 30 06:40:04 2010
@@ -91,12 +91,12 @@ STAGE PLANS:
             0 {VALUE._col0}
             1 {VALUE._col1}
           handleSkewJoin: false
-          outputColumnNames: _col0, _col3
+          outputColumnNames: _col0, _col5
           Select Operator
             expressions:
                   expr: _col0
                   type: string
-                  expr: _col3
+                  expr: _col5
                   type: string
             outputColumnNames: _col0, _col1
             File Output Operator
@@ -116,13 +116,13 @@ FROM join_rc1 JOIN join_rc2 ON join_rc1.
 PREHOOK: type: QUERY
 PREHOOK: Input: default@join_rc2
 PREHOOK: Input: default@join_rc1
-PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-36-04_236_2954829201146901111/10000
+PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-02-49_005_2982007273012088026/-mr-10000
 POSTHOOK: query: select join_rc1.key, join_rc2.value
 FROM join_rc1 JOIN join_rc2 ON join_rc1.key = join_rc2.key
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@join_rc2
 POSTHOOK: Input: default@join_rc1
-POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-36-04_236_2954829201146901111/10000
+POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-07-29_13-02-49_005_2982007273012088026/-mr-10000
 POSTHOOK: Lineage: join_rc1.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: join_rc1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
 POSTHOOK: Lineage: join_rc2.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]



Mime
View raw message