hadoop-hive-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From zs...@apache.org
Subject svn commit: r727337 [3/3] - in /hadoop/hive/trunk: ./ ql/src/java/org/apache/hadoop/hive/ql/exec/ ql/src/java/org/apache/hadoop/hive/ql/metadata/ ql/src/java/org/apache/hadoop/hive/ql/parse/ ql/src/java/org/apache/hadoop/hive/ql/plan/ ql/src/test/queri...
Date Wed, 17 Dec 2008 10:35:46 GMT
Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce6.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce6.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce6.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce6.q.out Wed Dec 17 02:35:44 2008
@@ -44,13 +44,23 @@
                       type: string
       Reduce Operator Tree:
         Extract
-          File Output Operator
-            compressed: false
-            table:
-                input format: org.apache.hadoop.mapred.TextInputFormat
-                output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
-                name: dest1
+          Select Operator
+            expressions:
+                  expr: UDFToInteger(0)
+                  type: int
+                  expr: 1
+                  type: int
+                  expr: 2
+                  type: int
+                  expr: 3
+                  type: string
+            File Output Operator
+              compressed: false
+              table:
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+                  serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+                  name: dest1
 
   Stage: Stage-0
     Move Operator
@@ -59,7 +69,7 @@
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+                serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
                 name: dest1
 
 

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias1.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias1.q.out Wed Dec 17 02:35:44 2008
@@ -49,7 +49,7 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        /tmp/hive-jssarma/238416396/213974514.10001 
+        /tmp/hive-zshao/10020404/161743246.10001 
           Reduce Output Operator
             key expressions:
                   expr: 0
@@ -69,7 +69,7 @@
           keys:
                 expr: KEY.0
                 type: string
-          mode: unknown
+          mode: final
           Select Operator
             expressions:
                   expr: '1234'
@@ -78,13 +78,21 @@
                   type: string
                   expr: 1
                   type: bigint
-            File Output Operator
-              compressed: false
-              table:
-                  input format: org.apache.hadoop.mapred.TextInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                  serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
-                  name: dest1
+            Select Operator
+              expressions:
+                    expr: 0
+                    type: string
+                    expr: UDFToInteger(1)
+                    type: int
+                    expr: UDFToDouble(2)
+                    type: double
+              File Output Operator
+                compressed: false
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+                    serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+                    name: dest1
 
   Stage: Stage-0
     Move Operator
@@ -93,64 +101,64 @@
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+                serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
                 name: dest1
 
 
-1234	0	3
-1234	10	1
-1234	11	1
-1234	12	2
-1234	15	2
-1234	17	1
-1234	18	2
-1234	19	1
-1234	2	1
-1234	20	1
-1234	24	2
-1234	26	2
-1234	27	1
-1234	28	1
-1234	30	1
-1234	33	1
-1234	34	1
-1234	35	3
-1234	37	2
-1234	4	1
-1234	41	1
-1234	42	2
-1234	43	1
-1234	44	1
-1234	47	1
-1234	5	3
-1234	51	2
-1234	53	1
-1234	54	1
-1234	57	1
-1234	58	2
-1234	64	1
-1234	65	1
-1234	66	1
-1234	67	2
-1234	69	1
-1234	70	3
-1234	72	2
-1234	74	1
-1234	76	2
-1234	77	1
-1234	78	1
-1234	8	1
-1234	80	1
-1234	82	1
-1234	83	2
-1234	84	2
-1234	85	1
-1234	86	1
-1234	87	1
-1234	9	1
-1234	90	3
-1234	92	1
-1234	95	2
-1234	96	1
-1234	97	2
-1234	98	2
+1234	0	3.0
+1234	10	1.0
+1234	11	1.0
+1234	12	2.0
+1234	15	2.0
+1234	17	1.0
+1234	18	2.0
+1234	19	1.0
+1234	2	1.0
+1234	20	1.0
+1234	24	2.0
+1234	26	2.0
+1234	27	1.0
+1234	28	1.0
+1234	30	1.0
+1234	33	1.0
+1234	34	1.0
+1234	35	3.0
+1234	37	2.0
+1234	4	1.0
+1234	41	1.0
+1234	42	2.0
+1234	43	1.0
+1234	44	1.0
+1234	47	1.0
+1234	5	3.0
+1234	51	2.0
+1234	53	1.0
+1234	54	1.0
+1234	57	1.0
+1234	58	2.0
+1234	64	1.0
+1234	65	1.0
+1234	66	1.0
+1234	67	2.0
+1234	69	1.0
+1234	70	3.0
+1234	72	2.0
+1234	74	1.0
+1234	76	2.0
+1234	77	1.0
+1234	78	1.0
+1234	8	1.0
+1234	80	1.0
+1234	82	1.0
+1234	83	2.0
+1234	84	2.0
+1234	85	1.0
+1234	86	1.0
+1234	87	1.0
+1234	9	1.0
+1234	90	3.0
+1234	92	1.0
+1234	95	2.0
+1234	96	1.0
+1234	97	2.0
+1234	98	2.0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias2.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias2.q.out Wed Dec 17 02:35:44 2008
@@ -49,7 +49,7 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        /tmp/hive-jssarma/506681419/141455468.10001 
+        /tmp/hive-zshao/55994326/602553370.10001 
           Reduce Output Operator
             key expressions:
                   expr: 0
@@ -69,7 +69,7 @@
           keys:
                 expr: KEY.0
                 type: string
-          mode: unknown
+          mode: final
           Select Operator
             expressions:
                   expr: '1234'
@@ -78,13 +78,21 @@
                   type: string
                   expr: 1
                   type: bigint
-            File Output Operator
-              compressed: false
-              table:
-                  input format: org.apache.hadoop.mapred.TextInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                  serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
-                  name: dest1
+            Select Operator
+              expressions:
+                    expr: 0
+                    type: string
+                    expr: UDFToInteger(1)
+                    type: int
+                    expr: UDFToDouble(2)
+                    type: double
+              File Output Operator
+                compressed: false
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+                    serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+                    name: dest1
 
   Stage: Stage-0
     Move Operator
@@ -93,64 +101,64 @@
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+                serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
                 name: dest1
 
 
-1234	0	3
-1234	10	1
-1234	11	1
-1234	12	2
-1234	15	2
-1234	17	1
-1234	18	2
-1234	19	1
-1234	2	1
-1234	20	1
-1234	24	2
-1234	26	2
-1234	27	1
-1234	28	1
-1234	30	1
-1234	33	1
-1234	34	1
-1234	35	3
-1234	37	2
-1234	4	1
-1234	41	1
-1234	42	2
-1234	43	1
-1234	44	1
-1234	47	1
-1234	5	3
-1234	51	2
-1234	53	1
-1234	54	1
-1234	57	1
-1234	58	2
-1234	64	1
-1234	65	1
-1234	66	1
-1234	67	2
-1234	69	1
-1234	70	3
-1234	72	2
-1234	74	1
-1234	76	2
-1234	77	1
-1234	78	1
-1234	8	1
-1234	80	1
-1234	82	1
-1234	83	2
-1234	84	2
-1234	85	1
-1234	86	1
-1234	87	1
-1234	9	1
-1234	90	3
-1234	92	1
-1234	95	2
-1234	96	1
-1234	97	2
-1234	98	2
+1234	0	3.0
+1234	10	1.0
+1234	11	1.0
+1234	12	2.0
+1234	15	2.0
+1234	17	1.0
+1234	18	2.0
+1234	19	1.0
+1234	2	1.0
+1234	20	1.0
+1234	24	2.0
+1234	26	2.0
+1234	27	1.0
+1234	28	1.0
+1234	30	1.0
+1234	33	1.0
+1234	34	1.0
+1234	35	3.0
+1234	37	2.0
+1234	4	1.0
+1234	41	1.0
+1234	42	2.0
+1234	43	1.0
+1234	44	1.0
+1234	47	1.0
+1234	5	3.0
+1234	51	2.0
+1234	53	1.0
+1234	54	1.0
+1234	57	1.0
+1234	58	2.0
+1234	64	1.0
+1234	65	1.0
+1234	66	1.0
+1234	67	2.0
+1234	69	1.0
+1234	70	3.0
+1234	72	2.0
+1234	74	1.0
+1234	76	2.0
+1234	77	1.0
+1234	78	1.0
+1234	8	1.0
+1234	80	1.0
+1234	82	1.0
+1234	83	2.0
+1234	84	2.0
+1234	85	1.0
+1234	86	1.0
+1234	87	1.0
+1234	9	1.0
+1234	90	3.0
+1234	92	1.0
+1234	95	2.0
+1234	96	1.0
+1234	97	2.0
+1234	98	2.0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/quote1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/quote1.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/quote1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/quote1.q.out Wed Dec 17 02:35:44 2008
@@ -20,13 +20,19 @@
                       type: string
                       expr: value
                       type: string
-                File Output Operator
-                  compressed: false
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
-                      name: dest1
+                Select Operator
+                  expressions:
+                        expr: UDFToInteger(0)
+                        type: int
+                        expr: 1
+                        type: string
+                  File Output Operator
+                    compressed: false
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+                        serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+                        name: dest1
 
   Stage: Stage-0
     Move Operator
@@ -37,7 +43,7 @@
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+                serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
                 name: dest1
 
 
@@ -60,7 +66,7 @@
               Select Operator
                 expressions:
                       expr: location
-                      type: string
+                      type: int
                       expr: type
                       type: string
                       expr: table

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample1.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample1.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample1.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample1.q.out Wed Dec 17 02:35:44 2008
@@ -28,29 +28,39 @@
                         type: string
                         expr: hr
                         type: string
-                  File Output Operator
-                    compressed: false
-                    directory: /tmp/hive-jssarma/37742099/114802476.10000.insclause-0
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                        properties:
-                          name dest1
-                          serialization.ddl struct dest1 { i32 key, string value, string dt, string hr}
-                          serialization.format 1
-                          columns key,value,dt,hr
-                          bucket_count -1
-                          serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
-                          file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                          file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                          location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
-                        serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
-                        name: dest1
+                  Select Operator
+                    expressions:
+                          expr: UDFToInteger(0)
+                          type: int
+                          expr: 1
+                          type: string
+                          expr: 2
+                          type: string
+                          expr: 3
+                          type: string
+                    File Output Operator
+                      compressed: false
+                      directory: /tmp/hive-zshao/860547727/34027035.10000.insclause-0
+                      table:
+                          input format: org.apache.hadoop.mapred.TextInputFormat
+                          output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+                          properties:
+                            name dest1
+                            serialization.ddl struct dest1 { i32 key, string value, string dt, string hr}
+                            serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
+                            columns key,value,dt,hr
+                            bucket_count -1
+                            serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+                            file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                            file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+                            location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+                          serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+                          name: dest1
       Needs Tagging: false
       Path -> Alias:
-        file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
+        file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
       Path -> Partition:
-        file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
+        file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 
           Partition
             partition values:
               ds 2008-04-08
@@ -68,7 +78,7 @@
                 serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcpart
+                location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcpart
               serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
               name: srcpart
 
@@ -76,21 +86,21 @@
     Move Operator
       tables:
             replace: true
-            source: /tmp/hive-jssarma/37742099/114802476.10000.insclause-0
+            source: /tmp/hive-zshao/860547727/34027035.10000.insclause-0
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
                 properties:
                   name dest1
                   serialization.ddl struct dest1 { i32 key, string value, string dt, string hr}
-                  serialization.format 1
+                  serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
                   columns key,value,dt,hr
                   bucket_count -1
-                  serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+                  serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
                   file.inputformat org.apache.hadoop.mapred.TextInputFormat
                   file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                  location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
-                serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+                  location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+                serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
                 name: dest1
 
 

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample2.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample2.q.out Wed Dec 17 02:35:44 2008
@@ -16,29 +16,35 @@
                     type: string
                     expr: value
                     type: string
-              File Output Operator
-                compressed: false
-                directory: /tmp/hive-jssarma/515988900/822291761.10000.insclause-0
-                table:
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                    properties:
-                      name dest1
-                      serialization.ddl struct dest1 { i32 key, string value}
-                      serialization.format 1
-                      columns key,value
-                      bucket_count -1
-                      serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
-                      file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                      file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                      location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
-                    serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
-                    name: dest1
+              Select Operator
+                expressions:
+                      expr: UDFToInteger(0)
+                      type: int
+                      expr: 1
+                      type: string
+                File Output Operator
+                  compressed: false
+                  directory: /tmp/hive-zshao/427867399/126053246.10000.insclause-0
+                  table:
+                      input format: org.apache.hadoop.mapred.TextInputFormat
+                      output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+                      properties:
+                        name dest1
+                        serialization.ddl struct dest1 { i32 key, string value}
+                        serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
+                        columns key,value
+                        bucket_count -1
+                        serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+                        file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                        file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+                        location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+                      serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+                      name: dest1
       Needs Tagging: false
       Path -> Alias:
-        file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket/kv1.txt 
+        file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket/kv1.txt 
       Path -> Partition:
-        file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket/kv1.txt 
+        file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket/kv1.txt 
           Partition
           
               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -53,7 +59,7 @@
                 serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket
+                location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket
               serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
               name: srcbucket
 
@@ -61,21 +67,21 @@
     Move Operator
       tables:
             replace: true
-            source: /tmp/hive-jssarma/515988900/822291761.10000.insclause-0
+            source: /tmp/hive-zshao/427867399/126053246.10000.insclause-0
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
                 properties:
                   name dest1
                   serialization.ddl struct dest1 { i32 key, string value}
-                  serialization.format 1
+                  serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
                   columns key,value
                   bucket_count -1
-                  serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+                  serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
                   file.inputformat org.apache.hadoop.mapred.TextInputFormat
                   file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                  location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
-                serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+                  location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+                serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
                 name: dest1
 
 

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample4.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample4.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample4.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample4.q.out Wed Dec 17 02:35:44 2008
@@ -16,29 +16,35 @@
                     type: string
                     expr: value
                     type: string
-              File Output Operator
-                compressed: false
-                directory: /tmp/hive-jssarma/1539308576/222195283.10000.insclause-0
-                table:
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                    properties:
-                      name dest1
-                      serialization.ddl struct dest1 { i32 key, string value}
-                      serialization.format 1
-                      columns key,value
-                      bucket_count -1
-                      serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
-                      file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                      file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                      location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
-                    serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
-                    name: dest1
+              Select Operator
+                expressions:
+                      expr: UDFToInteger(0)
+                      type: int
+                      expr: 1
+                      type: string
+                File Output Operator
+                  compressed: false
+                  directory: /tmp/hive-zshao/134470671/579412318.10000.insclause-0
+                  table:
+                      input format: org.apache.hadoop.mapred.TextInputFormat
+                      output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+                      properties:
+                        name dest1
+                        serialization.ddl struct dest1 { i32 key, string value}
+                        serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
+                        columns key,value
+                        bucket_count -1
+                        serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+                        file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                        file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+                        location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+                      serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+                      name: dest1
       Needs Tagging: false
       Path -> Alias:
-        file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket/kv1.txt 
+        file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket/kv1.txt 
       Path -> Partition:
-        file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket/kv1.txt 
+        file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket/kv1.txt 
           Partition
           
               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -53,7 +59,7 @@
                 serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket
+                location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket
               serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
               name: srcbucket
 
@@ -61,21 +67,21 @@
     Move Operator
       tables:
             replace: true
-            source: /tmp/hive-jssarma/1539308576/222195283.10000.insclause-0
+            source: /tmp/hive-zshao/134470671/579412318.10000.insclause-0
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
                 properties:
                   name dest1
                   serialization.ddl struct dest1 { i32 key, string value}
-                  serialization.format 1
+                  serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
                   columns key,value
                   bucket_count -1
-                  serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+                  serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
                   file.inputformat org.apache.hadoop.mapred.TextInputFormat
                   file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                  location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
-                serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+                  location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+                serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
                 name: dest1
 
 

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample5.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample5.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample5.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample5.q.out Wed Dec 17 02:35:44 2008
@@ -20,29 +20,35 @@
                       type: string
                       expr: value
                       type: string
-                File Output Operator
-                  compressed: false
-                  directory: /tmp/hive-jssarma/611887541/74229442.10000.insclause-0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                      properties:
-                        name dest1
-                        serialization.ddl struct dest1 { i32 key, string value}
-                        serialization.format 1
-                        columns key,value
-                        bucket_count -1
-                        serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
-                        file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                        file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                        location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
-                      serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
-                      name: dest1
+                Select Operator
+                  expressions:
+                        expr: UDFToInteger(0)
+                        type: int
+                        expr: 1
+                        type: string
+                  File Output Operator
+                    compressed: false
+                    directory: /tmp/hive-zshao/211838955/159322479.10000.insclause-0
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+                        properties:
+                          name dest1
+                          serialization.ddl struct dest1 { i32 key, string value}
+                          serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
+                          columns key,value
+                          bucket_count -1
+                          serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+                          file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                          file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+                          location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+                        serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+                        name: dest1
       Needs Tagging: false
       Path -> Alias:
-        file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket 
+        file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket 
       Path -> Partition:
-        file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket 
+        file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket 
           Partition
           
               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -57,7 +63,7 @@
                 serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket
+                location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket
               serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
               name: srcbucket
 
@@ -65,21 +71,21 @@
     Move Operator
       tables:
             replace: true
-            source: /tmp/hive-jssarma/611887541/74229442.10000.insclause-0
+            source: /tmp/hive-zshao/211838955/159322479.10000.insclause-0
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
                 properties:
                   name dest1
                   serialization.ddl struct dest1 { i32 key, string value}
-                  serialization.format 1
+                  serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
                   columns key,value
                   bucket_count -1
-                  serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+                  serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
                   file.inputformat org.apache.hadoop.mapred.TextInputFormat
                   file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                  location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
-                serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+                  location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+                serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
                 name: dest1
 
 

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample6.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample6.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample6.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample6.q.out Wed Dec 17 02:35:44 2008
@@ -20,29 +20,35 @@
                       type: string
                       expr: value
                       type: string
-                File Output Operator
-                  compressed: false
-                  directory: /tmp/hive-jssarma/862051292/1235999863.10000.insclause-0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                      properties:
-                        name dest1
-                        serialization.ddl struct dest1 { i32 key, string value}
-                        serialization.format 1
-                        columns key,value
-                        bucket_count -1
-                        serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
-                        file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                        file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                        location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
-                      serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
-                      name: dest1
+                Select Operator
+                  expressions:
+                        expr: UDFToInteger(0)
+                        type: int
+                        expr: 1
+                        type: string
+                  File Output Operator
+                    compressed: false
+                    directory: /tmp/hive-zshao/25298408/261228818.10000.insclause-0
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+                        properties:
+                          name dest1
+                          serialization.ddl struct dest1 { i32 key, string value}
+                          serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
+                          columns key,value
+                          bucket_count -1
+                          serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+                          file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                          file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+                          location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+                        serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+                        name: dest1
       Needs Tagging: false
       Path -> Alias:
-        file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket/kv1.txt 
+        file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket/kv1.txt 
       Path -> Partition:
-        file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket/kv1.txt 
+        file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket/kv1.txt 
           Partition
           
               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -57,7 +63,7 @@
                 serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket
+                location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket
               serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
               name: srcbucket
 
@@ -65,21 +71,21 @@
     Move Operator
       tables:
             replace: true
-            source: /tmp/hive-jssarma/862051292/1235999863.10000.insclause-0
+            source: /tmp/hive-zshao/25298408/261228818.10000.insclause-0
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
                 properties:
                   name dest1
                   serialization.ddl struct dest1 { i32 key, string value}
-                  serialization.format 1
+                  serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
                   columns key,value
                   bucket_count -1
-                  serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+                  serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
                   file.inputformat org.apache.hadoop.mapred.TextInputFormat
                   file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                  location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
-                serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+                  location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+                serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
                 name: dest1
 
 

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out Wed Dec 17 02:35:44 2008
@@ -24,29 +24,35 @@
                         type: string
                         expr: value
                         type: string
-                  File Output Operator
-                    compressed: false
-                    directory: /tmp/hive-jssarma/429463772/457590891.10000.insclause-0
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                        properties:
-                          name dest1
-                          serialization.ddl struct dest1 { i32 key, string value}
-                          serialization.format 1
-                          columns key,value
-                          bucket_count -1
-                          serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
-                          file.inputformat org.apache.hadoop.mapred.TextInputFormat
-                          file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                          location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
-                        serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
-                        name: dest1
+                  Select Operator
+                    expressions:
+                          expr: UDFToInteger(0)
+                          type: int
+                          expr: 1
+                          type: string
+                    File Output Operator
+                      compressed: false
+                      directory: /tmp/hive-zshao/444520495/211065154.10000.insclause-0
+                      table:
+                          input format: org.apache.hadoop.mapred.TextInputFormat
+                          output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+                          properties:
+                            name dest1
+                            serialization.ddl struct dest1 { i32 key, string value}
+                            serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
+                            columns key,value
+                            bucket_count -1
+                            serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+                            file.inputformat org.apache.hadoop.mapred.TextInputFormat
+                            file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+                            location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+                          serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
+                          name: dest1
       Needs Tagging: false
       Path -> Alias:
-        file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket/kv1.txt 
+        file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket/kv1.txt 
       Path -> Partition:
-        file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket/kv1.txt 
+        file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket/kv1.txt 
           Partition
           
               input format: org.apache.hadoop.mapred.TextInputFormat
@@ -61,7 +67,7 @@
                 serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
                 file.inputformat org.apache.hadoop.mapred.TextInputFormat
                 file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/srcbucket
+                location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/srcbucket
               serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
               name: srcbucket
 
@@ -69,21 +75,21 @@
     Move Operator
       tables:
             replace: true
-            source: /tmp/hive-jssarma/429463772/457590891.10000.insclause-0
+            source: /tmp/hive-zshao/444520495/211065154.10000.insclause-0
             table:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
                 properties:
                   name dest1
                   serialization.ddl struct dest1 { i32 key, string value}
-                  serialization.format 1
+                  serialization.format org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
                   columns key,value
                   bucket_count -1
-                  serialization.lib org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+                  serialization.lib org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
                   file.inputformat org.apache.hadoop.mapred.TextInputFormat
                   file.outputformat org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-                  location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive-trunk-compress/build/ql/test/data/warehouse/dest1
-                serde: org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe
+                  location file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/dest1
+                serde: org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe
                 name: dest1
 
 

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out Wed Dec 17 02:35:44 2008
@@ -45,7 +45,7 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        /tmp/hive-jssarma/438302216/1483130688.10002 
+        /tmp/hive-zshao/368989435/823759952.10002 
           Reduce Output Operator
             key expressions:
                   expr: 0
@@ -65,7 +65,7 @@
           keys:
                 expr: KEY.0
                 type: string
-          mode: unknown
+          mode: final
           Select Operator
             expressions:
                   expr: 0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out Wed Dec 17 02:35:44 2008
@@ -40,7 +40,7 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        /tmp/hive-jssarma/1415018212/1021817615.10001 
+        /tmp/hive-zshao/546652046/1647731334.10001 
           Reduce Output Operator
             sort order: 
             tag: -1
@@ -63,7 +63,7 @@
                 expr: count(VALUE.2)
                 expr: sum(VALUE.3)
                 expr: min(VALUE.4)
-          mode: unknown
+          mode: final
           Select Operator
             expressions:
                   expr: 2
@@ -71,7 +71,7 @@
                   expr: 3
                   type: double
                   expr: 1
-                  type: string
+                  type: double
                   expr: 4
                   type: double
                   expr: 0

Modified: hadoop/hive/trunk/ql/src/test/results/clientpositive/union2.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/clientpositive/union2.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/clientpositive/union2.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/clientpositive/union2.q.out Wed Dec 17 02:35:44 2008
@@ -57,7 +57,7 @@
   Stage: Stage-2
     Map Reduce
       Alias -> Map Operator Tree:
-        /tmp/hive-jssarma/32027196/47394207.10002 
+        /tmp/hive-zshao/175321401/121204523.10002 
           Reduce Output Operator
             sort order: 
             tag: -1
@@ -68,7 +68,7 @@
         Group By Operator
           aggregations:
                 expr: count(VALUE.0)
-          mode: unknown
+          mode: final
           Select Operator
             expressions:
                   expr: 0

Modified: hadoop/hive/trunk/ql/src/test/results/compiler/errors/invalid_index.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/compiler/errors/invalid_index.q.out?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/compiler/errors/invalid_index.q.out (original)
+++ hadoop/hive/trunk/ql/src/test/results/compiler/errors/invalid_index.q.out Wed Dec 17 02:35:44 2008
@@ -1,2 +1,2 @@
 Semantic Exception: 
-line 2:36 [] not Valid on Non Collection Types 0
\ No newline at end of file
+line 2:36 [] not Valid on Non Collection Types 0: string
\ No newline at end of file

Modified: hadoop/hive/trunk/ql/src/test/results/compiler/plan/groupby3.q.xml
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/compiler/plan/groupby3.q.xml?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/compiler/plan/groupby3.q.xml (original)
+++ hadoop/hive/trunk/ql/src/test/results/compiler/plan/groupby3.q.xml Wed Dec 17 02:35:44 2008
@@ -20,7 +20,7 @@
         <void property="aliasToWork"> 
          <object class="java.util.HashMap"> 
           <void method="put"> 
-           <string>/tmp/hive-njain/332856862/73535516.10002</string> 
+           <string>/tmp/hive-zshao/1268121/343796403.10002</string> 
            <object id="ReduceSinkOperator0" class="org.apache.hadoop.hive.ql.exec.ReduceSinkOperator"> 
             <void property="conf"> 
              <object class="org.apache.hadoop.hive.ql.plan.reduceSinkDesc"> 
@@ -176,7 +176,7 @@
                     <void property="conf"> 
                      <object class="org.apache.hadoop.hive.ql.plan.fileSinkDesc"> 
                       <void property="dirName"> 
-                       <string>/tmp/hive-njain/332856862/73535516.10002</string> 
+                       <string>/tmp/hive-zshao/1268121/343796403.10002</string> 
                       </void> 
                       <void property="tableInfo"> 
                        <object id="tableDesc2" class="org.apache.hadoop.hive.ql.plan.tableDesc"> 
@@ -853,10 +853,10 @@
         <void property="pathToAliases"> 
          <object class="java.util.LinkedHashMap"> 
           <void method="put"> 
-           <string>/tmp/hive-njain/332856862/73535516.10002</string> 
+           <string>/tmp/hive-zshao/1268121/343796403.10002</string> 
            <object class="java.util.ArrayList"> 
             <void method="add"> 
-             <string>/tmp/hive-njain/332856862/73535516.10002</string> 
+             <string>/tmp/hive-zshao/1268121/343796403.10002</string> 
             </void> 
            </object> 
           </void> 
@@ -865,7 +865,7 @@
         <void property="pathToPartitionInfo"> 
          <object class="java.util.LinkedHashMap"> 
           <void method="put"> 
-           <string>/tmp/hive-njain/332856862/73535516.10002</string> 
+           <string>/tmp/hive-zshao/1268121/343796403.10002</string> 
            <object class="org.apache.hadoop.hive.ql.plan.partitionDesc"> 
             <void property="tableDesc"> 
              <object idref="tableDesc2"/> 
@@ -887,7 +887,7 @@
                   <void property="conf"> 
                    <object class="org.apache.hadoop.hive.ql.plan.fileSinkDesc"> 
                     <void property="dirName"> 
-                     <string>/tmp/hive-njain/229394155.10001.insclause-0</string> 
+                     <string>/tmp/hive-zshao/147248168.10001.insclause-0</string> 
                     </void> 
                     <void property="tableInfo"> 
                      <object class="org.apache.hadoop.hive.ql.plan.tableDesc"> 
@@ -943,7 +943,7 @@
                          <string>1</string> 
                         </void> 
                         <void property="type"> 
-                         <object idref="PrimitiveTypeInfo0"/> 
+                         <object idref="PrimitiveTypeInfo1"/> 
                         </void> 
                        </object> 
                       </void> 
@@ -953,7 +953,7 @@
                          <string>2</string> 
                         </void> 
                         <void property="type"> 
-                         <object idref="PrimitiveTypeInfo0"/> 
+                         <object idref="PrimitiveTypeInfo1"/> 
                         </void> 
                        </object> 
                       </void> 
@@ -1005,7 +1005,7 @@
                      <string>2</string> 
                     </void> 
                     <void property="typeInfo"> 
-                     <object idref="PrimitiveTypeInfo0"/> 
+                     <object idref="PrimitiveTypeInfo1"/> 
                     </void> 
                    </object> 
                   </void> 
@@ -1015,7 +1015,7 @@
                      <string>0</string> 
                     </void> 
                     <void property="typeInfo"> 
-                     <object idref="PrimitiveTypeInfo0"/> 
+                     <object idref="PrimitiveTypeInfo1"/> 
                     </void> 
                    </object> 
                   </void> 
@@ -1199,7 +1199,7 @@
                  <string>0</string> 
                 </void> 
                 <void property="type"> 
-                 <object idref="PrimitiveTypeInfo0"/> 
+                 <object idref="PrimitiveTypeInfo1"/> 
                 </void> 
                </object> 
               </void> 
@@ -1219,7 +1219,7 @@
                  <string>2</string> 
                 </void> 
                 <void property="type"> 
-                 <object idref="PrimitiveTypeInfo0"/> 
+                 <object idref="PrimitiveTypeInfo1"/> 
                 </void> 
                </object> 
               </void> 
@@ -1281,7 +1281,7 @@
     <void property="pathToAliases"> 
      <object class="java.util.LinkedHashMap"> 
       <void method="put"> 
-       <string>file:/home/njain/workspace/hadoophive/trunk/build/ql/test/data/warehouse/src</string> 
+       <string>file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/src</string> 
        <object class="java.util.ArrayList"> 
         <void method="add"> 
          <string>src</string> 
@@ -1293,7 +1293,7 @@
     <void property="pathToPartitionInfo"> 
      <object class="java.util.LinkedHashMap"> 
       <void method="put"> 
-       <string>file:/home/njain/workspace/hadoophive/trunk/build/ql/test/data/warehouse/src</string> 
+       <string>file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/src</string> 
        <object class="org.apache.hadoop.hive.ql.plan.partitionDesc"> 
         <void property="partSpec"> 
          <object class="java.util.LinkedHashMap"/> 
@@ -1345,7 +1345,7 @@
             </void> 
             <void method="put"> 
              <string>location</string> 
-             <string>file:/home/njain/workspace/hadoophive/trunk/build/ql/test/data/warehouse/src</string> 
+             <string>file:/data/users/zshao/sync/apache-trunk-HIVE-104/build/ql/test/data/warehouse/src</string> 
             </void> 
            </object> 
           </void> 

Modified: hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/thrift/TCTLSeparatedProtocol.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/thrift/TCTLSeparatedProtocol.java?rev=727337&r1=727336&r2=727337&view=diff
==============================================================================
--- hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/thrift/TCTLSeparatedProtocol.java (original)
+++ hadoop/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/thrift/TCTLSeparatedProtocol.java Wed Dec 17 02:35:44 2008
@@ -266,11 +266,11 @@
    */
 
   public TCTLSeparatedProtocol(TTransport trans) {
-    this(trans, defaultPrimarySeparator, defaultSecondarySeparator, defaultMapSeparator, defaultRowSeparator, false, 4096);
+    this(trans, defaultPrimarySeparator, defaultSecondarySeparator, defaultMapSeparator, defaultRowSeparator, true, 4096);
   }
 
   public TCTLSeparatedProtocol(TTransport trans, int buffer_size) {
-    this(trans, defaultPrimarySeparator, defaultSecondarySeparator, defaultMapSeparator, defaultRowSeparator, false, buffer_size);
+    this(trans, defaultPrimarySeparator, defaultSecondarySeparator, defaultMapSeparator, defaultRowSeparator, true, buffer_size);
   }
 
   /**
@@ -287,8 +287,7 @@
                                int bufferSize) {
     super(trans);
 
-    returnNulls = returnNulls;
-
+    this.returnNulls = returnNulls;
 
     this.primarySeparator = primarySeparator;
     this.secondarySeparator = secondarySeparator;
@@ -702,31 +701,56 @@
   public byte readByte() throws TException {
     String val = readString();
     lastPrimitiveWasNullFlag = val == null;
-    return  val == null || val.isEmpty() ? 0 : Byte.valueOf(val).byteValue();
+    try {
+      return val == null || val.isEmpty() ? 0 : Byte.valueOf(val).byteValue();
+    } catch (NumberFormatException e) {
+      lastPrimitiveWasNullFlag = true;
+      return 0;
+    }
   }
 
   public short readI16() throws TException {
     String val = readString();
     lastPrimitiveWasNullFlag = val == null;
-    return val == null || val.isEmpty() ? 0 : Short.valueOf(val).shortValue();
+    try {
+      return val == null || val.isEmpty() ? 0 : Short.valueOf(val).shortValue();
+    } catch (NumberFormatException e) {
+      lastPrimitiveWasNullFlag = true;
+      return 0;
+    }
   }
 
   public int readI32() throws TException {
     String val = readString();
     lastPrimitiveWasNullFlag = val == null;
-    return val == null || val.isEmpty() ? 0 : Integer.valueOf(val).intValue();
+    try {
+      return val == null || val.isEmpty() ? 0 : Integer.valueOf(val).intValue();
+    } catch (NumberFormatException e) {
+      lastPrimitiveWasNullFlag = true;
+      return 0;
+    }
   }
 
   public long readI64() throws TException {
     String val = readString();
     lastPrimitiveWasNullFlag = val == null;
-    return val == null || val.isEmpty() ? 0 : Long.valueOf(val).longValue();
+    try {
+      return val == null || val.isEmpty() ? 0 : Long.valueOf(val).longValue();
+    } catch (NumberFormatException e) {
+      lastPrimitiveWasNullFlag = true;
+      return 0;
+    }
   }
 
   public double readDouble() throws TException {
     String val = readString();
     lastPrimitiveWasNullFlag = val == null;
-    return val == null || val.isEmpty() ? 0 :Double.valueOf(val).doubleValue();
+    try {
+      return val == null || val.isEmpty() ? 0 :Double.valueOf(val).doubleValue();
+    } catch (NumberFormatException e) {
+      lastPrimitiveWasNullFlag = true;
+      return 0;
+    }
   }
 
   public String readString() throws TException {



Mime
View raw message