hive-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From rhbut...@apache.org
Subject svn commit: r1573445 [2/3] - in /hive/trunk/ql/src: java/org/apache/hadoop/hive/ql/parse/ test/org/apache/hadoop/hive/ql/parse/ test/queries/clientnegative/ test/queries/clientpositive/ test/results/clientnegative/ test/results/clientpositive/
Date Mon, 03 Mar 2014 05:15:09 GMT
Added: hive/trunk/ql/src/test/results/clientpositive/join_cond_pushdown_unqual3.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/join_cond_pushdown_unqual3.q.out?rev=1573445&view=auto
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/join_cond_pushdown_unqual3.q.out (added)
+++ hive/trunk/ql/src/test/results/clientpositive/join_cond_pushdown_unqual3.q.out Mon Mar
 3 05:15:08 2014
@@ -0,0 +1,435 @@
+PREHOOK: query: DROP TABLE part
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: DROP TABLE part
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: -- data setup
+CREATE TABLE part( 
+    p_partkey INT,
+    p_name STRING,
+    p_mfgr STRING,
+    p_brand STRING,
+    p_type STRING,
+    p_size INT,
+    p_container STRING,
+    p_retailprice DOUBLE,
+    p_comment STRING
+)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+POSTHOOK: query: -- data setup
+CREATE TABLE part( 
+    p_partkey INT,
+    p_name STRING,
+    p_mfgr STRING,
+    p_brand STRING,
+    p_type STRING,
+    p_size INT,
+    p_container STRING,
+    p_retailprice DOUBLE,
+    p_comment STRING
+)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@part
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table
part
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@part
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table
part
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@part
+PREHOOK: query: create table part2( 
+    p2_partkey INT,
+    p2_name STRING,
+    p2_mfgr STRING,
+    p2_brand STRING,
+    p2_type STRING,
+    p2_size INT,
+    p2_container STRING,
+    p2_retailprice DOUBLE,
+    p2_comment STRING
+)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+POSTHOOK: query: create table part2( 
+    p2_partkey INT,
+    p2_name STRING,
+    p2_mfgr STRING,
+    p2_brand STRING,
+    p2_type STRING,
+    p2_size INT,
+    p2_container STRING,
+    p2_retailprice DOUBLE,
+    p2_comment STRING
+)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@part2
+PREHOOK: query: create table part3( 
+    p3_partkey INT,
+    p3_name STRING,
+    p3_mfgr STRING,
+    p3_brand STRING,
+    p3_type STRING,
+    p3_size INT,
+    p3_container STRING,
+    p3_retailprice DOUBLE,
+    p3_comment STRING
+)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+POSTHOOK: query: create table part3( 
+    p3_partkey INT,
+    p3_name STRING,
+    p3_mfgr STRING,
+    p3_brand STRING,
+    p3_type STRING,
+    p3_size INT,
+    p3_container STRING,
+    p3_retailprice DOUBLE,
+    p3_comment STRING
+)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@part3
+PREHOOK: query: explain select *
+from part p1 join part2 p2 join part3 p3 
+where p1.p_name = p2_name and p2_name = p3_name
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select *
+from part p1 join part2 p2 join part3 p3 
+where p1.p_name = p2_name and p2_name = p3_name
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Map Operator Tree:
+          TableScan
+            alias: p3
+            Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+            Reduce Output Operator
+              key expressions: p3_name (type: string)
+              sort order: +
+              Map-reduce partition columns: p3_name (type: string)
+              Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+              value expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr
(type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container
(type: string), p3_retailprice (type: double), p3_comment (type: string)
+          TableScan
+            alias: p2
+            Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+            Reduce Output Operator
+              key expressions: p2_name (type: string)
+              sort order: +
+              Map-reduce partition columns: p2_name (type: string)
+              Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+              value expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr
(type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container
(type: string), p2_retailprice (type: double), p2_comment (type: string)
+          TableScan
+            alias: p1
+            Statistics: Num rows: 5 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+            Reduce Output Operator
+              key expressions: p_name (type: string)
+              sort order: +
+              Map-reduce partition columns: p_name (type: string)
+              Statistics: Num rows: 5 Data size: 3173 Basic stats: COMPLETE Column stats:
NONE
+              value expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type:
string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type:
string), p_retailprice (type: double), p_comment (type: string)
+      Reduce Operator Tree:
+        Join Operator
+          condition map:
+               Inner Join 0 to 1
+               Inner Join 1 to 2
+          condition expressions:
+            0 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+            1 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+            2 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+          outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8,
_col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col22, _col23, _col24,
_col25, _col26, _col27, _col28, _col29, _col30
+          Statistics: Num rows: 11 Data size: 6980 Basic stats: COMPLETE Column stats: NONE
+          Filter Operator
+            predicate: ((_col1 = _col12) and (_col12 = _col23)) (type: boolean)
+            Statistics: Num rows: 2 Data size: 1269 Basic stats: COMPLETE Column stats: NONE
+            Select Operator
+              expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string),
_col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7
(type: double), _col8 (type: string), _col11 (type: int), _col12 (type: string), _col13 (type:
string), _col14 (type: string), _col15 (type: string), _col16 (type: int), _col17 (type: string),
_col18 (type: double), _col19 (type: string), _col22 (type: int), _col23 (type: string), _col24
(type: string), _col25 (type: string), _col26 (type: string), _col27 (type: int), _col28 (type:
string), _col29 (type: double), _col30 (type: string)
+              outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7,
_col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19,
_col20, _col21, _col22, _col23, _col24, _col25, _col26
+              Statistics: Num rows: 2 Data size: 1269 Basic stats: COMPLETE Column stats:
NONE
+              File Output Operator
+                compressed: false
+                Statistics: Num rows: 2 Data size: 1269 Basic stats: COMPLETE Column stats:
NONE
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+
+PREHOOK: query: explain select *
+from part p1 join part2 p2 join part3 p3 
+where p2_name = p1.p_name and p3_name = p2_name
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select *
+from part p1 join part2 p2 join part3 p3 
+where p2_name = p1.p_name and p3_name = p2_name
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Map Operator Tree:
+          TableScan
+            alias: p3
+            Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+            Reduce Output Operator
+              key expressions: p3_name (type: string)
+              sort order: +
+              Map-reduce partition columns: p3_name (type: string)
+              Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+              value expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr
(type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container
(type: string), p3_retailprice (type: double), p3_comment (type: string)
+          TableScan
+            alias: p2
+            Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+            Reduce Output Operator
+              key expressions: p2_name (type: string)
+              sort order: +
+              Map-reduce partition columns: p2_name (type: string)
+              Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+              value expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr
(type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container
(type: string), p2_retailprice (type: double), p2_comment (type: string)
+          TableScan
+            alias: p1
+            Statistics: Num rows: 5 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+            Reduce Output Operator
+              key expressions: p_name (type: string)
+              sort order: +
+              Map-reduce partition columns: p_name (type: string)
+              Statistics: Num rows: 5 Data size: 3173 Basic stats: COMPLETE Column stats:
NONE
+              value expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type:
string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type:
string), p_retailprice (type: double), p_comment (type: string)
+      Reduce Operator Tree:
+        Join Operator
+          condition map:
+               Inner Join 0 to 1
+               Inner Join 1 to 2
+          condition expressions:
+            0 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+            1 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+            2 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+          outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8,
_col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col22, _col23, _col24,
_col25, _col26, _col27, _col28, _col29, _col30
+          Statistics: Num rows: 11 Data size: 6980 Basic stats: COMPLETE Column stats: NONE
+          Filter Operator
+            predicate: ((_col12 = _col1) and (_col23 = _col12)) (type: boolean)
+            Statistics: Num rows: 2 Data size: 1269 Basic stats: COMPLETE Column stats: NONE
+            Select Operator
+              expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string),
_col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7
(type: double), _col8 (type: string), _col11 (type: int), _col12 (type: string), _col13 (type:
string), _col14 (type: string), _col15 (type: string), _col16 (type: int), _col17 (type: string),
_col18 (type: double), _col19 (type: string), _col22 (type: int), _col23 (type: string), _col24
(type: string), _col25 (type: string), _col26 (type: string), _col27 (type: int), _col28 (type:
string), _col29 (type: double), _col30 (type: string)
+              outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7,
_col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19,
_col20, _col21, _col22, _col23, _col24, _col25, _col26
+              Statistics: Num rows: 2 Data size: 1269 Basic stats: COMPLETE Column stats:
NONE
+              File Output Operator
+                compressed: false
+                Statistics: Num rows: 2 Data size: 1269 Basic stats: COMPLETE Column stats:
NONE
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+
+PREHOOK: query: explain select *
+from part p1 join part2 p2 join part3 p3 
+where p2_partkey + p1.p_partkey = p1.p_partkey and p3_name = p2_name
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select *
+from part p1 join part2 p2 join part3 p3 
+where p2_partkey + p1.p_partkey = p1.p_partkey and p3_name = p2_name
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-2 is a root stage
+  Stage-1 depends on stages: Stage-2
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-2
+    Map Reduce
+      Map Operator Tree:
+          TableScan
+            alias: p2
+            Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+            Reduce Output Operator
+              sort order: 
+              Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+              value expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr
(type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container
(type: string), p2_retailprice (type: double), p2_comment (type: string)
+          TableScan
+            alias: p1
+            Statistics: Num rows: 5 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+            Reduce Output Operator
+              sort order: 
+              Statistics: Num rows: 5 Data size: 3173 Basic stats: COMPLETE Column stats:
NONE
+              value expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type:
string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type:
string), p_retailprice (type: double), p_comment (type: string)
+      Reduce Operator Tree:
+        Join Operator
+          condition map:
+               Inner Join 0 to 1
+          condition expressions:
+            0 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+            1 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+          outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8,
_col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19
+          Statistics: Num rows: 5 Data size: 3490 Basic stats: COMPLETE Column stats: NONE
+          Filter Operator
+            predicate: ((_col11 + _col0) = _col0) (type: boolean)
+            Statistics: Num rows: 2 Data size: 1396 Basic stats: COMPLETE Column stats: NONE
+            File Output Operator
+              compressed: false
+              table:
+                  input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+                  output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                  serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
+
+  Stage: Stage-1
+    Map Reduce
+      Map Operator Tree:
+          TableScan
+            alias: p3
+            Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+            Reduce Output Operator
+              key expressions: p3_name (type: string)
+              sort order: +
+              Map-reduce partition columns: p3_name (type: string)
+              Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+              value expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr
(type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container
(type: string), p3_retailprice (type: double), p3_comment (type: string)
+          TableScan
+            Reduce Output Operator
+              key expressions: _col12 (type: string)
+              sort order: +
+              Map-reduce partition columns: _col12 (type: string)
+              Statistics: Num rows: 2 Data size: 1396 Basic stats: COMPLETE Column stats:
NONE
+              value expressions: _col11 (type: int), _col12 (type: string), _col13 (type:
string), _col14 (type: string), _col15 (type: string), _col16 (type: int), _col17 (type: string),
_col18 (type: double), _col19 (type: string), _col0 (type: int), _col1 (type: string), _col2
(type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type:
string), _col7 (type: double), _col8 (type: string)
+      Reduce Operator Tree:
+        Join Operator
+          condition map:
+               Inner Join 0 to 1
+          condition expressions:
+            0 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8} {VALUE._col11} {VALUE._col12} {VALUE._col13} {VALUE._col14}
{VALUE._col15} {VALUE._col16} {VALUE._col17} {VALUE._col18} {VALUE._col19}
+            1 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+          outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8,
_col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col22, _col23, _col24,
_col25, _col26, _col27, _col28, _col29, _col30
+          Statistics: Num rows: 2 Data size: 1535 Basic stats: COMPLETE Column stats: NONE
+          Filter Operator
+            predicate: (((_col0 + _col11) = _col11) and (_col23 = _col1)) (type: boolean)
+            Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+            Select Operator
+              expressions: _col11 (type: int), _col12 (type: string), _col13 (type: string),
_col14 (type: string), _col15 (type: string), _col16 (type: int), _col17 (type: string), _col18
(type: double), _col19 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type:
string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string),
_col7 (type: double), _col8 (type: string), _col22 (type: int), _col23 (type: string), _col24
(type: string), _col25 (type: string), _col26 (type: string), _col27 (type: int), _col28 (type:
string), _col29 (type: double), _col30 (type: string)
+              outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7,
_col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19,
_col20, _col21, _col22, _col23, _col24, _col25, _col26
+              Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+              File Output Operator
+                compressed: false
+                Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+
+PREHOOK: query: explain select *
+from part p1 join part2 p2 join part3 p3 
+where p2_partkey = 1 and p3_name = p2_name
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select *
+from part p1 join part2 p2 join part3 p3 
+where p2_partkey = 1 and p3_name = p2_name
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-2 is a root stage
+  Stage-1 depends on stages: Stage-2
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-2
+    Map Reduce
+      Map Operator Tree:
+          TableScan
+            alias: p2
+            Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+            Filter Operator
+              predicate: (p2_partkey = 1) (type: boolean)
+              Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+              Reduce Output Operator
+                sort order: 
+                Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+                value expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr
(type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container
(type: string), p2_retailprice (type: double), p2_comment (type: string)
+          TableScan
+            alias: p1
+            Statistics: Num rows: 5 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+            Reduce Output Operator
+              sort order: 
+              Statistics: Num rows: 5 Data size: 3173 Basic stats: COMPLETE Column stats:
NONE
+              value expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type:
string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type:
string), p_retailprice (type: double), p_comment (type: string)
+      Reduce Operator Tree:
+        Join Operator
+          condition map:
+               Inner Join 0 to 1
+          condition expressions:
+            0 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+            1 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+          outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8,
_col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19
+          Statistics: Num rows: 5 Data size: 3490 Basic stats: COMPLETE Column stats: NONE
+          File Output Operator
+            compressed: false
+            table:
+                input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+                output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
+
+  Stage: Stage-1
+    Map Reduce
+      Map Operator Tree:
+          TableScan
+            alias: p3
+            Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+            Reduce Output Operator
+              key expressions: p3_name (type: string)
+              sort order: +
+              Map-reduce partition columns: p3_name (type: string)
+              Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+              value expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr
(type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container
(type: string), p3_retailprice (type: double), p3_comment (type: string)
+          TableScan
+            Reduce Output Operator
+              key expressions: _col12 (type: string)
+              sort order: +
+              Map-reduce partition columns: _col12 (type: string)
+              Statistics: Num rows: 5 Data size: 3490 Basic stats: COMPLETE Column stats:
NONE
+              value expressions: _col11 (type: int), _col12 (type: string), _col13 (type:
string), _col14 (type: string), _col15 (type: string), _col16 (type: int), _col17 (type: string),
_col18 (type: double), _col19 (type: string), _col0 (type: int), _col1 (type: string), _col2
(type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type:
string), _col7 (type: double), _col8 (type: string)
+      Reduce Operator Tree:
+        Join Operator
+          condition map:
+               Inner Join 0 to 1
+          condition expressions:
+            0 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8} {VALUE._col11} {VALUE._col12} {VALUE._col13} {VALUE._col14}
{VALUE._col15} {VALUE._col16} {VALUE._col17} {VALUE._col18} {VALUE._col19}
+            1 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+          outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8,
_col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col22, _col23, _col24,
_col25, _col26, _col27, _col28, _col29, _col30
+          Statistics: Num rows: 5 Data size: 3839 Basic stats: COMPLETE Column stats: NONE
+          Filter Operator
+            predicate: ((_col0 = 1) and (_col23 = _col1)) (type: boolean)
+            Statistics: Num rows: 1 Data size: 767 Basic stats: COMPLETE Column stats: NONE
+            Select Operator
+              expressions: _col11 (type: int), _col12 (type: string), _col13 (type: string),
_col14 (type: string), _col15 (type: string), _col16 (type: int), _col17 (type: string), _col18
(type: double), _col19 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type:
string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string),
_col7 (type: double), _col8 (type: string), _col22 (type: int), _col23 (type: string), _col24
(type: string), _col25 (type: string), _col26 (type: string), _col27 (type: int), _col28 (type:
string), _col29 (type: double), _col30 (type: string)
+              outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7,
_col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19,
_col20, _col21, _col22, _col23, _col24, _col25, _col26
+              Statistics: Num rows: 1 Data size: 767 Basic stats: COMPLETE Column stats:
NONE
+              File Output Operator
+                compressed: false
+                Statistics: Num rows: 1 Data size: 767 Basic stats: COMPLETE Column stats:
NONE
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+

Added: hive/trunk/ql/src/test/results/clientpositive/join_cond_pushdown_unqual4.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/join_cond_pushdown_unqual4.q.out?rev=1573445&view=auto
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/join_cond_pushdown_unqual4.q.out (added)
+++ hive/trunk/ql/src/test/results/clientpositive/join_cond_pushdown_unqual4.q.out Mon Mar
 3 05:15:08 2014
@@ -0,0 +1,315 @@
+PREHOOK: query: DROP TABLE part
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: DROP TABLE part
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: -- data setup
+CREATE TABLE part( 
+    p_partkey INT,
+    p_name STRING,
+    p_mfgr STRING,
+    p_brand STRING,
+    p_type STRING,
+    p_size INT,
+    p_container STRING,
+    p_retailprice DOUBLE,
+    p_comment STRING
+)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+POSTHOOK: query: -- data setup
+CREATE TABLE part( 
+    p_partkey INT,
+    p_name STRING,
+    p_mfgr STRING,
+    p_brand STRING,
+    p_type STRING,
+    p_size INT,
+    p_container STRING,
+    p_retailprice DOUBLE,
+    p_comment STRING
+)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@part
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table
part
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@part
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table
part
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@part
+PREHOOK: query: create table part2( 
+    p2_partkey INT,
+    p2_name STRING,
+    p2_mfgr STRING,
+    p2_brand STRING,
+    p2_type STRING,
+    p2_size INT,
+    p2_container STRING,
+    p2_retailprice DOUBLE,
+    p2_comment STRING
+)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+POSTHOOK: query: create table part2( 
+    p2_partkey INT,
+    p2_name STRING,
+    p2_mfgr STRING,
+    p2_brand STRING,
+    p2_type STRING,
+    p2_size INT,
+    p2_container STRING,
+    p2_retailprice DOUBLE,
+    p2_comment STRING
+)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@part2
+PREHOOK: query: create table part3( 
+    p3_partkey INT,
+    p3_name STRING,
+    p3_mfgr STRING,
+    p3_brand STRING,
+    p3_type STRING,
+    p3_size INT,
+    p3_container STRING,
+    p3_retailprice DOUBLE,
+    p3_comment STRING
+)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+POSTHOOK: query: create table part3( 
+    p3_partkey INT,
+    p3_name STRING,
+    p3_mfgr STRING,
+    p3_brand STRING,
+    p3_type STRING,
+    p3_size INT,
+    p3_container STRING,
+    p3_retailprice DOUBLE,
+    p3_comment STRING
+)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@part3
+PREHOOK: query: explain select *
+from part p1 join part2 p2 join part3 p3 on p1.p_name = p2_name join part p4 
+where p2_name = p3_name and p1.p_name = p4.p_name
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select *
+from part p1 join part2 p2 join part3 p3 on p1.p_name = p2_name join part p4 
+where p2_name = p3_name and p1.p_name = p4.p_name
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Map Operator Tree:
+          TableScan
+            alias: p4
+            Statistics: Num rows: 5 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+            Reduce Output Operator
+              key expressions: p_name (type: string)
+              sort order: +
+              Map-reduce partition columns: p_name (type: string)
+              Statistics: Num rows: 5 Data size: 3173 Basic stats: COMPLETE Column stats:
NONE
+              value expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type:
string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type:
string), p_retailprice (type: double), p_comment (type: string)
+          TableScan
+            alias: p1
+            Statistics: Num rows: 5 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+            Reduce Output Operator
+              key expressions: p_name (type: string)
+              sort order: +
+              Map-reduce partition columns: p_name (type: string)
+              Statistics: Num rows: 5 Data size: 3173 Basic stats: COMPLETE Column stats:
NONE
+              value expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type:
string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type:
string), p_retailprice (type: double), p_comment (type: string)
+          TableScan
+            alias: p3
+            Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+            Reduce Output Operator
+              key expressions: p3_name (type: string)
+              sort order: +
+              Map-reduce partition columns: p3_name (type: string)
+              Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+              value expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr
(type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container
(type: string), p3_retailprice (type: double), p3_comment (type: string)
+          TableScan
+            alias: p2
+            Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+            Reduce Output Operator
+              key expressions: p2_name (type: string)
+              sort order: +
+              Map-reduce partition columns: p2_name (type: string)
+              Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+              value expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr
(type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container
(type: string), p2_retailprice (type: double), p2_comment (type: string)
+      Reduce Operator Tree:
+        Join Operator
+          condition map:
+               Inner Join 0 to 1
+               Inner Join 1 to 2
+               Inner Join 0 to 3
+          condition expressions:
+            0 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+            1 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+            2 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+            3 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+          outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8,
_col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col22, _col23, _col24,
_col25, _col26, _col27, _col28, _col29, _col30, _col33, _col34, _col35, _col36, _col37, _col38,
_col39, _col40, _col41
+          Statistics: Num rows: 16 Data size: 10470 Basic stats: COMPLETE Column stats: NONE
+          Filter Operator
+            predicate: ((_col12 = _col23) and (_col1 = _col34)) (type: boolean)
+            Statistics: Num rows: 4 Data size: 2617 Basic stats: COMPLETE Column stats: NONE
+            Select Operator
+              expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string),
_col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7
(type: double), _col8 (type: string), _col11 (type: int), _col12 (type: string), _col13 (type:
string), _col14 (type: string), _col15 (type: string), _col16 (type: int), _col17 (type: string),
_col18 (type: double), _col19 (type: string), _col22 (type: int), _col23 (type: string), _col24
(type: string), _col25 (type: string), _col26 (type: string), _col27 (type: int), _col28 (type:
string), _col29 (type: double), _col30 (type: string), _col33 (type: int), _col34 (type: string),
_col35 (type: string), _col36 (type: string), _col37 (type: string), _col38 (type: int), _col39
(type: string), _col40 (type: double), _col41 (type: string)
+              outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7,
_col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19,
_col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31,
_col32, _col33, _col34, _col35
+              Statistics: Num rows: 4 Data size: 2617 Basic stats: COMPLETE Column stats:
NONE
+              File Output Operator
+                compressed: false
+                Statistics: Num rows: 4 Data size: 2617 Basic stats: COMPLETE Column stats:
NONE
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+
+PREHOOK: query: explain select *
+from part p1 join part2 p2 join part3 p3 on p2_name = p1.p_name join part p4 
+where p2_name = p3_name and p1.p_partkey = p4.p_partkey 
+            and p1.p_partkey = p2_partkey
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select *
+from part p1 join part2 p2 join part3 p3 on p2_name = p1.p_name join part p4 
+where p2_name = p3_name and p1.p_partkey = p4.p_partkey 
+            and p1.p_partkey = p2_partkey
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-3 is a root stage
+  Stage-2 depends on stages: Stage-3
+  Stage-1 depends on stages: Stage-2
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-3
+    Map Reduce
+      Map Operator Tree:
+          TableScan
+            alias: p2
+            Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+            Reduce Output Operator
+              key expressions: p2_name (type: string), p2_partkey (type: int)
+              sort order: ++
+              Map-reduce partition columns: p2_name (type: string), p2_partkey (type: int)
+              Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+              value expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr
(type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container
(type: string), p2_retailprice (type: double), p2_comment (type: string)
+          TableScan
+            alias: p1
+            Statistics: Num rows: 5 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+            Reduce Output Operator
+              key expressions: p_name (type: string), p_partkey (type: int)
+              sort order: ++
+              Map-reduce partition columns: p_name (type: string), p_partkey (type: int)
+              Statistics: Num rows: 5 Data size: 3173 Basic stats: COMPLETE Column stats:
NONE
+              value expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type:
string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type:
string), p_retailprice (type: double), p_comment (type: string)
+      Reduce Operator Tree:
+        Join Operator
+          condition map:
+               Inner Join 0 to 1
+          condition expressions:
+            0 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+            1 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+          outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8,
_col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19
+          Statistics: Num rows: 5 Data size: 3490 Basic stats: COMPLETE Column stats: NONE
+          File Output Operator
+            compressed: false
+            table:
+                input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+                output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
+
+  Stage: Stage-2
+    Map Reduce
+      Map Operator Tree:
+          TableScan
+            alias: p3
+            Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+            Reduce Output Operator
+              key expressions: p3_name (type: string)
+              sort order: +
+              Map-reduce partition columns: p3_name (type: string)
+              Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+              value expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr
(type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container
(type: string), p3_retailprice (type: double), p3_comment (type: string)
+          TableScan
+            Reduce Output Operator
+              key expressions: _col12 (type: string)
+              sort order: +
+              Map-reduce partition columns: _col12 (type: string)
+              Statistics: Num rows: 5 Data size: 3490 Basic stats: COMPLETE Column stats:
NONE
+              value expressions: _col11 (type: int), _col12 (type: string), _col13 (type:
string), _col14 (type: string), _col15 (type: string), _col16 (type: int), _col17 (type: string),
_col18 (type: double), _col19 (type: string), _col0 (type: int), _col1 (type: string), _col2
(type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type:
string), _col7 (type: double), _col8 (type: string)
+      Reduce Operator Tree:
+        Join Operator
+          condition map:
+               Inner Join 0 to 1
+          condition expressions:
+            0 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8} {VALUE._col11} {VALUE._col12} {VALUE._col13} {VALUE._col14}
{VALUE._col15} {VALUE._col16} {VALUE._col17} {VALUE._col18} {VALUE._col19}
+            1 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+          outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8,
_col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col22, _col23, _col24,
_col25, _col26, _col27, _col28, _col29, _col30
+          Statistics: Num rows: 5 Data size: 3839 Basic stats: COMPLETE Column stats: NONE
+          File Output Operator
+            compressed: false
+            table:
+                input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+                output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
+
+  Stage: Stage-1
+    Map Reduce
+      Map Operator Tree:
+          TableScan
+            alias: p4
+            Statistics: Num rows: 5 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+            Reduce Output Operator
+              key expressions: p_partkey (type: int)
+              sort order: +
+              Map-reduce partition columns: p_partkey (type: int)
+              Statistics: Num rows: 5 Data size: 3173 Basic stats: COMPLETE Column stats:
NONE
+              value expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type:
string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type:
string), p_retailprice (type: double), p_comment (type: string)
+          TableScan
+            Reduce Output Operator
+              key expressions: _col11 (type: int)
+              sort order: +
+              Map-reduce partition columns: _col11 (type: int)
+              Statistics: Num rows: 5 Data size: 3839 Basic stats: COMPLETE Column stats:
NONE
+              value expressions: _col22 (type: int), _col23 (type: string), _col24 (type:
string), _col25 (type: string), _col26 (type: string), _col27 (type: int), _col28 (type: string),
_col29 (type: double), _col30 (type: string), _col0 (type: int), _col1 (type: string), _col2
(type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type:
string), _col7 (type: double), _col8 (type: string), _col11 (type: int), _col12 (type: string),
_col13 (type: string), _col14 (type: string), _col15 (type: string), _col16 (type: int), _col17
(type: string), _col18 (type: double), _col19 (type: string)
+      Reduce Operator Tree:
+        Join Operator
+          condition map:
+               Inner Join 0 to 1
+          condition expressions:
+            0 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8} {VALUE._col11} {VALUE._col12} {VALUE._col13} {VALUE._col14}
{VALUE._col15} {VALUE._col16} {VALUE._col17} {VALUE._col18} {VALUE._col19} {VALUE._col22}
{VALUE._col23} {VALUE._col24} {VALUE._col25} {VALUE._col26} {VALUE._col27} {VALUE._col28}
{VALUE._col29} {VALUE._col30}
+            1 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} {VALUE._col4} {VALUE._col5}
{VALUE._col6} {VALUE._col7} {VALUE._col8}
+          outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8,
_col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col22, _col23, _col24,
_col25, _col26, _col27, _col28, _col29, _col30, _col33, _col34, _col35, _col36, _col37, _col38,
_col39, _col40, _col41
+          Statistics: Num rows: 5 Data size: 4222 Basic stats: COMPLETE Column stats: NONE
+          Filter Operator
+            predicate: (((_col12 = _col1) and (_col22 = _col33)) and (_col22 = _col11)) (type:
boolean)
+            Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+            Select Operator
+              expressions: _col22 (type: int), _col23 (type: string), _col24 (type: string),
_col25 (type: string), _col26 (type: string), _col27 (type: int), _col28 (type: string), _col29
(type: double), _col30 (type: string), _col11 (type: int), _col12 (type: string), _col13 (type:
string), _col14 (type: string), _col15 (type: string), _col16 (type: int), _col17 (type: string),
_col18 (type: double), _col19 (type: string), _col0 (type: int), _col1 (type: string), _col2
(type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type:
string), _col7 (type: double), _col8 (type: string), _col33 (type: int), _col34 (type: string),
_col35 (type: string), _col36 (type: string), _col37 (type: string), _col38 (type: int), _col39
(type: string), _col40 (type: double), _col41 (type: string)
+              outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7,
_col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19,
_col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31,
_col32, _col33, _col34, _col35
+              Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+              File Output Operator
+                compressed: false
+                Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+                table:
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+



Mime
View raw message