hive-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From br...@apache.org
Subject svn commit: r1621416 [9/11] - in /hive/branches/spark: ./ common/src/java/org/apache/hadoop/hive/common/ common/src/java/org/apache/hadoop/hive/conf/ contrib/src/test/results/clientnegative/ contrib/src/test/results/clientpositive/ hbase-handler/src/te...
Date Sat, 30 Aug 2014 06:44:53 GMT
Modified: hive/branches/spark/ql/src/test/results/clientpositive/parquet_ctas.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/parquet_ctas.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/parquet_ctas.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/parquet_ctas.q.out Sat Aug 30 06:44:46 2014
@@ -39,9 +39,12 @@ POSTHOOK: Lineage: staging.value SIMPLE 
 PREHOOK: query: create table parquet_ctas stored as parquet as select * from staging
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@staging
+PREHOOK: Output: database:default
+PREHOOK: Output: default@parquet_ctas
 POSTHOOK: query: create table parquet_ctas stored as parquet as select * from staging
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@staging
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@parquet_ctas
 PREHOOK: query: describe parquet_ctas
 PREHOOK: type: DESCTABLE
@@ -72,9 +75,12 @@ POSTHOOK: Input: default@parquet_ctas
 PREHOOK: query: create table parquet_ctas_advanced stored as parquet as select key+1,concat(value,"value") from staging
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@staging
+PREHOOK: Output: database:default
+PREHOOK: Output: default@parquet_ctas_advanced
 POSTHOOK: query: create table parquet_ctas_advanced stored as parquet as select key+1,concat(value,"value") from staging
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@staging
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@parquet_ctas_advanced
 PREHOOK: query: describe parquet_ctas_advanced
 PREHOOK: type: DESCTABLE
@@ -105,9 +111,12 @@ POSTHOOK: Input: default@parquet_ctas_ad
 PREHOOK: query: create table parquet_ctas_alias stored as parquet as select key+1 as mykey,concat(value,"value") as myvalue from staging
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@staging
+PREHOOK: Output: database:default
+PREHOOK: Output: default@parquet_ctas_alias
 POSTHOOK: query: create table parquet_ctas_alias stored as parquet as select key+1 as mykey,concat(value,"value") as myvalue from staging
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@staging
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@parquet_ctas_alias
 PREHOOK: query: describe parquet_ctas_alias
 PREHOOK: type: DESCTABLE
@@ -138,9 +147,12 @@ POSTHOOK: Input: default@parquet_ctas_al
 PREHOOK: query: create table parquet_ctas_mixed stored as parquet as select key,key+1,concat(value,"value") as myvalue from staging
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@staging
+PREHOOK: Output: database:default
+PREHOOK: Output: default@parquet_ctas_mixed
 POSTHOOK: query: create table parquet_ctas_mixed stored as parquet as select key,key+1,concat(value,"value") as myvalue from staging
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@staging
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@parquet_ctas_mixed
 PREHOOK: query: describe parquet_ctas_mixed
 PREHOOK: type: DESCTABLE

Modified: hive/branches/spark/ql/src/test/results/clientpositive/parquet_join.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/parquet_join.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/parquet_join.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/parquet_join.q.out Sat Aug 30 06:44:46 2014
@@ -39,16 +39,22 @@ POSTHOOK: Lineage: staging.value SIMPLE 
 PREHOOK: query: create table parquet_jointable1 stored as parquet as select * from staging
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@staging
+PREHOOK: Output: database:default
+PREHOOK: Output: default@parquet_jointable1
 POSTHOOK: query: create table parquet_jointable1 stored as parquet as select * from staging
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@staging
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@parquet_jointable1
 PREHOOK: query: create table parquet_jointable2 stored as parquet as select key,key+1,concat(value,"value") as myvalue from staging
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@staging
+PREHOOK: Output: database:default
+PREHOOK: Output: default@parquet_jointable2
 POSTHOOK: query: create table parquet_jointable2 stored as parquet as select key,key+1,concat(value,"value") as myvalue from staging
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@staging
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@parquet_jointable2
 PREHOOK: query: -- MR join
 

Modified: hive/branches/spark/ql/src/test/results/clientpositive/partition_decode_name.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/partition_decode_name.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/partition_decode_name.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/partition_decode_name.q.out Sat Aug 30 06:44:46 2014
@@ -6,6 +6,8 @@ from (select '2011-01-11', '2011-01-11+1
       select '2011-01-11', '2011-01-11+16:18:26' from src tablesample (1 rows) ) s
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@sc
 POSTHOOK: query: create table sc as select * 
 from (select '2011-01-11', '2011-01-11+14:18:26' from src tablesample (1 rows)
       union all 
@@ -14,6 +16,7 @@ from (select '2011-01-11', '2011-01-11+1
       select '2011-01-11', '2011-01-11+16:18:26' from src tablesample (1 rows) ) s
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@sc
 PREHOOK: query: create table sc_part (key string) partitioned by (ts string) stored as rcfile
 PREHOOK: type: CREATETABLE

Modified: hive/branches/spark/ql/src/test/results/clientpositive/partition_special_char.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/partition_special_char.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/partition_special_char.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/partition_special_char.q.out Sat Aug 30 06:44:46 2014
@@ -6,6 +6,8 @@ from (select '2011-01-11', '2011-01-11+1
       select '2011-01-11', '2011-01-11+16:18:26' from src tablesample (1 rows) ) s
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@sc
 POSTHOOK: query: create table sc as select * 
 from (select '2011-01-11', '2011-01-11+14:18:26' from src tablesample (1 rows)
       union all 
@@ -14,6 +16,7 @@ from (select '2011-01-11', '2011-01-11+1
       select '2011-01-11', '2011-01-11+16:18:26' from src tablesample (1 rows) ) s
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@sc
 PREHOOK: query: create table sc_part (key string) partitioned by (ts string) stored as rcfile
 PREHOOK: type: CREATETABLE

Modified: hive/branches/spark/ql/src/test/results/clientpositive/ppd_field_garbage.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/ppd_field_garbage.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/ppd_field_garbage.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/ppd_field_garbage.q.out Sat Aug 30 06:44:46 2014
@@ -11,9 +11,12 @@ POSTHOOK: Output: default@test_issue
 PREHOOK: query: CREATE VIEW v_test_issue AS SELECT fileid, i.user, test_c.user_c.age FROM test_issue LATERAL VIEW explode(infos) info AS i
 PREHOOK: type: CREATEVIEW
 PREHOOK: Input: default@test_issue
+PREHOOK: Output: database:default
+PREHOOK: Output: default@v_test_issue
 POSTHOOK: query: CREATE VIEW v_test_issue AS SELECT fileid, i.user, test_c.user_c.age FROM test_issue LATERAL VIEW explode(infos) info AS i
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Input: default@test_issue
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@v_test_issue
 PREHOOK: query: -- dummy data
 LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE test_issue

Modified: hive/branches/spark/ql/src/test/results/clientpositive/ppd_union_view.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/ppd_union_view.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/ppd_union_view.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/ppd_union_view.q.out Sat Aug 30 06:44:46 2014 differ

Modified: hive/branches/spark/ql/src/test/results/clientpositive/ptf.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/ptf.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/ptf.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/ptf.q.out Sat Aug 30 06:44:46 2014
@@ -772,6 +772,8 @@ from part 
 group by p_mfgr, p_brand
 PREHOOK: type: CREATEVIEW
 PREHOOK: Input: default@part
+PREHOOK: Output: database:default
+PREHOOK: Output: default@mfgr_price_view
 POSTHOOK: query: -- 16. testViewAsTableInputToPTF
 create view IF NOT EXISTS mfgr_price_view as 
 select p_mfgr, p_brand, 
@@ -780,6 +782,7 @@ from part 
 group by p_mfgr, p_brand
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Input: default@part
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@mfgr_price_view
 PREHOOK: query: select p_mfgr, p_brand, s, 
 sum(s) over w1  as s1

Modified: hive/branches/spark/ql/src/test/results/clientpositive/query_result_fileformat.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/query_result_fileformat.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/query_result_fileformat.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/query_result_fileformat.q.out Sat Aug 30 06:44:46 2014
@@ -4,12 +4,15 @@ PREHOOK: query: create table nzhang_test
 http://asdf' value from src limit 1
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@nzhang_test1
 POSTHOOK: query: create table nzhang_test1 stored as sequencefile as select 'key1' as key, 'value
 1
 
 http://asdf' value from src limit 1
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@nzhang_test1
 PREHOOK: query: select * from nzhang_test1
 PREHOOK: type: QUERY

Modified: hive/branches/spark/ql/src/test/results/clientpositive/quotedid_basic.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/quotedid_basic.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/quotedid_basic.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/quotedid_basic.q.out Sat Aug 30 06:44:46 2014
@@ -359,12 +359,15 @@ select `x+1```, `y&y`
 from t4 where `x+1``` < '200'
 PREHOOK: type: CREATEVIEW
 PREHOOK: Input: default@t4
+PREHOOK: Output: database:default
+PREHOOK: Output: default@v1
 POSTHOOK: query: -- view
 create view v1 as 
 select `x+1```, `y&y`
 from t4 where `x+1``` < '200'
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Input: default@t4
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@v1
 PREHOOK: query: select `x+1```, `y&y`, rank() over(partition by `x+1``` order by  `y&y`)
 from v1

Modified: hive/branches/spark/ql/src/test/results/clientpositive/rcfile_createas1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/rcfile_createas1.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/rcfile_createas1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/rcfile_createas1.q.out Sat Aug 30 06:44:46 2014
@@ -105,7 +105,7 @@ STAGE PLANS:
           input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
           output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
           serde name: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
-          name: rcfile_createas1b
+          name: default.rcfile_createas1b
 
   Stage: Stage-2
     Stats-Aggr Operator
@@ -134,6 +134,8 @@ PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@rcfile_createas1a
 PREHOOK: Input: default@rcfile_createas1a@ds=1
 PREHOOK: Input: default@rcfile_createas1a@ds=2
+PREHOOK: Output: database:default
+PREHOOK: Output: default@rcfile_createas1b
 POSTHOOK: query: CREATE TABLE rcfile_createas1b
     STORED AS RCFILE AS 
         SELECT key, value, PMOD(HASH(key), 50) as part
@@ -142,6 +144,7 @@ POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@rcfile_createas1a
 POSTHOOK: Input: default@rcfile_createas1a@ds=1
 POSTHOOK: Input: default@rcfile_createas1a@ds=2
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@rcfile_createas1b
 PREHOOK: query: SELECT SUM(HASH(c)) FROM (
     SELECT TRANSFORM(key, value) USING 'tr \t _' AS (c)

Modified: hive/branches/spark/ql/src/test/results/clientpositive/rcfile_default_format.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/rcfile_default_format.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/rcfile_default_format.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/rcfile_default_format.q.out Sat Aug 30 06:44:46 2014
@@ -39,9 +39,12 @@ Storage Desc Params:	 	 
 PREHOOK: query: CREATE TABLE rcfile_default_format_ctas AS SELECT key,value FROM src
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@rcfile_default_format_ctas
 POSTHOOK: query: CREATE TABLE rcfile_default_format_ctas AS SELECT key,value FROM src
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@rcfile_default_format_ctas
 PREHOOK: query: DESCRIBE FORMATTED rcfile_default_format_ctas
 PREHOOK: type: DESCTABLE
@@ -134,9 +137,12 @@ Storage Desc Params:	 	 
 PREHOOK: query: CREATE TABLE textfile_default_format_ctas AS SELECT key,value FROM rcfile_default_format_ctas
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@rcfile_default_format_ctas
+PREHOOK: Output: database:default
+PREHOOK: Output: default@textfile_default_format_ctas
 POSTHOOK: query: CREATE TABLE textfile_default_format_ctas AS SELECT key,value FROM rcfile_default_format_ctas
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@rcfile_default_format_ctas
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@textfile_default_format_ctas
 PREHOOK: query: DESCRIBE FORMATTED textfile_default_format_ctas
 PREHOOK: type: DESCTABLE
@@ -177,9 +183,12 @@ Storage Desc Params:	 	 
 PREHOOK: query: CREATE TABLE rcfile_default_format_ctas_default_serde AS SELECT key,value FROM rcfile_default_format_ctas
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@rcfile_default_format_ctas
+PREHOOK: Output: database:default
+PREHOOK: Output: default@rcfile_default_format_ctas_default_serde
 POSTHOOK: query: CREATE TABLE rcfile_default_format_ctas_default_serde AS SELECT key,value FROM rcfile_default_format_ctas
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@rcfile_default_format_ctas
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@rcfile_default_format_ctas_default_serde
 PREHOOK: query: DESCRIBE FORMATTED rcfile_default_format_ctas_default_serde
 PREHOOK: type: DESCTABLE
@@ -258,9 +267,12 @@ Storage Desc Params:	 	 
 PREHOOK: query: CREATE TABLE rcfile_ctas_default_serde STORED AS rcfile AS SELECT key,value FROM rcfile_default_format_ctas
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@rcfile_default_format_ctas
+PREHOOK: Output: database:default
+PREHOOK: Output: default@rcfile_ctas_default_serde
 POSTHOOK: query: CREATE TABLE rcfile_ctas_default_serde STORED AS rcfile AS SELECT key,value FROM rcfile_default_format_ctas
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@rcfile_default_format_ctas
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@rcfile_ctas_default_serde
 PREHOOK: query: DESCRIBE FORMATTED rcfile_ctas_default_serde
 PREHOOK: type: DESCTABLE

Modified: hive/branches/spark/ql/src/test/results/clientpositive/sample_islocalmode_hook.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/sample_islocalmode_hook.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/sample_islocalmode_hook.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/sample_islocalmode_hook.q.out Sat Aug 30 06:44:46 2014
@@ -48,19 +48,25 @@ PREHOOK: Input: default@sih_i_part
 PREHOOK: Input: default@sih_i_part@p=1
 PREHOOK: Input: default@sih_i_part@p=2
 PREHOOK: Input: default@sih_i_part@p=3
+PREHOOK: Output: database:default
+PREHOOK: Output: default@sih_src
 POSTHOOK: query: create table sih_src as select key, value from sih_i_part order by key, value
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@sih_i_part
 POSTHOOK: Input: default@sih_i_part@p=1
 POSTHOOK: Input: default@sih_i_part@p=2
 POSTHOOK: Input: default@sih_i_part@p=3
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@sih_src
 PREHOOK: query: create table sih_src2 as select key, value from sih_src order by key, value
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@sih_src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@sih_src2
 POSTHOOK: query: create table sih_src2 as select key, value from sih_src order by key, value
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@sih_src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@sih_src2
 PREHOOK: query: -- Relaxing hive.exec.mode.local.auto.input.files.max=1.
 -- Hadoop20 will not generate more splits than there are files (one).

Modified: hive/branches/spark/ql/src/test/results/clientpositive/semijoin.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/semijoin.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/semijoin.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/semijoin.q.out Sat Aug 30 06:44:46 2014
@@ -1,9 +1,12 @@
 PREHOOK: query: create table t1 as select cast(key as int) key, value from src where key <= 10
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@t1
 POSTHOOK: query: create table t1 as select cast(key as int) key, value from src where key <= 10
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@t1
 PREHOOK: query: select * from t1 sort by key
 PREHOOK: type: QUERY
@@ -27,9 +30,12 @@ POSTHOOK: Input: default@t1
 PREHOOK: query: create table t2 as select cast(2*key as int) key, value from t1
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@t1
+PREHOOK: Output: database:default
+PREHOOK: Output: default@t2
 POSTHOOK: query: create table t2 as select cast(2*key as int) key, value from t1
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@t1
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@t2
 PREHOOK: query: select * from t2 sort by key
 PREHOOK: type: QUERY
@@ -54,10 +60,13 @@ PREHOOK: query: create table t3 as selec
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@t1
 PREHOOK: Input: default@t2
+PREHOOK: Output: database:default
+PREHOOK: Output: default@t3
 POSTHOOK: query: create table t3 as select * from (select * from t1 union all select * from t2) b
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@t1
 POSTHOOK: Input: default@t2
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@t3
 PREHOOK: query: select * from t3 sort by key, value
 PREHOOK: type: QUERY

Modified: hive/branches/spark/ql/src/test/results/clientpositive/serde_regex.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/serde_regex.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/serde_regex.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/serde_regex.q.out Sat Aug 30 06:44:46 2014
@@ -45,7 +45,7 @@ STAGE PLANS:
           serde name: org.apache.hadoop.hive.serde2.RegexSerDe
           serde properties:
             input.regex ([^ ]*) ([^ ]*) ([^ ]*) (-|\[[^\]]*\]) ([^ "]*|"[^"]*") (-|[0-9]*) (-|[0-9]*)(?: ([^ "]*|"[^"]*") ([^ "]*|"[^"]*"))?
-          name: serde_regex
+          name: default.serde_regex
 
 PREHOOK: query: CREATE TABLE serde_regex(
   host STRING,
@@ -160,7 +160,7 @@ STAGE PLANS:
           serde name: org.apache.hadoop.hive.serde2.RegexSerDe
           serde properties:
             input.regex ([^ ]*) ([^ ]*)
-          name: serde_regex1
+          name: default.serde_regex1
 
 PREHOOK: query: CREATE TABLE serde_regex1(
   key decimal(38,18),

Modified: hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_db_table.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_db_table.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_db_table.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_db_table.q.out Sat Aug 30 06:44:46 2014
@@ -17,11 +17,10 @@ tmp_feng
 PREHOOK: query: CREATE TABLE tmp_feng.tmp_showcrt (key string, value int)
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:tmp_feng
-PREHOOK: Output: tmp_feng@tmp_feng.tmp_showcrt
+PREHOOK: Output: tmp_feng@tmp_showcrt
 POSTHOOK: query: CREATE TABLE tmp_feng.tmp_showcrt (key string, value int)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: database:tmp_feng
-POSTHOOK: Output: tmp_feng@tmp_feng.tmp_showcrt
 POSTHOOK: Output: tmp_feng@tmp_showcrt
 PREHOOK: query: USE default
 PREHOOK: type: SWITCHDATABASE

Modified: hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_temp_table.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_temp_table.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_temp_table.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_temp_table.q.out Sat Aug 30 06:44:46 2014
@@ -7,12 +7,11 @@ POSTHOOK: Output: database:tmpdb
 PREHOOK: query: create temporary table tmpdb.tmp1 (c1 string, c2 string)
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:tmpdb
-PREHOOK: Output: tmpdb@tmpdb.tmp1
+PREHOOK: Output: tmpdb@tmp1
 POSTHOOK: query: create temporary table tmpdb.tmp1 (c1 string, c2 string)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: database:tmpdb
 POSTHOOK: Output: tmpdb@tmp1
-POSTHOOK: Output: tmpdb@tmpdb.tmp1
 PREHOOK: query: show create table tmpdb.tmp1
 PREHOOK: type: SHOW_CREATETABLE
 PREHOOK: Input: tmpdb@tmp1

Modified: hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_view.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_view.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_view.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_view.q.out Sat Aug 30 06:44:46 2014
@@ -3,11 +3,14 @@ PREHOOK: query: -- Test SHOW CREATE TABL
 CREATE VIEW tmp_copy_src AS SELECT * FROM src
 PREHOOK: type: CREATEVIEW
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@tmp_copy_src
 POSTHOOK: query: -- Test SHOW CREATE TABLE on a view name.
 
 CREATE VIEW tmp_copy_src AS SELECT * FROM src
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@tmp_copy_src
 PREHOOK: query: SHOW CREATE TABLE tmp_copy_src
 PREHOOK: type: SHOW_CREATETABLE

Modified: hive/branches/spark/ql/src/test/results/clientpositive/skewjoin_noskew.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/skewjoin_noskew.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/skewjoin_noskew.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/skewjoin_noskew.q.out Sat Aug 30 06:44:46 2014
@@ -147,7 +147,7 @@ STAGE PLANS:
           input format: org.apache.hadoop.mapred.TextInputFormat
           output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
           serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-          name: noskew
+          name: default.noskew
 
   Stage: Stage-3
     Stats-Aggr Operator
@@ -155,9 +155,12 @@ STAGE PLANS:
 PREHOOK: query: create table noskew as select a.* from src a join src b on a.key=b.key order by a.key limit 30
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@noskew
 POSTHOOK: query: create table noskew as select a.* from src a join src b on a.key=b.key order by a.key limit 30
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@noskew
 PREHOOK: query: select * from noskew
 PREHOOK: type: QUERY

Modified: hive/branches/spark/ql/src/test/results/clientpositive/smb_mapjoin9.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/smb_mapjoin9.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/smb_mapjoin9.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/smb_mapjoin9.q.out Sat Aug 30 06:44:46 2014
@@ -352,7 +352,7 @@ STAGE PLANS:
           input format: org.apache.hadoop.mapred.TextInputFormat
           output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
           serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-          name: smb_mapjoin9_results
+          name: default.smb_mapjoin9_results
 
   Stage: Stage-3
     Stats-Aggr Operator
@@ -397,6 +397,8 @@ PREHOOK: Input: default@hive_test_smb_bu
 PREHOOK: Input: default@hive_test_smb_bucket1@ds=2010-10-15
 PREHOOK: Input: default@hive_test_smb_bucket2
 PREHOOK: Input: default@hive_test_smb_bucket2@ds=2010-10-15
+PREHOOK: Output: database:default
+PREHOOK: Output: default@smb_mapjoin9_results
 POSTHOOK: query: create table smb_mapjoin9_results as
 SELECT /* + MAPJOIN(b) */ b.key as k1, b.value, b.ds, a.key as k2
 FROM hive_test_smb_bucket1 a JOIN 
@@ -407,6 +409,7 @@ POSTHOOK: Input: default@hive_test_smb_b
 POSTHOOK: Input: default@hive_test_smb_bucket1@ds=2010-10-15
 POSTHOOK: Input: default@hive_test_smb_bucket2
 POSTHOOK: Input: default@hive_test_smb_bucket2@ds=2010-10-15
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@smb_mapjoin9_results
 PREHOOK: query: drop table smb_mapjoin9_results
 PREHOOK: type: DROPTABLE

Modified: hive/branches/spark/ql/src/test/results/clientpositive/stats5.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/stats5.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/stats5.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/stats5.q.out Sat Aug 30 06:44:46 2014
@@ -1,9 +1,12 @@
 PREHOOK: query: create table analyze_src as select * from src
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@analyze_src
 POSTHOOK: query: create table analyze_src as select * from src
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@analyze_src
 PREHOOK: query: explain analyze table analyze_src compute statistics
 PREHOOK: type: QUERY

Modified: hive/branches/spark/ql/src/test/results/clientpositive/stats_counter.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/stats_counter.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/stats_counter.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/stats_counter.q.out Sat Aug 30 06:44:46 2014
@@ -2,10 +2,13 @@ PREHOOK: query: -- by analyze
 create table dummy1 as select * from src
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@dummy1
 POSTHOOK: query: -- by analyze
 create table dummy1 as select * from src
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@dummy1
 PREHOOK: query: analyze table dummy1 compute statistics
 PREHOOK: type: QUERY
@@ -55,10 +58,13 @@ PREHOOK: query: -- by autogather
 create table dummy2 as select * from src
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@dummy2
 POSTHOOK: query: -- by autogather
 create table dummy2 as select * from src
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@dummy2
 PREHOOK: query: desc formatted dummy2
 PREHOOK: type: DESCTABLE

Modified: hive/branches/spark/ql/src/test/results/clientpositive/str_to_map.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/str_to_map.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/str_to_map.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/str_to_map.q.out Sat Aug 30 06:44:46 2014
@@ -187,9 +187,12 @@ POSTHOOK: type: DROPTABLE
 PREHOOK: query: create table tbl_s2m as select 'ABC=CC_333=444' as t from src tablesample (3 rows)
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@tbl_s2m
 POSTHOOK: query: create table tbl_s2m as select 'ABC=CC_333=444' as t from src tablesample (3 rows)
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@tbl_s2m
 PREHOOK: query: select str_to_map(t,'_','=')['333'] from tbl_s2m
 PREHOOK: type: QUERY

Modified: hive/branches/spark/ql/src/test/results/clientpositive/subquery_exists.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/subquery_exists.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/subquery_exists.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/subquery_exists.q.out Sat Aug 30 06:44:46 2014
@@ -125,6 +125,8 @@ where exists
   where b.value = a.value  and a.key = b.key and a.value > 'val_9')
 PREHOOK: type: CREATEVIEW
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@cv1
 POSTHOOK: query: -- view test
 create view cv1 as 
 select * 
@@ -135,6 +137,7 @@ where exists
   where b.value = a.value  and a.key = b.key and a.value > 'val_9')
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@cv1
 PREHOOK: query: select * from cv1
 PREHOOK: type: QUERY

Modified: hive/branches/spark/ql/src/test/results/clientpositive/subquery_exists_having.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/subquery_exists_having.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/subquery_exists_having.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/subquery_exists_having.q.out Sat Aug 30 06:44:46 2014
@@ -320,6 +320,8 @@ having exists
   )
 PREHOOK: type: CREATEVIEW
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@cv1
 POSTHOOK: query: -- view test
 create view cv1 as 
 select b.key, count(*) as c
@@ -332,6 +334,7 @@ having exists
   )
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@cv1
 PREHOOK: query: select * from cv1
 PREHOOK: type: QUERY

Modified: hive/branches/spark/ql/src/test/results/clientpositive/subquery_notin.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/subquery_notin.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/subquery_notin.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/subquery_notin.q.out Sat Aug 30 06:44:46 2014
@@ -1413,22 +1413,28 @@ create view T1_v as 
 select key from src where key <'11'
 PREHOOK: type: CREATEVIEW
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@T1_v
 POSTHOOK: query: -- null check
 create view T1_v as 
 select key from src where key <'11'
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@T1_v
 PREHOOK: query: create view T2_v as 
 select case when key > '104' then null else key end as key from T1_v
 PREHOOK: type: CREATEVIEW
 PREHOOK: Input: default@src
 PREHOOK: Input: default@t1_v
+PREHOOK: Output: database:default
+PREHOOK: Output: default@T2_v
 POSTHOOK: query: create view T2_v as 
 select case when key > '104' then null else key end as key from T1_v
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Input: default@src
 POSTHOOK: Input: default@t1_v
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@T2_v
 Warning: Shuffle Join JOIN[24][tables = [t1_v, sq_1_notin_nullcheck]] in Stage 'Stage-2:MAPRED' is a cross product
 PREHOOK: query: explain

Modified: hive/branches/spark/ql/src/test/results/clientpositive/subquery_views.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/subquery_views.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/subquery_views.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/subquery_views.q.out Sat Aug 30 06:44:46 2014
@@ -10,6 +10,8 @@ where exists
   where b.value = a.value  and a.key = b.key and a.value > 'val_9')
 PREHOOK: type: CREATEVIEW
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@cv1
 POSTHOOK: query: -- SORT_QUERY_RESULTS
 
 -- exists test
@@ -22,6 +24,7 @@ where exists
   where b.value = a.value  and a.key = b.key and a.value > 'val_9')
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@cv1
 PREHOOK: query: select * 
 from cv1 where cv1.key in (select key from cv1 c where c.key > '95')
@@ -51,6 +54,8 @@ where b.key not in
   )
 PREHOOK: type: CREATEVIEW
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@cv2
 POSTHOOK: query: -- not in test
 create view cv2 as 
 select * 
@@ -62,6 +67,7 @@ where b.key not in
   )
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@cv2
 Warning: Shuffle Join JOIN[42][tables = [b, sq_1_notin_nullcheck]] in Stage 'Stage-5:MAPRED' is a cross product
 Warning: Shuffle Join JOIN[18][tables = [b, sq_1_notin_nullcheck]] in Stage 'Stage-1:MAPRED' is a cross product
@@ -416,6 +422,8 @@ group by key, value
 having count(*) in (select count(*) from src s1 where s1.key > '9' group by s1.key )
 PREHOOK: type: CREATEVIEW
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@cv3
 POSTHOOK: query: -- in where + having
 create view cv3 as
 select key, value, count(*) 
@@ -425,6 +433,7 @@ group by key, value
 having count(*) in (select count(*) from src s1 where s1.key > '9' group by s1.key )
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@cv3
 PREHOOK: query: select * from cv3
 PREHOOK: type: QUERY

Modified: hive/branches/spark/ql/src/test/results/clientpositive/symlink_text_input_format.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/symlink_text_input_format.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/symlink_text_input_format.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/symlink_text_input_format.q.out Sat Aug 30 06:44:46 2014
@@ -18,7 +18,7 @@ STAGE PLANS:
           columns: key string, value string
           input format: org.apache.hadoop.hive.ql.io.SymlinkTextInputFormat
           output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-          name: symlink_text_input_format
+          name: default.symlink_text_input_format
 
 PREHOOK: query: CREATE TABLE symlink_text_input_format (key STRING, value STRING) STORED AS INPUTFORMAT 'org.apache.hadoop.hive.ql.io.SymlinkTextInputFormat' OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat'
 PREHOOK: type: CREATETABLE

Modified: hive/branches/spark/ql/src/test/results/clientpositive/temp_table.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/temp_table.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/temp_table.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/temp_table.q.out Sat Aug 30 06:44:46 2014
@@ -59,7 +59,7 @@ STAGE PLANS:
 #### A masked pattern was here ####
           output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
           serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-          name: foo
+          name: default.foo
           isTemporary: true
 
   Stage: Stage-2
@@ -98,9 +98,12 @@ STAGE PLANS:
 PREHOOK: query: CREATE TEMPORARY TABLE foo AS SELECT * FROM src WHERE key % 2 = 0
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@foo
 POSTHOOK: query: CREATE TEMPORARY TABLE foo AS SELECT * FROM src WHERE key % 2 = 0
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@foo
 PREHOOK: query: EXPLAIN CREATE TEMPORARY TABLE bar AS SELECT * FROM src WHERE key % 2 = 1
 PREHOOK: type: CREATETABLE_AS_SELECT
@@ -163,7 +166,7 @@ STAGE PLANS:
 #### A masked pattern was here ####
           output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
           serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-          name: bar
+          name: default.bar
           isTemporary: true
 
   Stage: Stage-2
@@ -202,9 +205,12 @@ STAGE PLANS:
 PREHOOK: query: CREATE TEMPORARY TABLE bar AS SELECT * FROM src WHERE key % 2 = 1
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@bar
 POSTHOOK: query: CREATE TEMPORARY TABLE bar AS SELECT * FROM src WHERE key % 2 = 1
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@bar
 PREHOOK: query: DESCRIBE foo
 PREHOOK: type: DESCTABLE
@@ -452,9 +458,12 @@ POSTHOOK: type: SHOWTABLES
 PREHOOK: query: CREATE TEMPORARY TABLE foo AS SELECT * FROM default.foo
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@foo
+PREHOOK: Output: database:two
+PREHOOK: Output: two@foo
 POSTHOOK: query: CREATE TEMPORARY TABLE foo AS SELECT * FROM default.foo
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@foo
+POSTHOOK: Output: database:two
 POSTHOOK: Output: two@foo
 PREHOOK: query: SHOW TABLES
 PREHOOK: type: SHOWTABLES

Modified: hive/branches/spark/ql/src/test/results/clientpositive/temp_table_gb1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/temp_table_gb1.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/temp_table_gb1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/temp_table_gb1.q.out Sat Aug 30 06:44:46 2014
@@ -11,9 +11,12 @@ POSTHOOK: Output: default@dest_g2
 PREHOOK: query: CREATE TEMPORARY TABLE src_temp AS SELECT * FROM src
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@src_temp
 POSTHOOK: query: CREATE TEMPORARY TABLE src_temp AS SELECT * FROM src
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@src_temp
 PREHOOK: query: FROM src_temp
 INSERT OVERWRITE TABLE dest_g2 SELECT substr(src_temp.key,1,1), count(DISTINCT substr(src_temp.value,5)), concat(substr(src_temp.key,1,1),sum(substr(src_temp.value,5))) GROUP BY substr(src_temp.key,1,1)

Modified: hive/branches/spark/ql/src/test/results/clientpositive/temp_table_join1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/temp_table_join1.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/temp_table_join1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/temp_table_join1.q.out Sat Aug 30 06:44:46 2014
@@ -1,16 +1,22 @@
 PREHOOK: query: CREATE TABLE src_nontemp AS SELECT * FROM src limit 10
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@src_nontemp
 POSTHOOK: query: CREATE TABLE src_nontemp AS SELECT * FROM src limit 10
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@src_nontemp
 PREHOOK: query: CREATE TEMPORARY TABLE src_temp AS SELECT * FROM src limit 10
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@src_temp
 POSTHOOK: query: CREATE TEMPORARY TABLE src_temp AS SELECT * FROM src limit 10
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@src_temp
 PREHOOK: query: -- Non temp table join
 EXPLAIN

Modified: hive/branches/spark/ql/src/test/results/clientpositive/temp_table_names.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/temp_table_names.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/temp_table_names.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/temp_table_names.q.out Sat Aug 30 06:44:46 2014
@@ -1,12 +1,11 @@
 PREHOOK: query: -- Test temp tables with upper/lower case names
 create temporary table Default.Temp_Table_Names (C1 string, c2 string)
 PREHOOK: type: CREATETABLE
-PREHOOK: Output: Default@Default.Temp_Table_Names
+PREHOOK: Output: Default@Temp_Table_Names
 PREHOOK: Output: database:default
 POSTHOOK: query: -- Test temp tables with upper/lower case names
 create temporary table Default.Temp_Table_Names (C1 string, c2 string)
 POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: Default@Default.Temp_Table_Names
 POSTHOOK: Output: Default@Temp_Table_Names
 POSTHOOK: Output: database:default
 PREHOOK: query: show tables 'Temp_Table*'

Modified: hive/branches/spark/ql/src/test/results/clientpositive/temp_table_precedence.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/temp_table_precedence.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/temp_table_precedence.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/temp_table_precedence.q.out Sat Aug 30 06:44:46 2014
@@ -8,13 +8,12 @@ PREHOOK: query: -- Create non-temp table
 create table ttp.tab1 (a1 string, a2 string)
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:ttp
-PREHOOK: Output: ttp@ttp.tab1
+PREHOOK: Output: ttp@tab1
 POSTHOOK: query: -- Create non-temp tables
 create table ttp.tab1 (a1 string, a2 string)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: database:ttp
 POSTHOOK: Output: ttp@tab1
-POSTHOOK: Output: ttp@ttp.tab1
 PREHOOK: query: insert overwrite table ttp.tab1 select * from src where key = 5 limit 5
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
@@ -47,12 +46,11 @@ POSTHOOK: Input: ttp@tab1
 PREHOOK: query: create table ttp.tab2 (b1 string, b2 string)
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:ttp
-PREHOOK: Output: ttp@ttp.tab2
+PREHOOK: Output: ttp@tab2
 POSTHOOK: query: create table ttp.tab2 (b1 string, b2 string)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: database:ttp
 POSTHOOK: Output: ttp@tab2
-POSTHOOK: Output: ttp@ttp.tab2
 PREHOOK: query: insert overwrite table ttp.tab2 select * from src where key = 2 limit 5
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
@@ -84,13 +82,12 @@ PREHOOK: query: -- Now create temp table
 create temporary table ttp.tab1 (c1 int, c2 string)
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:ttp
-PREHOOK: Output: ttp@ttp.tab1
+PREHOOK: Output: ttp@tab1
 POSTHOOK: query: -- Now create temp table with same name
 create temporary table ttp.tab1 (c1 int, c2 string)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: database:ttp
 POSTHOOK: Output: ttp@tab1
-POSTHOOK: Output: ttp@ttp.tab1
 PREHOOK: query: insert overwrite table ttp.tab1 select * from src where key = 0 limit 5
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src

Modified: hive/branches/spark/ql/src/test/results/clientpositive/temp_table_subquery1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/temp_table_subquery1.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/temp_table_subquery1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/temp_table_subquery1.q.out Sat Aug 30 06:44:46 2014
@@ -1,9 +1,12 @@
 PREHOOK: query: create temporary table src_temp as select * from src
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@src_temp
 POSTHOOK: query: create temporary table src_temp as select * from src
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@src_temp
 PREHOOK: query: -- subquery exists
 select *

Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_1.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_1.q.out Sat Aug 30 06:44:46 2014
@@ -2,21 +2,27 @@ PREHOOK: query: create table A as
 select * from src
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@A
 POSTHOOK: query: create table A as
 select * from src
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@A
 PREHOOK: query: create table B as
 select * from src
 limit 10
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@B
 POSTHOOK: query: create table B as
 select * from src
 limit 10
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@B
 Warning: Shuffle Join JOIN[4][tables = [a, b]] in Stage 'Reducer 2' is a cross product
 PREHOOK: query: explain select * from A join B

Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_2.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_2.q.out Sat Aug 30 06:44:46 2014
@@ -2,21 +2,27 @@ PREHOOK: query: create table A as
 select * from src
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@A
 POSTHOOK: query: create table A as
 select * from src
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@A
 PREHOOK: query: create table B as
 select * from src
 limit 10
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@B
 POSTHOOK: query: create table B as
 select * from src
 limit 10
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@B
 Warning: Map Join MAPJOIN[7][bigTable=a] in task 'Map 2' is a cross product
 PREHOOK: query: explain select * from A join B

Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/ctas.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/ctas.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/ctas.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/ctas.q.out Sat Aug 30 06:44:46 2014
@@ -91,7 +91,7 @@ STAGE PLANS:
           input format: org.apache.hadoop.mapred.TextInputFormat
           output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
           serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-          name: nzhang_CTAS1
+          name: default.nzhang_CTAS1
 
   Stage: Stage-3
     Stats-Aggr Operator
@@ -105,9 +105,12 @@ STAGE PLANS:
 PREHOOK: query: create table nzhang_CTAS1 as select key k, value from src sort by k, value limit 10
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@nzhang_CTAS1
 POSTHOOK: query: create table nzhang_CTAS1 as select key k, value from src sort by k, value limit 10
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@nzhang_CTAS1
 PREHOOK: query: select * from nzhang_CTAS1
 PREHOOK: type: QUERY
@@ -236,7 +239,7 @@ STAGE PLANS:
           input format: org.apache.hadoop.mapred.TextInputFormat
           output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
           serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-          name: nzhang_ctas2
+          name: default.nzhang_ctas2
 
   Stage: Stage-3
     Stats-Aggr Operator
@@ -250,9 +253,12 @@ STAGE PLANS:
 PREHOOK: query: create table nzhang_ctas2 as select * from src sort by key, value limit 10
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@nzhang_ctas2
 POSTHOOK: query: create table nzhang_ctas2 as select * from src sort by key, value limit 10
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@nzhang_ctas2
 PREHOOK: query: select * from nzhang_ctas2
 PREHOOK: type: QUERY
@@ -381,7 +387,7 @@ STAGE PLANS:
           input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
           output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
           serde name: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
-          name: nzhang_ctas3
+          name: default.nzhang_ctas3
 
   Stage: Stage-3
     Stats-Aggr Operator
@@ -395,9 +401,12 @@ STAGE PLANS:
 PREHOOK: query: create table nzhang_ctas3 row format serde "org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe" stored as RCFile as select key/2 half_key, concat(value, "_con") conb  from src sort by half_key, conb limit 10
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@nzhang_ctas3
 POSTHOOK: query: create table nzhang_ctas3 row format serde "org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe" stored as RCFile as select key/2 half_key, concat(value, "_con") conb  from src sort by half_key, conb limit 10
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@nzhang_ctas3
 PREHOOK: query: select * from nzhang_ctas3
 PREHOOK: type: QUERY
@@ -592,7 +601,7 @@ STAGE PLANS:
           input format: org.apache.hadoop.mapred.TextInputFormat
           output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
           serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-          name: nzhang_ctas4
+          name: default.nzhang_ctas4
 
   Stage: Stage-3
     Stats-Aggr Operator
@@ -606,9 +615,12 @@ STAGE PLANS:
 PREHOOK: query: create table nzhang_ctas4 row format delimited fields terminated by ',' stored as textfile as select key, value from src sort by key, value limit 10
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@nzhang_ctas4
 POSTHOOK: query: create table nzhang_ctas4 row format delimited fields terminated by ',' stored as textfile as select key, value from src sort by key, value limit 10
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@nzhang_ctas4
 PREHOOK: query: select * from nzhang_ctas4
 PREHOOK: type: QUERY
@@ -854,7 +866,7 @@ STAGE PLANS:
 
           output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
           serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-          name: nzhang_ctas5
+          name: default.nzhang_ctas5
 
   Stage: Stage-3
     Stats-Aggr Operator
@@ -869,9 +881,12 @@ STAGE PLANS:
 PREHOOK: query: create table nzhang_ctas5 row format delimited fields terminated by ',' lines terminated by '\012' stored as textfile as select key, value from src sort by key, value limit 10
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@nzhang_ctas5
 POSTHOOK: query: create table nzhang_ctas5 row format delimited fields terminated by ',' lines terminated by '\012' stored as textfile as select key, value from src sort by key, value limit 10
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@nzhang_ctas5
 PREHOOK: query: create table nzhang_ctas6 (key string, `to` string)
 PREHOOK: type: CREATETABLE
@@ -894,7 +909,10 @@ POSTHOOK: Lineage: nzhang_ctas6.to SIMPL
 PREHOOK: query: create table nzhang_ctas7 as select key, `to` from nzhang_ctas6
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@nzhang_ctas6
+PREHOOK: Output: database:default
+PREHOOK: Output: default@nzhang_ctas7
 POSTHOOK: query: create table nzhang_ctas7 as select key, `to` from nzhang_ctas6
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@nzhang_ctas6
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@nzhang_ctas7

Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out Sat Aug 30 06:44:46 2014
@@ -14,6 +14,7 @@ PREHOOK: query: create table over1k(
        fields terminated by '|'
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:default
+PREHOOK: Output: default@over1k
 POSTHOOK: query: create table over1k(
            t tinyint,
            si smallint,
@@ -42,6 +43,7 @@ POSTHOOK: Output: default@over1k
 PREHOOK: query: create table over1k_orc like over1k
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:default
+PREHOOK: Output: default@over1k_orc
 POSTHOOK: query: create table over1k_orc like over1k
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: database:default
@@ -81,6 +83,7 @@ PREHOOK: query: create table over1k_part
        partitioned by (ds string, t tinyint) stored as orc
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:default
+PREHOOK: Output: default@over1k_part_orc
 POSTHOOK: query: create table over1k_part_orc(
            si smallint,
            i int,
@@ -93,6 +96,7 @@ POSTHOOK: Output: default@over1k_part_or
 PREHOOK: query: create table over1k_part_limit_orc like over1k_part_orc
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:default
+PREHOOK: Output: default@over1k_part_limit_orc
 POSTHOOK: query: create table over1k_part_limit_orc like over1k_part_orc
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: database:default
@@ -114,6 +118,7 @@ PREHOOK: query: create table over1k_part
        clustered by (si) into 4 buckets stored as orc
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:default
+PREHOOK: Output: default@over1k_part_buck_orc
 POSTHOOK: query: create table over1k_part_buck_orc(
            si smallint,
            i int,
@@ -134,6 +139,7 @@ PREHOOK: query: create table over1k_part
        sorted by (f) into 4 buckets stored as orc
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:default
+PREHOOK: Output: default@over1k_part_buck_sort_orc
 POSTHOOK: query: create table over1k_part_buck_sort_orc(
            si smallint,
            i int,
@@ -195,6 +201,7 @@ STAGE PLANS:
                   Map-reduce partition columns: _col4 (type: tinyint)
                   Statistics: Num rows: 1048 Data size: 310873 Basic stats: COMPLETE Column stats: NONE
                   value expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
+            Execution mode: vectorized
         Reducer 3 
             Reduce Operator Tree:
               Extract
@@ -280,6 +287,7 @@ STAGE PLANS:
                     Map-reduce partition columns: _col4 (type: tinyint)
                     Statistics: Num rows: 10 Data size: 2960 Basic stats: COMPLETE Column stats: NONE
                     value expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
+            Execution mode: vectorized
         Reducer 3 
             Reduce Operator Tree:
               Extract
@@ -564,6 +572,7 @@ STAGE PLANS:
                   Map-reduce partition columns: _col4 (type: tinyint)
                   Statistics: Num rows: 1048 Data size: 310873 Basic stats: COMPLETE Column stats: NONE
                   value expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
+            Execution mode: vectorized
         Reducer 3 
             Reduce Operator Tree:
               Extract
@@ -649,6 +658,7 @@ STAGE PLANS:
                     Map-reduce partition columns: _col4 (type: tinyint)
                     Statistics: Num rows: 10 Data size: 2960 Basic stats: COMPLETE Column stats: NONE
                     value expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
+            Execution mode: vectorized
         Reducer 3 
             Reduce Operator Tree:
               Extract
@@ -1292,6 +1302,7 @@ create table over1k_part2_orc(
        partitioned by (ds string, t tinyint)
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:default
+PREHOOK: Output: default@over1k_part2_orc
 POSTHOOK: query: -- tests for HIVE-6883
 create table over1k_part2_orc(
            si smallint,
@@ -1419,6 +1430,7 @@ STAGE PLANS:
                   Map-reduce partition columns: _col4 (type: tinyint)
                   Statistics: Num rows: 1048 Data size: 310873 Basic stats: COMPLETE Column stats: NONE
                   value expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint)
+            Execution mode: vectorized
         Reducer 3 
             Reduce Operator Tree:
               Extract
@@ -1762,6 +1774,7 @@ create table over1k_part_buck_sort2_orc(
        sorted by (f) into 1 buckets
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:default
+PREHOOK: Output: default@over1k_part_buck_sort2_orc
 POSTHOOK: query: -- hadoop-1 does not honor number of reducers in local mode. There is always only 1 reducer irrespective of the number of buckets.
 -- Hence all records go to one bucket and all other buckets will be empty. Similar to HIVE-6867. However, hadoop-2 honors number
 -- of reducers and records are spread across all reducers. To avoid this inconsistency we will make number of buckets to 1 for this test.

Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/insert1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/insert1.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/insert1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/insert1.q.out Sat Aug 30 06:44:46 2014
@@ -143,12 +143,11 @@ POSTHOOK: Output: database:x
 PREHOOK: query: create table x.insert1(key int, value string) stored as textfile
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:x
-PREHOOK: Output: x@x.insert1
+PREHOOK: Output: x@insert1
 POSTHOOK: query: create table x.insert1(key int, value string) stored as textfile
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: database:x
 POSTHOOK: Output: x@insert1
-POSTHOOK: Output: x@x.insert1
 PREHOOK: query: explain insert into table x.INSERT1 select a.key, a.value from insert2 a WHERE (a.key=-1)
 PREHOOK: type: QUERY
 POSTHOOK: query: explain insert into table x.INSERT1 select a.key, a.value from insert2 a WHERE (a.key=-1)
@@ -406,11 +405,10 @@ POSTHOOK: Output: database:db1
 PREHOOK: query: CREATE TABLE db1.result(col1 STRING)
 PREHOOK: type: CREATETABLE
 PREHOOK: Output: database:db1
-PREHOOK: Output: db1@db1.result
+PREHOOK: Output: db1@result
 POSTHOOK: query: CREATE TABLE db1.result(col1 STRING)
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: database:db1
-POSTHOOK: Output: db1@db1.result
 POSTHOOK: Output: db1@result
 PREHOOK: query: INSERT OVERWRITE TABLE db1.result SELECT 'db1_insert1' FROM src LIMIT 1
 PREHOOK: type: QUERY

Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/orc_analyze.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/orc_analyze.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/orc_analyze.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/orc_analyze.q.out Sat Aug 30 06:44:46 2014
@@ -105,7 +105,7 @@ Table Parameters:	 	 
 	numFiles            	1                   
 	numRows             	100                 
 	rawDataSize         	52600               
-	totalSize           	3042                
+	totalSize           	3098                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -195,7 +195,7 @@ Table Parameters:	 	 
 	numFiles            	1                   
 	numRows             	100                 
 	rawDataSize         	52600               
-	totalSize           	3042                
+	totalSize           	3098                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -309,7 +309,7 @@ Partition Parameters:	 	 
 	numFiles            	1                   
 	numRows             	50                  
 	rawDataSize         	21950               
-	totalSize           	1962                
+	totalSize           	2016                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -354,7 +354,7 @@ Partition Parameters:	 	 
 	numFiles            	1                   
 	numRows             	50                  
 	rawDataSize         	22050               
-	totalSize           	1981                
+	totalSize           	2036                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -456,7 +456,7 @@ Partition Parameters:	 	 
 	numFiles            	1                   
 	numRows             	50                  
 	rawDataSize         	21950               
-	totalSize           	1962                
+	totalSize           	2016                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -501,7 +501,7 @@ Partition Parameters:	 	 
 	numFiles            	1                   
 	numRows             	50                  
 	rawDataSize         	22050               
-	totalSize           	1981                
+	totalSize           	2036                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -621,7 +621,7 @@ Partition Parameters:	 	 
 	numFiles            	4                   
 	numRows             	50                  
 	rawDataSize         	21980               
-	totalSize           	4746                
+	totalSize           	4955                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -666,7 +666,7 @@ Partition Parameters:	 	 
 	numFiles            	4                   
 	numRows             	50                  
 	rawDataSize         	22048               
-	totalSize           	4829                
+	totalSize           	5046                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -774,7 +774,7 @@ Partition Parameters:	 	 
 	numFiles            	4                   
 	numRows             	50                  
 	rawDataSize         	21980               
-	totalSize           	4746                
+	totalSize           	4955                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -819,7 +819,7 @@ Partition Parameters:	 	 
 	numFiles            	4                   
 	numRows             	50                  
 	rawDataSize         	22048               
-	totalSize           	4829                
+	totalSize           	5046                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 
@@ -984,7 +984,7 @@ Partition Parameters:	 	 
 	numFiles            	1                   
 	numRows             	50                  
 	rawDataSize         	21950               
-	totalSize           	1962                
+	totalSize           	2016                
 #### A masked pattern was here ####
 	 	 
 # Storage Information	 	 

Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/ptf.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/ptf.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/ptf.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/ptf.q.out Sat Aug 30 06:44:46 2014
@@ -772,6 +772,8 @@ from part 
 group by p_mfgr, p_brand
 PREHOOK: type: CREATEVIEW
 PREHOOK: Input: default@part
+PREHOOK: Output: database:default
+PREHOOK: Output: default@mfgr_price_view
 POSTHOOK: query: -- 16. testViewAsTableInputToPTF
 create view IF NOT EXISTS mfgr_price_view as 
 select p_mfgr, p_brand, 
@@ -780,6 +782,7 @@ from part 
 group by p_mfgr, p_brand
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Input: default@part
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@mfgr_price_view
 PREHOOK: query: select p_mfgr, p_brand, s, 
 sum(s) over w1  as s1

Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/stats_counter.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/stats_counter.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/stats_counter.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/stats_counter.q.out Sat Aug 30 06:44:46 2014
@@ -2,10 +2,13 @@ PREHOOK: query: -- by analyze
 create table dummy1 as select * from src
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@dummy1
 POSTHOOK: query: -- by analyze
 create table dummy1 as select * from src
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@dummy1
 PREHOOK: query: analyze table dummy1 compute statistics
 PREHOOK: type: QUERY
@@ -55,10 +58,13 @@ PREHOOK: query: -- by autogather
 create table dummy2 as select * from src
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@dummy2
 POSTHOOK: query: -- by autogather
 create table dummy2 as select * from src
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@dummy2
 PREHOOK: query: desc formatted dummy2
 PREHOOK: type: DESCTABLE

Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/subquery_exists.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/subquery_exists.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/subquery_exists.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/subquery_exists.q.out Sat Aug 30 06:44:46 2014
@@ -133,6 +133,8 @@ where exists
   where b.value = a.value  and a.key = b.key and a.value > 'val_9')
 PREHOOK: type: CREATEVIEW
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@cv1
 POSTHOOK: query: -- view test
 create view cv1 as 
 select * 
@@ -143,6 +145,7 @@ where exists
   where b.value = a.value  and a.key = b.key and a.value > 'val_9')
 POSTHOOK: type: CREATEVIEW
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@cv1
 PREHOOK: query: select * from cv1
 PREHOOK: type: QUERY

Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/temp_table.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/temp_table.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/temp_table.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/temp_table.q.out Sat Aug 30 06:44:46 2014
@@ -46,7 +46,7 @@ STAGE PLANS:
 #### A masked pattern was here ####
           output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
           serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-          name: foo
+          name: default.foo
           isTemporary: true
 
   Stage: Stage-3
@@ -61,9 +61,12 @@ STAGE PLANS:
 PREHOOK: query: CREATE TEMPORARY TABLE foo AS SELECT * FROM src WHERE key % 2 = 0
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@foo
 POSTHOOK: query: CREATE TEMPORARY TABLE foo AS SELECT * FROM src WHERE key % 2 = 0
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@foo
 PREHOOK: query: EXPLAIN CREATE TEMPORARY TABLE bar AS SELECT * FROM src WHERE key % 2 = 1
 PREHOOK: type: CREATETABLE_AS_SELECT
@@ -113,7 +116,7 @@ STAGE PLANS:
 #### A masked pattern was here ####
           output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
           serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-          name: bar
+          name: default.bar
           isTemporary: true
 
   Stage: Stage-3
@@ -128,9 +131,12 @@ STAGE PLANS:
 PREHOOK: query: CREATE TEMPORARY TABLE bar AS SELECT * FROM src WHERE key % 2 = 1
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@bar
 POSTHOOK: query: CREATE TEMPORARY TABLE bar AS SELECT * FROM src WHERE key % 2 = 1
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@bar
 PREHOOK: query: DESCRIBE foo
 PREHOOK: type: DESCTABLE
@@ -378,9 +384,12 @@ POSTHOOK: type: SHOWTABLES
 PREHOOK: query: CREATE TEMPORARY TABLE foo AS SELECT * FROM default.foo
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@foo
+PREHOOK: Output: database:two
+PREHOOK: Output: two@foo
 POSTHOOK: query: CREATE TEMPORARY TABLE foo AS SELECT * FROM default.foo
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@foo
+POSTHOOK: Output: database:two
 POSTHOOK: Output: two@foo
 PREHOOK: query: SHOW TABLES
 PREHOOK: type: SHOWTABLES

Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/tez_dml.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/tez_dml.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/tez_dml.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/tez_dml.q.out Sat Aug 30 06:44:46 2014
@@ -82,7 +82,7 @@ STAGE PLANS:
           input format: org.apache.hadoop.mapred.TextInputFormat
           output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
           serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-          name: tmp_src
+          name: default.tmp_src
 
   Stage: Stage-3
     Stats-Aggr Operator
@@ -96,9 +96,12 @@ STAGE PLANS:
 PREHOOK: query: CREATE TABLE tmp_src AS SELECT * FROM (SELECT value, count(value) AS cnt FROM src GROUP BY value) f1 ORDER BY cnt
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@tmp_src
 POSTHOOK: query: CREATE TABLE tmp_src AS SELECT * FROM (SELECT value, count(value) AS cnt FROM src GROUP BY value) f1 ORDER BY cnt
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@tmp_src
 PREHOOK: query: SELECT * FROM tmp_src
 PREHOOK: type: QUERY
@@ -1478,10 +1481,13 @@ PREHOOK: query: -- create empty table
 CREATE TABLE empty STORED AS orc AS SELECT * FROM tmp_src_part WHERE d = -1000
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@tmp_src_part
+PREHOOK: Output: database:default
+PREHOOK: Output: default@empty
 POSTHOOK: query: -- create empty table
 CREATE TABLE empty STORED AS orc AS SELECT * FROM tmp_src_part WHERE d = -1000
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@tmp_src_part
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@empty
 PREHOOK: query: SELECT * FROM empty
 PREHOOK: type: QUERY

Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/tez_union.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/tez_union.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/tez_union.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/tez_union.q.out Sat Aug 30 06:44:46 2014
@@ -92,12 +92,15 @@ UNION  ALL  
 select s2.key as key, s2.value as value from src s2
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@ut
 POSTHOOK: query: create table ut as
 select s1.key as key, s1.value as value from src s1 join src s3 on s1.key=s3.key
 UNION  ALL  
 select s2.key as key, s2.value as value from src s2
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@ut
 PREHOOK: query: select * from ut order by key, value limit 20
 PREHOOK: type: QUERY
@@ -268,12 +271,15 @@ select count(*) as cnt from (select u1.k
 u as u1 join u as u2 on (u1.key = u2.key)) a
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@ut
 POSTHOOK: query: create table ut as
 with u as (select * from src union all select * from src)
 select count(*) as cnt from (select u1.key as k1, u2.key as k2 from
 u as u1 join u as u2 on (u1.key = u2.key)) a
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@ut
 PREHOOK: query: select * from ut order by cnt limit 20
 PREHOOK: type: QUERY
@@ -401,12 +407,15 @@ src s1
 join (select * from src union all select * from src) u1 on s1.key = u1.key
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@ut
 POSTHOOK: query: create table ut as
 select s1.key as skey, u1.key as ukey from
 src s1
 join (select * from src union all select * from src) u1 on s1.key = u1.key
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@ut
 PREHOOK: query: select * from ut order by skey, ukey limit 20
 PREHOOK: type: QUERY
@@ -708,6 +717,8 @@ join src s8 on (u1.key = s8.key)
 order by lkey
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@ut
 POSTHOOK: query: create table ut as
 select s1.key as skey, u1.key as ukey, s8.key as lkey from 
 src s1
@@ -718,6 +729,7 @@ join src s8 on (u1.key = s8.key)
 order by lkey
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@ut
 PREHOOK: query: select * from ut order by skey, ukey, lkey limit 100
 PREHOOK: type: QUERY
@@ -951,11 +963,14 @@ select s2.key as key from src s2 join sr
 union all select s4.key from src s4 join src s5 on s4.key = s5.key
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@ut
 POSTHOOK: query: create table ut as
 select s2.key as key from src s2 join src s3 on s2.key = s3.key
 union all select s4.key from src s4 join src s5 on s4.key = s5.key
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@ut
 PREHOOK: query: select * from ut order by key limit 30
 PREHOOK: type: QUERY
@@ -1190,12 +1205,15 @@ select u.key as ukey, s.key as skey from
 right outer join src s on u.key = s.key
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@ut
 POSTHOOK: query: create table ut as
 select u.key as ukey, s.key as skey from
 (select * from src union all select * from src) u
 right outer join src s on u.key = s.key
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@ut
 PREHOOK: query: select * from ut order by ukey, skey limit 20
 PREHOOK: type: QUERY

Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/vector_decimal_aggregate.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/vector_decimal_aggregate.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/vector_decimal_aggregate.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/vector_decimal_aggregate.q.out Sat Aug 30 06:44:46 2014
@@ -5,6 +5,8 @@ PREHOOK: query: CREATE TABLE decimal_vgb
 	FROM alltypesorc
 PREHOOK: type: CREATETABLE_AS_SELECT
 PREHOOK: Input: default@alltypesorc
+PREHOOK: Output: database:default
+PREHOOK: Output: default@decimal_vgby
 POSTHOOK: query: CREATE TABLE decimal_vgby STORED AS ORC AS 
 	SELECT cdouble, CAST (((cdouble*22.1)/37) AS DECIMAL(20,10)) AS cdecimal1, 
 	CAST (((cdouble*9.3)/13) AS DECIMAL(23,14)) AS cdecimal2,
@@ -12,6 +14,7 @@ POSTHOOK: query: CREATE TABLE decimal_vg
 	FROM alltypesorc
 POSTHOOK: type: CREATETABLE_AS_SELECT
 POSTHOOK: Input: default@alltypesorc
+POSTHOOK: Output: database:default
 POSTHOOK: Output: default@decimal_vgby
 PREHOOK: query: EXPLAIN SELECT cint,
 	COUNT(cdecimal1), MAX(cdecimal1), MIN(cdecimal1), SUM(cdecimal1), AVG(cdecimal1), STDDEV_POP(cdecimal1), STDDEV_SAMP(cdecimal1),

Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/vectorization_part_project.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/vectorization_part_project.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/tez/vectorization_part_project.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/tez/vectorization_part_project.q.out Sat Aug 30 06:44:46 2014
@@ -65,28 +65,28 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: alltypesorc_part
-                  Statistics: Num rows: 200 Data size: 4068 Basic stats: COMPLETE Column stats: NONE
+                  Statistics: Num rows: 200 Data size: 41576 Basic stats: COMPLETE Column stats: NONE
                   Select Operator
                     expressions: (cdouble + 2) (type: double)
                     outputColumnNames: _col0
-                    Statistics: Num rows: 200 Data size: 4068 Basic stats: COMPLETE Column stats: NONE
+                    Statistics: Num rows: 200 Data size: 41576 Basic stats: COMPLETE Column stats: NONE
                     Reduce Output Operator
                       key expressions: _col0 (type: double)
                       sort order: +
-                      Statistics: Num rows: 200 Data size: 4068 Basic stats: COMPLETE Column stats: NONE
+                      Statistics: Num rows: 200 Data size: 41576 Basic stats: COMPLETE Column stats: NONE
             Execution mode: vectorized
         Reducer 2 
             Reduce Operator Tree:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: double)
                 outputColumnNames: _col0
-                Statistics: Num rows: 200 Data size: 4068 Basic stats: COMPLETE Column stats: NONE
+                Statistics: Num rows: 200 Data size: 41576 Basic stats: COMPLETE Column stats: NONE
                 Limit
                   Number of rows: 10
-                  Statistics: Num rows: 10 Data size: 200 Basic stats: COMPLETE Column stats: NONE
+                  Statistics: Num rows: 10 Data size: 2070 Basic stats: COMPLETE Column stats: NONE
                   File Output Operator
                     compressed: false
-                    Statistics: Num rows: 10 Data size: 200 Basic stats: COMPLETE Column stats: NONE
+                    Statistics: Num rows: 10 Data size: 2070 Basic stats: COMPLETE Column stats: NONE
                     table:
                         input format: org.apache.hadoop.mapred.TextInputFormat
                         output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/vectorized_ptf.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/vectorized_ptf.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/tez/vectorized_ptf.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/tez/vectorized_ptf.q.out Sat Aug 30 06:44:46 2014 differ



Mime
View raw message