hadoop-hive-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From rmur...@apache.org
Subject svn commit: r819792 [1/24] - in /hadoop/hive/trunk: ./ common/src/java/org/apache/hadoop/hive/conf/ contrib/src/test/results/clientnegative/ contrib/src/test/results/clientpositive/ data/conf/ ql/src/java/org/apache/hadoop/hive/ql/ ql/src/java/org/apac...
Date Tue, 29 Sep 2009 01:25:30 GMT
Author: rmurthy
Date: Tue Sep 29 01:25:15 2009
New Revision: 819792

URL: http://svn.apache.org/viewvc?rev=819792&view=rev
Log:
HIVE-854. Provide for post-execute hooks.
(Namit Jain via rmurthy)

Added:
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/hooks/PostExecute.java
    hadoop/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/hooks/PostExecutePrinter.java
Modified:
    hadoop/hive/trunk/CHANGES.txt
    hadoop/hive/trunk/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java
    hadoop/hive/trunk/contrib/src/test/results/clientnegative/serde_regex.q.out
    hadoop/hive/trunk/contrib/src/test/results/clientpositive/dboutput.q.out
    hadoop/hive/trunk/contrib/src/test/results/clientpositive/fileformat_base64.q.out
    hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_regex.q.out
    hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_s3.q.out
    hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_typedbytes.q.out
    hadoop/hive/trunk/contrib/src/test/results/clientpositive/udaf_example_avg.q.out
    hadoop/hive/trunk/contrib/src/test/results/clientpositive/udaf_example_group_concat.q.out
    hadoop/hive/trunk/contrib/src/test/results/clientpositive/udf_example_add.q.out
    hadoop/hive/trunk/contrib/src/test/results/clientpositive/udf_example_arraymapstruct.q.out
    hadoop/hive/trunk/contrib/src/test/results/clientpositive/udf_example_format.q.out
    hadoop/hive/trunk/data/conf/hive-site.xml
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/Driver.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/MoveTask.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/hooks/ReadEntity.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/hooks/WriteEntity.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMRFileSink1.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/LoadSemanticAnalyzer.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzerFactory.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/DDLWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/moveWork.java
    hadoop/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/session/SessionState.java
    hadoop/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/hooks/PreExecutePrinter.java
    hadoop/hive/trunk/ql/src/test/results/clientnegative/altern1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/bad_sample_clause.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/clusterbydistributeby.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/clusterbysortby.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/create_insert_outputformat.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/create_udaf_failure.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/create_unknown_genericudf.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/create_unknown_udf_udaf.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/deletejar.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/describe_xpath1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/describe_xpath2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/describe_xpath3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/describe_xpath4.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/drop_native_udf.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/external1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/external2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/fetchtask_ioexception.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/fileformat_void_input.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/invalid_create_tbl1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_rc_seq.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/load_wrong_fileformat_txt_seq.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_insert.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/nopart_load.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/notable_alias3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/orderbysortby.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/script_error.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/show_tablestatus.q.out
    hadoop/hive/trunk/ql/src/test/results/clientnegative/union2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/alter1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/alter2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/alter3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/binarysortable_1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/case_sensitivity.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/cast1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/cluster.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/columnarserde_create_shortcut.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/create_1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/create_escape.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/create_genericudaf.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/create_genericudf.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/create_insert_outputformat.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/create_like.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/create_nested_type.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/create_struct_table.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/create_udaf.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/describe_function.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/describe_xpath.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/drop_udf.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/fileformat_sequencefile.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/fileformat_text.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby10.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby11.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_limit.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_map.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_map_nomap.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_map_skew.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby1_noskew.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_limit.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_map.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_map_skew.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby2_noskew.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby3_map.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby3_map_skew.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby3_noskew.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby4.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby4_map.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby4_map_skew.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby4_noskew.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby5.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby5_map.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby5_map_skew.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby5_noskew.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby6.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby6_map.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby6_map_skew.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby6_noskew.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby7.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby7_map.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby7_map_skew.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby7_noskew.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby8.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby8_map.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby8_map_skew.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby8_noskew.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby9.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby_bigdata.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby_map_ppr.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby_neg_float.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/groupby_ppr.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/implicit_cast1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input0.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input10.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input11.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input11_limit.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input12.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input13.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input14.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input14_limit.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input15.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input16.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input16_cc.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input17.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input18.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input19.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input1_limit.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input20.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input21.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input22.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input23.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input24.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input25.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input26.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input28.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input2_limit.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input30.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input31.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input32.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input33.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input34.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input35.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input36.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input37.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input38.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input39.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input3_limit.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input4.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input40.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input41.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input42.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input4_cb_delim.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input4_limit.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input5.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input6.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input7.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input8.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input9.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_columnarserde.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_dynamicserde.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_lazyserde.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_limit.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_part0.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_part1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_part2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_part3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_part4.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_part5.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_part6.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_part7.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_part8.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_part9.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_testsequencefile.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_testxpath.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_testxpath2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_testxpath3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/input_testxpath4.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl4.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl5.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl6.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl7.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/inputddl8.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/insert1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/insertexternal1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join0.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join10.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join11.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join12.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join13.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join14.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join15.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join16.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join17.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join18.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join19.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join20.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join21.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join22.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join23.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join24.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join25.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join26.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join27.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join28.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join29.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join30.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join31.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join32.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join33.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join34.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join35.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join36.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join37.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join4.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join5.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join6.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join7.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join8.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join9.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join_hive_626.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join_map_ppr.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join_rc.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/join_thrift.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/loadpart1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/loadpart_err.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/louter_join_ppr.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce4.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce5.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce6.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce7.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/mapreduce8.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/merge1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/no_hooks.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/noalias_subq1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/notable_alias2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/null_column.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/nullgroup.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/nullgroup2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/nullgroup3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/nullgroup4.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/nullgroup5.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/nullinput.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/nullscript.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/order.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/order2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/outer_join_ppr.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_clusterby.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_constant_expr.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_gby.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_gby2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_gby_join.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_join.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_join2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_join3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_multi_insert.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_outer_join4.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_random.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_transform.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_udf_case.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/ppd_union.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/query_with_semi.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/quote1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/quote2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/rand_partitionpruner3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/rcfile_bigdata.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/rcfile_columnar.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/regex_col.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/regexp_extract.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/router_join_ppr.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/sample1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/sample2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/sample3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/sample4.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/sample5.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/sample6.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/sample7.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/sample8.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/sample9.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/scriptfile1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/select_as_omitted.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/show_functions.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/show_tables.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/show_tablestatus.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/showparts.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/sort.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/subq.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/subq2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/transform_ppr1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/transform_ppr2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/type_cast_1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf1.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf4.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf5.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf6.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf7.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf8.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf9.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_10_trims.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_abs.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_acos.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_ascii.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_asin.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_bin.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_column_pruning.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_case_thrift.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_coalesce.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_conv.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_cos.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_divider.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_elt.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_hash.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_hex.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_hour_minute_second.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_if.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_instr.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_isnull_isnotnull.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_json.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_length.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_like.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_locate.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lower.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_lpad_rpad.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_parse_url.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_pmod.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_repeat.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_reverse.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_round.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_sin.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_size.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_space.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_split.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_substr.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_testlength.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_testlength2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_unhex.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_unix_timestamp.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/udf_when.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union10.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union11.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union12.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union13.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union14.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union15.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union16.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union17.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union18.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union19.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union2.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union20.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union21.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union3.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union4.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union5.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union6.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union7.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union8.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union9.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union_ppr.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/union_script.q.out
    hadoop/hive/trunk/ql/src/test/results/clientpositive/uniquejoin.q.out
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/case_sensitivity.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/groupby1.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input1.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input2.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input3.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input4.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input5.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input6.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input7.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input9.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/input_testsequencefile.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/join1.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/join2.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/join3.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/sample2.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/sample3.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/sample4.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/sample5.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/sample6.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/sample7.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/subq.q.xml
    hadoop/hive/trunk/ql/src/test/results/compiler/plan/union.q.xml

Modified: hadoop/hive/trunk/CHANGES.txt
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/CHANGES.txt?rev=819792&r1=819791&r2=819792&view=diff
==============================================================================
--- hadoop/hive/trunk/CHANGES.txt (original)
+++ hadoop/hive/trunk/CHANGES.txt Tue Sep 29 01:25:15 2009
@@ -45,6 +45,9 @@
     HIVE-843. Add UDFs for hour, minute and second
     (Zheng Shao via namit)
 
+    HIVE-854. Provide for post-execute hooks.
+    (Namit Jain via rmurthy)
+
   IMPROVEMENTS
 
     HIVE-760. Add version info to META-INF/MANIFEST.MF.

Modified: hadoop/hive/trunk/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java?rev=819792&r1=819791&r2=819792&view=diff
==============================================================================
--- hadoop/hive/trunk/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java (original)
+++ hadoop/hive/trunk/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java Tue Sep 29 01:25:15 2009
@@ -68,6 +68,7 @@
     BYTESPERREDUCER("hive.exec.reducers.bytes.per.reducer", (long)(1000*1000*1000)),
     MAXREDUCERS("hive.exec.reducers.max", 999),
     PREEXECHOOKS("hive.exec.pre.hooks", ""),
+    POSTEXECHOOKS("hive.exec.post.hooks", ""),
 
     // hadoop stuff
     HADOOPBIN("hadoop.bin.path", System.getenv("HADOOP_HOME") + "/bin/hadoop"),

Modified: hadoop/hive/trunk/contrib/src/test/results/clientnegative/serde_regex.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/contrib/src/test/results/clientnegative/serde_regex.q.out?rev=819792&r1=819791&r2=819792&view=diff
==============================================================================
--- hadoop/hive/trunk/contrib/src/test/results/clientnegative/serde_regex.q.out (original)
+++ hadoop/hive/trunk/contrib/src/test/results/clientnegative/serde_regex.q.out Tue Sep 29 01:25:15 2009
@@ -1,5 +1,8 @@
-query: DROP TABLE serde_regex
-query: EXPLAIN
+PREHOOK: query: DROP TABLE serde_regex
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: DROP TABLE serde_regex
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: EXPLAIN
 CREATE TABLE serde_regex(
   host STRING,
   identity STRING,
@@ -16,6 +19,25 @@
   "output.format.string" = "%1$s %2$s %3$s %4$s %5$s %6$s %7$s %8$s %9$s"
 )
 STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+POSTHOOK: query: EXPLAIN
+CREATE TABLE serde_regex(
+  host STRING,
+  identity STRING,
+  user STRING,
+  time STRING,
+  request STRING,
+  status INT,
+  size INT,
+  referer STRING,
+  agent STRING)
+ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.RegexSerDe'
+WITH SERDEPROPERTIES (
+  "input.regex" = "([^ ]*) ([^ ]*) ([^ ]*) (-|\\[[^\\]]*\\]) ([^ \"]*|\"[^\"]*\") (-|[0-9]*) (-|[0-9]*)(?: ([^ \"]*|\"[^\"]*\") ([^ \"]*|\"[^\"]*\"))?",
+  "output.format.string" = "%1$s %2$s %3$s %4$s %5$s %6$s %7$s %8$s %9$s"
+)
+STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
 ABSTRACT SYNTAX TREE:
   (TOK_CREATETABLE serde_regex TOK_LIKETABLE (TOK_TABCOLLIST (TOK_TABCOL host TOK_STRING) (TOK_TABCOL identity TOK_STRING) (TOK_TABCOL user TOK_STRING) (TOK_TABCOL time TOK_STRING) (TOK_TABCOL request TOK_STRING) (TOK_TABCOL status TOK_INT) (TOK_TABCOL size TOK_INT) (TOK_TABCOL referer TOK_STRING) (TOK_TABCOL agent TOK_STRING)) (TOK_TABLESERIALIZER (TOK_SERDENAME 'org.apache.hadoop.hive.contrib.serde2.RegexSerDe' (TOK_TABLEPROPERTIES (TOK_TABLEPROPLIST (TOK_TABLEPROPERTY "input.regex" "([^ ]*) ([^ ]*) ([^ ]*) (-|\\[[^\\]]*\\]) ([^ \"]*|\"[^\"]*\") (-|[0-9]*) (-|[0-9]*)(?: ([^ \"]*|\"[^\"]*\") ([^ \"]*|\"[^\"]*\"))?") (TOK_TABLEPROPERTY "output.format.string" "%1$s %2$s %3$s %4$s %5$s %6$s %7$s %8$s %9$s"))))) TOK_TBLTEXTFILE)
 
@@ -39,7 +61,7 @@
           isExternal: false
 
 
-query: CREATE TABLE serde_regex(
+PREHOOK: query: CREATE TABLE serde_regex(
   host STRING,
   identity STRING,
   user STRING,
@@ -55,5 +77,6 @@
   "output.format.string" = "%1$s %2$s %3$s %4$s %5$s %6$s %7$s %8$s %9$s"
 )
 STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
 FAILED: Error in metadata: org.apache.hadoop.hive.ql.metadata.HiveException: MetaException(message:org.apache.hadoop.hive.serde2.SerDeException org.apache.hadoop.hive.contrib.serde2.RegexSerDe only accepts string columns, but column[5] named status has type int)
 FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask

Modified: hadoop/hive/trunk/contrib/src/test/results/clientpositive/dboutput.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/contrib/src/test/results/clientpositive/dboutput.q.out?rev=819792&r1=819791&r2=819792&view=diff
==============================================================================
--- hadoop/hive/trunk/contrib/src/test/results/clientpositive/dboutput.q.out (original)
+++ hadoop/hive/trunk/contrib/src/test/results/clientpositive/dboutput.q.out Tue Sep 29 01:25:15 2009
@@ -1,7 +1,16 @@
-query: CREATE TEMPORARY FUNCTION dboutput AS 'org.apache.hadoop.hive.contrib.genericudf.example.GenericUDFDBOutput'
-query: DESCRIBE FUNCTION dboutput
+PREHOOK: query: CREATE TEMPORARY FUNCTION dboutput AS 'org.apache.hadoop.hive.contrib.genericudf.example.GenericUDFDBOutput'
+PREHOOK: type: CREATEFUNCTION
+POSTHOOK: query: CREATE TEMPORARY FUNCTION dboutput AS 'org.apache.hadoop.hive.contrib.genericudf.example.GenericUDFDBOutput'
+POSTHOOK: type: CREATEFUNCTION
+PREHOOK: query: DESCRIBE FUNCTION dboutput
+PREHOOK: type: DESCFUNCTION
+POSTHOOK: query: DESCRIBE FUNCTION dboutput
+POSTHOOK: type: DESCFUNCTION
 dboutput(jdbcstring,username,password,preparedstatement,[arguments]) - sends data to a jdbc driver
-query: DESCRIBE FUNCTION EXTENDED dboutput
+PREHOOK: query: DESCRIBE FUNCTION EXTENDED dboutput
+PREHOOK: type: DESCFUNCTION
+POSTHOOK: query: DESCRIBE FUNCTION EXTENDED dboutput
+POSTHOOK: type: DESCFUNCTION
 dboutput(jdbcstring,username,password,preparedstatement,[arguments]) - sends data to a jdbc driver
 argument 0 is the JDBC connection string
 argument 1 is the user name
@@ -9,7 +18,7 @@
 argument 3 is an SQL query to be used in the PreparedStatement
 argument (4-n) The remaining arguments must be primitive and are passed to the PreparedStatement object
 
-query: EXPLAIN FROM src
+PREHOOK: query: EXPLAIN FROM src
 
 SELECT dboutput ( 'jdbc:derby:../build/test_dboutput_db;create=true','','',
 'CREATE TABLE app_info ( kkey VARCHAR(255) NOT NULL, vvalue VARCHAR(255) NOT NULL, UNIQUE(kkey))' ),
@@ -21,6 +30,20 @@
 'INSERT INTO app_info (kkey,vvalue) VALUES (?,?)','20','b')
 
 limit 1
+PREHOOK: type: QUERY
+POSTHOOK: query: EXPLAIN FROM src
+
+SELECT dboutput ( 'jdbc:derby:../build/test_dboutput_db;create=true','','',
+'CREATE TABLE app_info ( kkey VARCHAR(255) NOT NULL, vvalue VARCHAR(255) NOT NULL, UNIQUE(kkey))' ),
+
+dboutput('jdbc:derby:../build/test_dboutput_db','','',
+'INSERT INTO app_info (kkey,vvalue) VALUES (?,?)','20','a'),
+
+dboutput('jdbc:derby:../build/test_dboutput_db','','',
+'INSERT INTO app_info (kkey,vvalue) VALUES (?,?)','20','b')
+
+limit 1
+POSTHOOK: type: QUERY
 ABSTRACT SYNTAX TREE:
   (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION dboutput 'jdbc:derby:../build/test_dboutput_db;create=true' '' '' 'CREATE TABLE app_info ( kkey VARCHAR(255) NOT NULL, vvalue VARCHAR(255) NOT NULL, UNIQUE(kkey))')) (TOK_SELEXPR (TOK_FUNCTION dboutput 'jdbc:derby:../build/test_dboutput_db' '' '' 'INSERT INTO app_info (kkey,vvalue) VALUES (?,?)' '20' 'a')) (TOK_SELEXPR (TOK_FUNCTION dboutput 'jdbc:derby:../build/test_dboutput_db' '' '' 'INSERT INTO app_info (kkey,vvalue) VALUES (?,?)' '20' 'b'))) (TOK_LIMIT 1)))
 
@@ -57,7 +80,22 @@
       limit: 1
 
 
-query: FROM src 
+PREHOOK: query: FROM src 
+
+SELECT dboutput ( 'jdbc:derby:../build/test_dboutput_db;create=true','','',
+'CREATE TABLE app_info ( kkey INTEGER NOT NULL, vvalue VARCHAR(255) NOT NULL, UNIQUE(kkey))' ),
+
+dboutput('jdbc:derby:../build/test_dboutput_db','','',
+'INSERT INTO app_info (kkey,vvalue) VALUES (?,?)','20','a'),
+
+dboutput('jdbc:derby:../build/test_dboutput_db','','',
+'INSERT INTO app_info (kkey,vvalue) VALUES (?,?)','20','b')
+
+limit 1
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/109195708/10000
+POSTHOOK: query: FROM src 
 
 SELECT dboutput ( 'jdbc:derby:../build/test_dboutput_db;create=true','','',
 'CREATE TABLE app_info ( kkey INTEGER NOT NULL, vvalue VARCHAR(255) NOT NULL, UNIQUE(kkey))' ),
@@ -69,15 +107,24 @@
 'INSERT INTO app_info (kkey,vvalue) VALUES (?,?)','20','b')
 
 limit 1
-Input: default/src
-Output: file:/home/ecapriolo/hive-617/hive-trunk/build/ql/tmp/753506919/10000
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/109195708/10000
 0	0	1
-query: EXPLAIN SELECT
+PREHOOK: query: EXPLAIN SELECT
+
+dboutput('jdbc:derby:../build/test_dboutput_db','','',
+'INSERT INTO app_info (kkey,vvalue) VALUES (?,?)',key,value)
+
+FROM src WHERE key < 10
+PREHOOK: type: QUERY
+POSTHOOK: query: EXPLAIN SELECT
 
 dboutput('jdbc:derby:../build/test_dboutput_db','','',
 'INSERT INTO app_info (kkey,vvalue) VALUES (?,?)',key,value)
 
 FROM src WHERE key < 10
+POSTHOOK: type: QUERY
 ABSTRACT SYNTAX TREE:
   (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION dboutput 'jdbc:derby:../build/test_dboutput_db' '' '' 'INSERT INTO app_info (kkey,vvalue) VALUES (?,?)' (TOK_TABLE_OR_COL key) (TOK_TABLE_OR_COL value)))) (TOK_WHERE (< (TOK_TABLE_OR_COL key) 10))))
 
@@ -117,14 +164,24 @@
       limit: -1
 
 
-query: SELECT
+PREHOOK: query: SELECT
+
+dboutput('jdbc:derby:../build/test_dboutput_db','','',
+'INSERT INTO app_info (kkey,vvalue) VALUES (?,?)',key,value)
+
+FROM src WHERE key < 10
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/667407949/10000
+POSTHOOK: query: SELECT
 
 dboutput('jdbc:derby:../build/test_dboutput_db','','',
 'INSERT INTO app_info (kkey,vvalue) VALUES (?,?)',key,value)
 
 FROM src WHERE key < 10
-Input: default/src
-Output: file:/home/ecapriolo/hive-617/hive-trunk/build/ql/tmp/1030277391/10000
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/667407949/10000
 0
 0
 0
@@ -135,6 +192,9 @@
 0
 1
 0
-Deleted file:/home/ecapriolo/hive-617/hive-trunk/build/test_dboutput_db
-Deleted file:/home/ecapriolo/hive-617/hive-trunk/contrib/derby.log
-query: DROP TEMPORARY FUNCTION dboutput
+Deleted file:/data/users/njain/hive5/hive5/build/test_dboutput_db
+Deleted file:/data/users/njain/hive5/hive5/contrib/derby.log
+PREHOOK: query: DROP TEMPORARY FUNCTION dboutput
+PREHOOK: type: DROPFUNCTION
+POSTHOOK: query: DROP TEMPORARY FUNCTION dboutput
+POSTHOOK: type: DROPFUNCTION

Modified: hadoop/hive/trunk/contrib/src/test/results/clientpositive/fileformat_base64.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/contrib/src/test/results/clientpositive/fileformat_base64.q.out?rev=819792&r1=819791&r2=819792&view=diff
==============================================================================
--- hadoop/hive/trunk/contrib/src/test/results/clientpositive/fileformat_base64.q.out (original)
+++ hadoop/hive/trunk/contrib/src/test/results/clientpositive/fileformat_base64.q.out Tue Sep 29 01:25:15 2009
@@ -1,8 +1,17 @@
-query: DROP TABLE base64_test
-query: EXPLAIN
+PREHOOK: query: DROP TABLE base64_test
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: DROP TABLE base64_test
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: EXPLAIN
 CREATE TABLE base64_test(key INT, value STRING) STORED AS
   INPUTFORMAT 'org.apache.hadoop.hive.contrib.fileformat.base64.Base64TextInputFormat'
   OUTPUTFORMAT 'org.apache.hadoop.hive.contrib.fileformat.base64.Base64TextOutputFormat'
+PREHOOK: type: CREATETABLE
+POSTHOOK: query: EXPLAIN
+CREATE TABLE base64_test(key INT, value STRING) STORED AS
+  INPUTFORMAT 'org.apache.hadoop.hive.contrib.fileformat.base64.Base64TextInputFormat'
+  OUTPUTFORMAT 'org.apache.hadoop.hive.contrib.fileformat.base64.Base64TextOutputFormat'
+POSTHOOK: type: CREATETABLE
 ABSTRACT SYNTAX TREE:
   (TOK_CREATETABLE base64_test TOK_LIKETABLE (TOK_TABCOLLIST (TOK_TABCOL key TOK_INT) (TOK_TABCOL value TOK_STRING)) (TOK_TABLEFILEFORMAT 'org.apache.hadoop.hive.contrib.fileformat.base64.Base64TextInputFormat' 'org.apache.hadoop.hive.contrib.fileformat.base64.Base64TextOutputFormat'))
 
@@ -22,22 +31,43 @@
           isExternal: false
 
 
-query: CREATE TABLE base64_test(key INT, value STRING) STORED AS
+PREHOOK: query: CREATE TABLE base64_test(key INT, value STRING) STORED AS
   INPUTFORMAT 'org.apache.hadoop.hive.contrib.fileformat.base64.Base64TextInputFormat'
   OUTPUTFORMAT 'org.apache.hadoop.hive.contrib.fileformat.base64.Base64TextOutputFormat'
-query: DESCRIBE EXTENDED base64_test
+PREHOOK: type: CREATETABLE
+POSTHOOK: query: CREATE TABLE base64_test(key INT, value STRING) STORED AS
+  INPUTFORMAT 'org.apache.hadoop.hive.contrib.fileformat.base64.Base64TextInputFormat'
+  OUTPUTFORMAT 'org.apache.hadoop.hive.contrib.fileformat.base64.Base64TextOutputFormat'
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: default@base64_test
+PREHOOK: query: DESCRIBE EXTENDED base64_test
+PREHOOK: type: DESCTABLE
+POSTHOOK: query: DESCRIBE EXTENDED base64_test
+POSTHOOK: type: DESCTABLE
 key	int	
 value	string	
 	 	 
-Detailed Table Information	Table(tableName:base64_test, dbName:default, owner:zshao, createTime:1248153315, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:int, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/data/users/zshao/tools/639-trunk-apache-hive/build/ql/test/data/warehouse/base64_test, inputFormat:org.apache.hadoop.hive.contrib.fileformat.base64.Base64TextInputFormat, outputFormat:org.apache.hadoop.hive.contrib.fileformat.base64.Base64TextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{})	
-query: FROM src
+Detailed Table Information	Table(tableName:base64_test, dbName:default, owner:njain, createTime:1253817673, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:int, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/data/users/njain/hive5/hive5/build/ql/test/data/warehouse/base64_test, inputFormat:org.apache.hadoop.hive.contrib.fileformat.base64.Base64TextInputFormat, outputFormat:org.apache.hadoop.hive.contrib.fileformat.base64.Base64TextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{})	
+PREHOOK: query: FROM src
+INSERT OVERWRITE TABLE base64_test
+SELECT key, value WHERE key < 10
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@base64_test
+POSTHOOK: query: FROM src
 INSERT OVERWRITE TABLE base64_test
 SELECT key, value WHERE key < 10
-Input: default/src
-Output: default/base64_test
-query: SELECT * FROM base64_test
-Input: default/base64_test
-Output: file:/data/users/zshao/tools/639-trunk-apache-hive/build/ql/tmp/239011143/10000
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@base64_test
+PREHOOK: query: SELECT * FROM base64_test
+PREHOOK: type: QUERY
+PREHOOK: Input: default@base64_test
+PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1384563275/10000
+POSTHOOK: query: SELECT * FROM base64_test
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@base64_test
+POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1384563275/10000
 0	val_0
 4	val_4
 8	val_8
@@ -48,17 +78,32 @@
 2	val_2
 5	val_5
 9	val_9
-query: -- Base64TextInput/OutputFormat supports signature (a prefix to check the validity of
+PREHOOK: query: -- Base64TextInput/OutputFormat supports signature (a prefix to check the validity of
+-- the data). These queries test that prefix capabilities.
+
+FROM src
+INSERT OVERWRITE TABLE base64_test
+SELECT key, value WHERE key < 10
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@base64_test
+POSTHOOK: query: -- Base64TextInput/OutputFormat supports signature (a prefix to check the validity of
 -- the data). These queries test that prefix capabilities.
 
 FROM src
 INSERT OVERWRITE TABLE base64_test
 SELECT key, value WHERE key < 10
-Input: default/src
-Output: default/base64_test
-query: SELECT * FROM base64_test
-Input: default/base64_test
-Output: file:/data/users/zshao/tools/639-trunk-apache-hive/build/ql/tmp/494942552/10000
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@base64_test
+PREHOOK: query: SELECT * FROM base64_test
+PREHOOK: type: QUERY
+PREHOOK: Input: default@base64_test
+PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1705395068/10000
+POSTHOOK: query: SELECT * FROM base64_test
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@base64_test
+POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1705395068/10000
 0	val_0
 4	val_4
 8	val_8
@@ -69,4 +114,8 @@
 2	val_2
 5	val_5
 9	val_9
-query: DROP TABLE base64_test
+PREHOOK: query: DROP TABLE base64_test
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: DROP TABLE base64_test
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Output: default@base64_test

Modified: hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_regex.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_regex.q.out?rev=819792&r1=819791&r2=819792&view=diff
==============================================================================
--- hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_regex.q.out (original)
+++ hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_regex.q.out Tue Sep 29 01:25:15 2009
@@ -1,5 +1,8 @@
-query: DROP TABLE serde_regex
-query: EXPLAIN
+PREHOOK: query: DROP TABLE serde_regex
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: DROP TABLE serde_regex
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: EXPLAIN
 CREATE TABLE serde_regex(
   host STRING,
   identity STRING,
@@ -16,6 +19,25 @@
   "output.format.string" = "%1$s %2$s %3$s %4$s %5$s %6$s %7$s %8$s %9$s"
 )
 STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+POSTHOOK: query: EXPLAIN
+CREATE TABLE serde_regex(
+  host STRING,
+  identity STRING,
+  user STRING,
+  time STRING,
+  request STRING,
+  status STRING,
+  size STRING,
+  referer STRING,
+  agent STRING)
+ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.RegexSerDe'
+WITH SERDEPROPERTIES (
+  "input.regex" = "([^ ]*) ([^ ]*) ([^ ]*) (-|\\[[^\\]]*\\]) ([^ \"]*|\"[^\"]*\") (-|[0-9]*) (-|[0-9]*)(?: ([^ \"]*|\"[^\"]*\") ([^ \"]*|\"[^\"]*\"))?",
+  "output.format.string" = "%1$s %2$s %3$s %4$s %5$s %6$s %7$s %8$s %9$s"
+)
+STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
 ABSTRACT SYNTAX TREE:
   (TOK_CREATETABLE serde_regex TOK_LIKETABLE (TOK_TABCOLLIST (TOK_TABCOL host TOK_STRING) (TOK_TABCOL identity TOK_STRING) (TOK_TABCOL user TOK_STRING) (TOK_TABCOL time TOK_STRING) (TOK_TABCOL request TOK_STRING) (TOK_TABCOL status TOK_STRING) (TOK_TABCOL size TOK_STRING) (TOK_TABCOL referer TOK_STRING) (TOK_TABCOL agent TOK_STRING)) (TOK_TABLESERIALIZER (TOK_SERDENAME 'org.apache.hadoop.hive.contrib.serde2.RegexSerDe' (TOK_TABLEPROPERTIES (TOK_TABLEPROPLIST (TOK_TABLEPROPERTY "input.regex" "([^ ]*) ([^ ]*) ([^ ]*) (-|\\[[^\\]]*\\]) ([^ \"]*|\"[^\"]*\") (-|[0-9]*) (-|[0-9]*)(?: ([^ \"]*|\"[^\"]*\") ([^ \"]*|\"[^\"]*\"))?") (TOK_TABLEPROPERTY "output.format.string" "%1$s %2$s %3$s %4$s %5$s %6$s %7$s %8$s %9$s"))))) TOK_TBLTEXTFILE)
 
@@ -39,7 +61,24 @@
           isExternal: false
 
 
-query: CREATE TABLE serde_regex(
+PREHOOK: query: CREATE TABLE serde_regex(
+  host STRING,
+  identity STRING,
+  user STRING,
+  time STRING,
+  request STRING,
+  status STRING,
+  size STRING,
+  referer STRING,
+  agent STRING)
+ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.RegexSerDe'
+WITH SERDEPROPERTIES (
+  "input.regex" = "([^ ]*) ([^ ]*) ([^ ]*) (-|\\[[^\\]]*\\]) ([^ \"]*|\"[^\"]*\") (-|[0-9]*) (-|[0-9]*)(?: ([^ \"]*|\"[^\"]*\") ([^ \"]*|\"[^\"]*\"))?",
+  "output.format.string" = "%1$s %2$s %3$s %4$s %5$s %6$s %7$s %8$s %9$s"
+)
+STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+POSTHOOK: query: CREATE TABLE serde_regex(
   host STRING,
   identity STRING,
   user STRING,
@@ -55,11 +94,30 @@
   "output.format.string" = "%1$s %2$s %3$s %4$s %5$s %6$s %7$s %8$s %9$s"
 )
 STORED AS TEXTFILE
-query: LOAD DATA LOCAL INPATH "../data/files/apache.access.log" INTO TABLE serde_regex
-query: LOAD DATA LOCAL INPATH "../data/files/apache.access.2.log" INTO TABLE serde_regex
-query: SELECT * FROM serde_regex ORDER BY time
-Input: default/serde_regex
-Output: file:/data/users/njain/hive3/hive3/build/ql/tmp/1069367320/10000
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: default@serde_regex
+PREHOOK: query: LOAD DATA LOCAL INPATH "../data/files/apache.access.log" INTO TABLE serde_regex
+PREHOOK: type: LOAD
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../data/files/apache.access.log" INTO TABLE serde_regex
+POSTHOOK: type: LOAD
+POSTHOOK: Output: default@serde_regex
+PREHOOK: query: LOAD DATA LOCAL INPATH "../data/files/apache.access.2.log" INTO TABLE serde_regex
+PREHOOK: type: LOAD
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../data/files/apache.access.2.log" INTO TABLE serde_regex
+POSTHOOK: type: LOAD
+POSTHOOK: Output: default@serde_regex
+PREHOOK: query: SELECT * FROM serde_regex ORDER BY time
+PREHOOK: type: QUERY
+PREHOOK: Input: default@serde_regex
+PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/569581247/10000
+POSTHOOK: query: SELECT * FROM serde_regex ORDER BY time
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@serde_regex
+POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/569581247/10000
 127.0.0.1	-	frank	[10/Oct/2000:13:55:36 -0700]	"GET /apache_pb.gif HTTP/1.0"	200	2326	NULL	NULL
 127.0.0.1	-	-	[26/May/2009:00:00:00 +0000]	"GET /someurl/?track=Blabla(Main) HTTP/1.1"	200	5864	-	"Mozilla/5.0 (Windows; U; Windows NT 6.0; en-US) AppleWebKit/525.19 (KHTML, like Gecko) Chrome/1.0.154.65 Safari/525.19"
-query: DROP TABLE serde_regex
+PREHOOK: query: DROP TABLE serde_regex
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: DROP TABLE serde_regex
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Output: default@serde_regex

Modified: hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_s3.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_s3.q.out?rev=819792&r1=819791&r2=819792&view=diff
==============================================================================
--- hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_s3.q.out (original)
+++ hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_s3.q.out Tue Sep 29 01:25:15 2009
@@ -1,8 +1,20 @@
-query: DROP TABLE s3log
-query: CREATE TABLE s3log
+PREHOOK: query: DROP TABLE s3log
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: DROP TABLE s3log
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: CREATE TABLE s3log
 ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.s3.S3LogDeserializer'
 STORED AS TEXTFILE
-query: DESCRIBE s3log
+PREHOOK: type: CREATETABLE
+POSTHOOK: query: CREATE TABLE s3log
+ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.s3.S3LogDeserializer'
+STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: default@s3log
+PREHOOK: query: DESCRIBE s3log
+PREHOOK: type: DESCTABLE
+POSTHOOK: query: DESCRIBE s3log
+POSTHOOK: type: DESCTABLE
 bucketowner	string	from deserializer
 bucketname	string	from deserializer
 rdatetime	string	from deserializer
@@ -20,10 +32,19 @@
 turnaroundtime	int	from deserializer
 referer	string	from deserializer
 useragent	string	from deserializer
-query: LOAD DATA LOCAL INPATH '../contrib/data/files/s3.log' INTO TABLE s3log
-query: SELECT a.* FROM s3log a
-Input: default/s3log
-Output: file:/data/users/zshao/tools/deploy-trunk-apache-hive/build/ql/tmp/267321215/10000
+PREHOOK: query: LOAD DATA LOCAL INPATH '../contrib/data/files/s3.log' INTO TABLE s3log
+PREHOOK: type: LOAD
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../contrib/data/files/s3.log' INTO TABLE s3log
+POSTHOOK: type: LOAD
+POSTHOOK: Output: default@s3log
+PREHOOK: query: SELECT a.* FROM s3log a
+PREHOOK: type: QUERY
+PREHOOK: Input: default@s3log
+PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/712726189/10000
+POSTHOOK: query: SELECT a.* FROM s3log a
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@s3log
+POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/712726189/10000
 04ff331638adc13885d6c42059584deabbdeabcd55bf0bee491172a79a87b196	img.zemanta.com	09/Apr/2009:22:00:01 +0000	212.143.99.188	65a011a29cdf8ec533ec3d1ccaae921c	D987234E52141DE7	REST.GET.OBJECT	pixy.gif	GET /pixy.gif?x-id=4560525f-2864-495c-842c-159ede7143f8 HTTP/1.1	200	-	828	828	3	2	http://www.mediafuturist.com/2009/02/marc-andreessen-on-the-charlie-rose-show-talking-about-mobile-technology-video.html	Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US) AppleWebKit/525.19 (KHTML, like Gecko) Chrome/1.0.154.53 Safari/525.19
 04ff331638adc13885d6c42059584deabbdeabcd55bf0bee491172a79a87b196	img.zemanta.com	09/Apr/2009:22:00:01 +0000	74.244.182.35	65a011a29cdf8ec533ec3d1ccaae921c	626EECA20AB12A5C	REST.GET.OBJECT	pixy.gif	GET /pixy.gif HTTP/1.1	200	-	828	828	20	20	http://trueslant.com/	Mozilla/5.0 (Macintosh; U; Intel Mac OS X 10_5_6; en-us) AppleWebKit/528.16 (KHTML, like Gecko) Version/4 Public Beta Safari/528.16
 04ff331638adc13885d6c42059584deabbdeabcd55bf0bee491172a79a87b196	img.zemanta.com	09/Apr/2009:22:00:02 +0000	62.149.175.120	65a011a29cdf8ec533ec3d1ccaae921c	3E93D70E69292C98	REST.GET.OBJECT	pixy.gif	GET /pixy.gif?x-id=9fec752e-2318-4da3-864e-ac5b9e47c4ae HTTP/1.0	200	-	828	828	4	3	-	-
@@ -33,4 +54,8 @@
 04ff331638adc13885d6c42059584deabbdeabcd55bf0bee491172a79a87b196	img.zemanta.com	09/Apr/2009:22:00:06 +0000	68.209.136.23	65a011a29cdf8ec533ec3d1ccaae921c	42C8E7EEF98D46BC	REST.GET.OBJECT	pixy.gif	GET /pixy.gif?x-id=78ad075e-832f-4006-a89e-7b5e2978674e HTTP/1.1	200	-	828	828	5	4	http://continuityblog.wordpress.com/category/ragdoll/	Mozilla/5.0 (Windows; U; []  Windows NT 5.1; en-US; rv:1.9.0.8) Gecko/2009032609 Firefox/3.0.8
 04ff331638adc13885d6c42059584deabbdeabcd55bf0bee491172a79a87b196	img.zemanta.com	09/Apr/2009:22:00:06 +0000	68.209.136.23	65a011a29cdf8ec533ec3d1ccaae921c	E5781B471524E1BB	REST.GET.OBJECT	pixy.gif	GET /pixy.gif?x-id=f596477d-b317-8882-8d14-0b8f168f5e8e HTTP/1.1	200	-	828	828	4	3	http://continuityblog.wordpress.com/category/ragdoll/	Mozilla/5.0 (Windows; U; Win[ s]dows NT 5.1; en-US; rv:1.9.0.8) Gecko/2009032609 Firefox/3.0.8
 04ff331638adc13885d6c42059584deabbdeabcd55bf0bee491172a79a87b196	img.zemanta.com	09/Apr/2009:22:00:07 +0000	190.225.84.114	65a011a29cdf8ec533ec3d1ccaae921c	F4FC3FEAD8C00024	REST.GET.OBJECT	pixy.gif	GET /pixy.gif?x-id=23d25db1-160b-48bb-a932-e7dc1e88c321 HTTP/1.1	304	-	NULL	828	3	NULL	http://www.viamujer.com/2009/03/horoscopo-acuario-abril-mayo-y-junio-2009/	Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; .NET CLR 1.1.4322; .NET CLR 2.0.50727)
-query: DROP TABLE s3log
+PREHOOK: query: DROP TABLE s3log
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: DROP TABLE s3log
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Output: default@s3log

Modified: hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_typedbytes.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_typedbytes.q.out?rev=819792&r1=819791&r2=819792&view=diff
==============================================================================
--- hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_typedbytes.q.out (original)
+++ hadoop/hive/trunk/contrib/src/test/results/clientpositive/serde_typedbytes.q.out Tue Sep 29 01:25:15 2009
@@ -1,6 +1,13 @@
-query: drop table dest1
-query: CREATE TABLE dest1(key INT, value STRING) STORED AS TEXTFILE
-query: EXPLAIN
+PREHOOK: query: drop table dest1
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table dest1
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: CREATE TABLE dest1(key INT, value STRING) STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+POSTHOOK: query: CREATE TABLE dest1(key INT, value STRING) STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: default@dest1
+PREHOOK: query: EXPLAIN
 FROM (
   FROM src
   SELECT TRANSFORM(src.key, src.value) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe' 
@@ -10,6 +17,18 @@
   RECORDREADER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordReader'
 ) tmap
 INSERT OVERWRITE TABLE dest1 SELECT tkey, tvalue
+PREHOOK: type: QUERY
+POSTHOOK: query: EXPLAIN
+FROM (
+  FROM src
+  SELECT TRANSFORM(src.key, src.value) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe' 
+  RECORDWRITER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordWriter'
+  USING '/bin/cat'
+  AS (tkey, tvalue) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe' 
+  RECORDREADER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordReader'
+) tmap
+INSERT OVERWRITE TABLE dest1 SELECT tkey, tvalue
+POSTHOOK: type: QUERY
 ABSTRACT SYNTAX TREE:
   (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL src) value)) (TOK_SERDE (TOK_SERDENAME 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe')) (TOK_RECORDWRITER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordWriter') '/bin/cat' (TOK_SERDE (TOK_SERDENAME 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe')) (TOK_RECORDREADER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordReader') (TOK_ALIASLIST tkey tvalue)))))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB dest1)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL tkey)) (TOK_SELEXPR (TOK_TABLE_OR_COL tvalue)))))
 
@@ -66,10 +85,10 @@
           Move Operator
             files:
                 hdfs directory: true
-                destination: file:/data/users/njain/hive1/hive1/build/ql/tmp/2032306747/10000
+                destination: file:/data/users/njain/hive5/hive5/build/ql/tmp/1381406508/10000
           Map Reduce
             Alias -> Map Operator Tree:
-              file:/data/users/njain/hive1/hive1/build/ql/tmp/90741617/10002 
+              file:/data/users/njain/hive5/hive5/build/ql/tmp/2145629671/10002 
                   Reduce Output Operator
                     sort order: 
                     Map-reduce partition columns:
@@ -103,7 +122,19 @@
               name: dest1
 
 
-query: FROM (
+PREHOOK: query: FROM (
+  FROM src
+  SELECT TRANSFORM(src.key, src.value) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe'
+  RECORDWRITER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordWriter'
+  USING '/bin/cat'
+  AS (tkey, tvalue) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe'
+  RECORDREADER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordReader'
+) tmap
+INSERT OVERWRITE TABLE dest1 SELECT tkey, tvalue
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@dest1
+POSTHOOK: query: FROM (
   FROM src
   SELECT TRANSFORM(src.key, src.value) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe'
   RECORDWRITER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordWriter'
@@ -112,11 +143,17 @@
   RECORDREADER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordReader'
 ) tmap
 INSERT OVERWRITE TABLE dest1 SELECT tkey, tvalue
-Input: default/src
-Output: default/dest1
-query: SELECT dest1.* FROM dest1
-Input: default/dest1
-Output: file:/data/users/njain/hive1/hive1/build/ql/tmp/1184607913/10000
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@dest1
+PREHOOK: query: SELECT dest1.* FROM dest1
+PREHOOK: type: QUERY
+PREHOOK: Input: default@dest1
+PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1901712227/10000
+POSTHOOK: query: SELECT dest1.* FROM dest1
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@dest1
+POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1901712227/10000
 238	val_238
 86	val_86
 311	val_311
@@ -617,4 +654,8 @@
 400	val_400
 200	val_200
 97	val_97
-query: drop table dest1
+PREHOOK: query: drop table dest1
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table dest1
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Output: default@dest1

Modified: hadoop/hive/trunk/contrib/src/test/results/clientpositive/udaf_example_avg.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/contrib/src/test/results/clientpositive/udaf_example_avg.q.out?rev=819792&r1=819791&r2=819792&view=diff
==============================================================================
--- hadoop/hive/trunk/contrib/src/test/results/clientpositive/udaf_example_avg.q.out (original)
+++ hadoop/hive/trunk/contrib/src/test/results/clientpositive/udaf_example_avg.q.out Tue Sep 29 01:25:15 2009
@@ -1,8 +1,17 @@
-query: CREATE TEMPORARY FUNCTION example_avg AS 'org.apache.hadoop.hive.contrib.udaf.example.UDAFExampleAvg'
-query: EXPLAIN
+PREHOOK: query: CREATE TEMPORARY FUNCTION example_avg AS 'org.apache.hadoop.hive.contrib.udaf.example.UDAFExampleAvg'
+PREHOOK: type: CREATEFUNCTION
+POSTHOOK: query: CREATE TEMPORARY FUNCTION example_avg AS 'org.apache.hadoop.hive.contrib.udaf.example.UDAFExampleAvg'
+POSTHOOK: type: CREATEFUNCTION
+PREHOOK: query: EXPLAIN
 SELECT example_avg(substr(value,5)),
        example_avg(IF(substr(value,5) > 250, NULL, substr(value,5)))
 FROM src
+PREHOOK: type: QUERY
+POSTHOOK: query: EXPLAIN
+SELECT example_avg(substr(value,5)),
+       example_avg(IF(substr(value,5) > 250, NULL, substr(value,5)))
+FROM src
+POSTHOOK: type: QUERY
 ABSTRACT SYNTAX TREE:
   (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION example_avg (TOK_FUNCTION substr (TOK_TABLE_OR_COL value) 5))) (TOK_SELEXPR (TOK_FUNCTION example_avg (TOK_FUNCTION IF (> (TOK_FUNCTION substr (TOK_TABLE_OR_COL value) 5) 250) TOK_NULL (TOK_FUNCTION substr (TOK_TABLE_OR_COL value) 5)))))))
 
@@ -62,10 +71,20 @@
       limit: -1
 
 
-query: SELECT example_avg(substr(value,5)),
+PREHOOK: query: SELECT example_avg(substr(value,5)),
+       example_avg(IF(substr(value,5) > 250, NULL, substr(value,5)))
+FROM src
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1148096071/10000
+POSTHOOK: query: SELECT example_avg(substr(value,5)),
        example_avg(IF(substr(value,5) > 250, NULL, substr(value,5)))
 FROM src
-Input: default/src
-Output: file:/data/users/zshao/tools/699-trunk-apache-hive/build/ql/tmp/532982360/10000
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1148096071/10000
 260.182	134.82926829268294
-query: DROP TEMPORARY FUNCTION example_avg
+PREHOOK: query: DROP TEMPORARY FUNCTION example_avg
+PREHOOK: type: DROPFUNCTION
+POSTHOOK: query: DROP TEMPORARY FUNCTION example_avg
+POSTHOOK: type: DROPFUNCTION

Modified: hadoop/hive/trunk/contrib/src/test/results/clientpositive/udaf_example_group_concat.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/contrib/src/test/results/clientpositive/udaf_example_group_concat.q.out?rev=819792&r1=819791&r2=819792&view=diff
==============================================================================
--- hadoop/hive/trunk/contrib/src/test/results/clientpositive/udaf_example_group_concat.q.out (original)
+++ hadoop/hive/trunk/contrib/src/test/results/clientpositive/udaf_example_group_concat.q.out Tue Sep 29 01:25:15 2009
@@ -1,8 +1,17 @@
-query: CREATE TEMPORARY FUNCTION example_group_concat AS 'org.apache.hadoop.hive.contrib.udaf.example.UDAFExampleGroupConcat'
-query: EXPLAIN
+PREHOOK: query: CREATE TEMPORARY FUNCTION example_group_concat AS 'org.apache.hadoop.hive.contrib.udaf.example.UDAFExampleGroupConcat'
+PREHOOK: type: CREATEFUNCTION
+POSTHOOK: query: CREATE TEMPORARY FUNCTION example_group_concat AS 'org.apache.hadoop.hive.contrib.udaf.example.UDAFExampleGroupConcat'
+POSTHOOK: type: CREATEFUNCTION
+PREHOOK: query: EXPLAIN
 SELECT substr(value,5,1), example_group_concat("(", key, ":", value, ")")
 FROM src
 GROUP BY substr(value,5,1)
+PREHOOK: type: QUERY
+POSTHOOK: query: EXPLAIN
+SELECT substr(value,5,1), example_group_concat("(", key, ":", value, ")")
+FROM src
+GROUP BY substr(value,5,1)
+POSTHOOK: type: QUERY
 ABSTRACT SYNTAX TREE:
   (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION substr (TOK_TABLE_OR_COL value) 5 1)) (TOK_SELEXPR (TOK_FUNCTION example_group_concat "(" (TOK_TABLE_OR_COL key) ":" (TOK_TABLE_OR_COL value) ")"))) (TOK_GROUPBY (TOK_FUNCTION substr (TOK_TABLE_OR_COL value) 5 1))))
 
@@ -72,11 +81,18 @@
       limit: -1
 
 
-query: SELECT substr(value,5,1), example_group_concat("(", key, ":", value, ")")
+PREHOOK: query: SELECT substr(value,5,1), example_group_concat("(", key, ":", value, ")")
+FROM src
+GROUP BY substr(value,5,1)
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1156831700/10000
+POSTHOOK: query: SELECT substr(value,5,1), example_group_concat("(", key, ":", value, ")")
 FROM src
 GROUP BY substr(value,5,1)
-Input: default/src
-Output: file:/data/users/zshao/tools/699-trunk-apache-hive/build/ql/tmp/148650144/10000
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1156831700/10000
 0	(0:val_0)(0:val_0)(0:val_0)
 1	(100:val_100)(100:val_100)(103:val_103)(103:val_103)(104:val_104)(104:val_104)(105:val_105)(10:val_10)(111:val_111)(113:val_113)(113:val_113)(114:val_114)(116:val_116)(118:val_118)(118:val_118)(119:val_119)(119:val_119)(119:val_119)(11:val_11)(120:val_120)(120:val_120)(125:val_125)(125:val_125)(126:val_126)(128:val_128)(128:val_128)(128:val_128)(129:val_129)(129:val_129)(12:val_12)(12:val_12)(131:val_131)(133:val_133)(134:val_134)(134:val_134)(136:val_136)(137:val_137)(137:val_137)(138:val_138)(138:val_138)(138:val_138)(138:val_138)(143:val_143)(145:val_145)(146:val_146)(146:val_146)(149:val_149)(149:val_149)(150:val_150)(152:val_152)(152:val_152)(153:val_153)(155:val_155)(156:val_156)(157:val_157)(158:val_158)(15:val_15)(15:val_15)(160:val_160)(162:val_162)(163:val_163)(164:val_164)(164:val_164)(165:val_165)(165:val_165)(166:val_166)(167:val_167)(167:val_167)(167:val_167)(168:val_168)(169:val_169)(169:val_169)(169:val_169)(169:val_169)(170:val_170)(172:val_172)(172:val_17
 2)(174:val_174)(174:val_174)(175:val_175)(175:val_175)(176:val_176)(176:val_176)(177:val_177)(178:val_178)(179:val_179)(179:val_179)(17:val_17)(180:val_180)(181:val_181)(183:val_183)(186:val_186)(187:val_187)(187:val_187)(187:val_187)(189:val_189)(18:val_18)(18:val_18)(190:val_190)(191:val_191)(191:val_191)(192:val_192)(193:val_193)(193:val_193)(193:val_193)(194:val_194)(195:val_195)(195:val_195)(196:val_196)(197:val_197)(197:val_197)(199:val_199)(199:val_199)(199:val_199)(19:val_19)
 2	(200:val_200)(200:val_200)(201:val_201)(202:val_202)(203:val_203)(203:val_203)(205:val_205)(205:val_205)(207:val_207)(207:val_207)(208:val_208)(208:val_208)(208:val_208)(209:val_209)(209:val_209)(20:val_20)(213:val_213)(213:val_213)(214:val_214)(216:val_216)(216:val_216)(217:val_217)(217:val_217)(218:val_218)(219:val_219)(219:val_219)(221:val_221)(221:val_221)(222:val_222)(223:val_223)(223:val_223)(224:val_224)(224:val_224)(226:val_226)(228:val_228)(229:val_229)(229:val_229)(230:val_230)(230:val_230)(230:val_230)(230:val_230)(230:val_230)(233:val_233)(233:val_233)(235:val_235)(237:val_237)(237:val_237)(238:val_238)(238:val_238)(239:val_239)(239:val_239)(241:val_241)(242:val_242)(242:val_242)(244:val_244)(247:val_247)(248:val_248)(249:val_249)(24:val_24)(24:val_24)(252:val_252)(255:val_255)(255:val_255)(256:val_256)(256:val_256)(257:val_257)(258:val_258)(260:val_260)(262:val_262)(263:val_263)(265:val_265)(265:val_265)(266:val_266)(26:val_26)(26:val_26)(272:val_272)(272:val_
 272)(273:val_273)(273:val_273)(273:val_273)(274:val_274)(275:val_275)(277:val_277)(277:val_277)(277:val_277)(277:val_277)(278:val_278)(278:val_278)(27:val_27)(280:val_280)(280:val_280)(281:val_281)(281:val_281)(282:val_282)(282:val_282)(283:val_283)(284:val_284)(285:val_285)(286:val_286)(287:val_287)(288:val_288)(288:val_288)(289:val_289)(28:val_28)(291:val_291)(292:val_292)(296:val_296)(298:val_298)(298:val_298)(298:val_298)(2:val_2)
@@ -87,4 +103,7 @@
 7	(70:val_70)(70:val_70)(70:val_70)(72:val_72)(72:val_72)(74:val_74)(76:val_76)(76:val_76)(77:val_77)(78:val_78)
 8	(80:val_80)(82:val_82)(83:val_83)(83:val_83)(84:val_84)(84:val_84)(85:val_85)(86:val_86)(87:val_87)(8:val_8)
 9	(90:val_90)(90:val_90)(90:val_90)(92:val_92)(95:val_95)(95:val_95)(96:val_96)(97:val_97)(97:val_97)(98:val_98)(98:val_98)(9:val_9)
-query: DROP TEMPORARY FUNCTION example_group_concat
+PREHOOK: query: DROP TEMPORARY FUNCTION example_group_concat
+PREHOOK: type: DROPFUNCTION
+POSTHOOK: query: DROP TEMPORARY FUNCTION example_group_concat
+POSTHOOK: type: DROPFUNCTION

Modified: hadoop/hive/trunk/contrib/src/test/results/clientpositive/udf_example_add.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/contrib/src/test/results/clientpositive/udf_example_add.q.out?rev=819792&r1=819791&r2=819792&view=diff
==============================================================================
--- hadoop/hive/trunk/contrib/src/test/results/clientpositive/udf_example_add.q.out (original)
+++ hadoop/hive/trunk/contrib/src/test/results/clientpositive/udf_example_add.q.out Tue Sep 29 01:25:15 2009
@@ -1,5 +1,8 @@
-query: CREATE TEMPORARY FUNCTION example_add AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleAdd'
-query: EXPLAIN
+PREHOOK: query: CREATE TEMPORARY FUNCTION example_add AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleAdd'
+PREHOOK: type: CREATEFUNCTION
+POSTHOOK: query: CREATE TEMPORARY FUNCTION example_add AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleAdd'
+POSTHOOK: type: CREATEFUNCTION
+PREHOOK: query: EXPLAIN
 SELECT example_add(1, 2),
        example_add(1, 2, 3),
        example_add(1, 2, 3, 4),
@@ -8,6 +11,17 @@
        example_add(1.1, 2.2, 3.3, 4.4),
        example_add(1, 2, 3, 4.4)
 FROM src LIMIT 1
+PREHOOK: type: QUERY
+POSTHOOK: query: EXPLAIN
+SELECT example_add(1, 2),
+       example_add(1, 2, 3),
+       example_add(1, 2, 3, 4),
+       example_add(1.1, 2.2),
+       example_add(1.1, 2.2, 3.3),
+       example_add(1.1, 2.2, 3.3, 4.4),
+       example_add(1, 2, 3, 4.4)
+FROM src LIMIT 1
+POSTHOOK: type: QUERY
 ABSTRACT SYNTAX TREE:
   (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION example_add 1 2)) (TOK_SELEXPR (TOK_FUNCTION example_add 1 2 3)) (TOK_SELEXPR (TOK_FUNCTION example_add 1 2 3 4)) (TOK_SELEXPR (TOK_FUNCTION example_add 1.1 2.2)) (TOK_SELEXPR (TOK_FUNCTION example_add 1.1 2.2 3.3)) (TOK_SELEXPR (TOK_FUNCTION example_add 1.1 2.2 3.3 4.4)) (TOK_SELEXPR (TOK_FUNCTION example_add 1 2 3 4.4))) (TOK_LIMIT 1)))
 
@@ -52,7 +66,18 @@
       limit: 1
 
 
-query: SELECT example_add(1, 2),
+PREHOOK: query: SELECT example_add(1, 2),
+       example_add(1, 2, 3),
+       example_add(1, 2, 3, 4),
+       example_add(1.1, 2.2),
+       example_add(1.1, 2.2, 3.3),
+       example_add(1.1, 2.2, 3.3, 4.4),
+       example_add(1, 2, 3, 4.4)
+FROM src LIMIT 1
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/504154641/10000
+POSTHOOK: query: SELECT example_add(1, 2),
        example_add(1, 2, 3),
        example_add(1, 2, 3, 4),
        example_add(1.1, 2.2),
@@ -60,7 +85,11 @@
        example_add(1.1, 2.2, 3.3, 4.4),
        example_add(1, 2, 3, 4.4)
 FROM src LIMIT 1
-Input: default/src
-Output: file:/data/users/zshao/tools/699-trunk-apache-hive/build/ql/tmp/221019346/10000
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/504154641/10000
 3	6	10	3.3000000000000003	6.6	11.0	10.4
-query: DROP TEMPORARY FUNCTION example_add
+PREHOOK: query: DROP TEMPORARY FUNCTION example_add
+PREHOOK: type: DROPFUNCTION
+POSTHOOK: query: DROP TEMPORARY FUNCTION example_add
+POSTHOOK: type: DROPFUNCTION

Modified: hadoop/hive/trunk/contrib/src/test/results/clientpositive/udf_example_arraymapstruct.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/contrib/src/test/results/clientpositive/udf_example_arraymapstruct.q.out?rev=819792&r1=819791&r2=819792&view=diff
==============================================================================
--- hadoop/hive/trunk/contrib/src/test/results/clientpositive/udf_example_arraymapstruct.q.out (original)
+++ hadoop/hive/trunk/contrib/src/test/results/clientpositive/udf_example_arraymapstruct.q.out Tue Sep 29 01:25:15 2009
@@ -1,9 +1,23 @@
-query: CREATE TEMPORARY FUNCTION example_arraysum    AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleArraySum'
-query: CREATE TEMPORARY FUNCTION example_mapconcat   AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleMapConcat'
-query: CREATE TEMPORARY FUNCTION example_structprint AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleStructPrint'
-query: EXPLAIN
+PREHOOK: query: CREATE TEMPORARY FUNCTION example_arraysum    AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleArraySum'
+PREHOOK: type: CREATEFUNCTION
+POSTHOOK: query: CREATE TEMPORARY FUNCTION example_arraysum    AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleArraySum'
+POSTHOOK: type: CREATEFUNCTION
+PREHOOK: query: CREATE TEMPORARY FUNCTION example_mapconcat   AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleMapConcat'
+PREHOOK: type: CREATEFUNCTION
+POSTHOOK: query: CREATE TEMPORARY FUNCTION example_mapconcat   AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleMapConcat'
+POSTHOOK: type: CREATEFUNCTION
+PREHOOK: query: CREATE TEMPORARY FUNCTION example_structprint AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleStructPrint'
+PREHOOK: type: CREATEFUNCTION
+POSTHOOK: query: CREATE TEMPORARY FUNCTION example_structprint AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleStructPrint'
+POSTHOOK: type: CREATEFUNCTION
+PREHOOK: query: EXPLAIN
 SELECT example_arraysum(lint), example_mapconcat(mstringstring), example_structprint(lintstring[0])
 FROM src_thrift
+PREHOOK: type: QUERY
+POSTHOOK: query: EXPLAIN
+SELECT example_arraysum(lint), example_mapconcat(mstringstring), example_structprint(lintstring[0])
+FROM src_thrift
+POSTHOOK: type: QUERY
 ABSTRACT SYNTAX TREE:
   (TOK_QUERY (TOK_FROM (TOK_TABREF src_thrift)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION example_arraysum (TOK_TABLE_OR_COL lint))) (TOK_SELEXPR (TOK_FUNCTION example_mapconcat (TOK_TABLE_OR_COL mstringstring))) (TOK_SELEXPR (TOK_FUNCTION example_structprint ([ (TOK_TABLE_OR_COL lintstring) 0))))))
 
@@ -39,10 +53,16 @@
       limit: -1
 
 
-query: SELECT example_arraysum(lint), example_mapconcat(mstringstring), example_structprint(lintstring[0])
+PREHOOK: query: SELECT example_arraysum(lint), example_mapconcat(mstringstring), example_structprint(lintstring[0])
+FROM src_thrift
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src_thrift
+PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/571432173/10000
+POSTHOOK: query: SELECT example_arraysum(lint), example_mapconcat(mstringstring), example_structprint(lintstring[0])
 FROM src_thrift
-Input: default/src_thrift
-Output: file:/data/users/zshao/tools/699-trunk-apache-hive/build/ql/tmp/1971586889/10000
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src_thrift
+POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/571432173/10000
 0.0	(key_0:value_0)	(0:0)(1:0)(2:0)
 6.0	(key_1:value_1)	(0:1)(1:1)(2:1)
 12.0	(key_2:value_2)	(0:4)(1:8)(2:2)
@@ -54,6 +74,15 @@
 48.0	(key_8:value_8)	(0:64)(1:512)(2:8)
 54.0	(key_9:value_9)	(0:81)(1:729)(2:9)
 NULL	NULL	NULL
-query: DROP TEMPORARY FUNCTION example_arraysum
-query: DROP TEMPORARY FUNCTION example_mapconcat
-query: DROP TEMPORARY FUNCTION example_structprint
+PREHOOK: query: DROP TEMPORARY FUNCTION example_arraysum
+PREHOOK: type: DROPFUNCTION
+POSTHOOK: query: DROP TEMPORARY FUNCTION example_arraysum
+POSTHOOK: type: DROPFUNCTION
+PREHOOK: query: DROP TEMPORARY FUNCTION example_mapconcat
+PREHOOK: type: DROPFUNCTION
+POSTHOOK: query: DROP TEMPORARY FUNCTION example_mapconcat
+POSTHOOK: type: DROPFUNCTION
+PREHOOK: query: DROP TEMPORARY FUNCTION example_structprint
+PREHOOK: type: DROPFUNCTION
+POSTHOOK: query: DROP TEMPORARY FUNCTION example_structprint
+POSTHOOK: type: DROPFUNCTION

Modified: hadoop/hive/trunk/contrib/src/test/results/clientpositive/udf_example_format.q.out
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/contrib/src/test/results/clientpositive/udf_example_format.q.out?rev=819792&r1=819791&r2=819792&view=diff
==============================================================================
--- hadoop/hive/trunk/contrib/src/test/results/clientpositive/udf_example_format.q.out (original)
+++ hadoop/hive/trunk/contrib/src/test/results/clientpositive/udf_example_format.q.out Tue Sep 29 01:25:15 2009
@@ -1,10 +1,21 @@
-query: CREATE TEMPORARY FUNCTION example_format AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleFormat'
-query: EXPLAIN
+PREHOOK: query: CREATE TEMPORARY FUNCTION example_format AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleFormat'
+PREHOOK: type: CREATEFUNCTION
+POSTHOOK: query: CREATE TEMPORARY FUNCTION example_format AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleFormat'
+POSTHOOK: type: CREATEFUNCTION
+PREHOOK: query: EXPLAIN
 SELECT example_format("abc"),
        example_format("%1$s", 1.1),
        example_format("%1$s %2$e", 1.1, 1.2),
        example_format("%1$x %2$o %3$d", 10, 10, 10)
 FROM src LIMIT 1
+PREHOOK: type: QUERY
+POSTHOOK: query: EXPLAIN
+SELECT example_format("abc"),
+       example_format("%1$s", 1.1),
+       example_format("%1$s %2$e", 1.1, 1.2),
+       example_format("%1$x %2$o %3$d", 10, 10, 10)
+FROM src LIMIT 1
+POSTHOOK: type: QUERY
 ABSTRACT SYNTAX TREE:
   (TOK_QUERY (TOK_FROM (TOK_TABREF src)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION example_format "abc")) (TOK_SELEXPR (TOK_FUNCTION example_format "%1$s" 1.1)) (TOK_SELEXPR (TOK_FUNCTION example_format "%1$s %2$e" 1.1 1.2)) (TOK_SELEXPR (TOK_FUNCTION example_format "%1$x %2$o %3$d" 10 10 10))) (TOK_LIMIT 1)))
 
@@ -43,12 +54,24 @@
       limit: 1
 
 
-query: SELECT example_format("abc"),
+PREHOOK: query: SELECT example_format("abc"),
+       example_format("%1$s", 1.1),
+       example_format("%1$s %2$e", 1.1, 1.2),
+       example_format("%1$x %2$o %3$d", 10, 10, 10)
+FROM src LIMIT 1
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/508406280/10000
+POSTHOOK: query: SELECT example_format("abc"),
        example_format("%1$s", 1.1),
        example_format("%1$s %2$e", 1.1, 1.2),
        example_format("%1$x %2$o %3$d", 10, 10, 10)
 FROM src LIMIT 1
-Input: default/src
-Output: file:/data/users/zshao/tools/699-trunk-apache-hive/build/ql/tmp/857451543/10000
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/508406280/10000
 abc	1.1	1.1 1.200000e+00	a 12 10
-query: DROP TEMPORARY FUNCTION example_format
+PREHOOK: query: DROP TEMPORARY FUNCTION example_format
+PREHOOK: type: DROPFUNCTION
+POSTHOOK: query: DROP TEMPORARY FUNCTION example_format
+POSTHOOK: type: DROPFUNCTION

Modified: hadoop/hive/trunk/data/conf/hive-site.xml
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/data/conf/hive-site.xml?rev=819792&r1=819791&r2=819792&view=diff
==============================================================================
--- hadoop/hive/trunk/data/conf/hive-site.xml (original)
+++ hadoop/hive/trunk/data/conf/hive-site.xml Tue Sep 29 01:25:15 2009
@@ -58,7 +58,7 @@
 <property>
   <name>hive.metastore.metadb.dir</name>
   <value>file://${user.dir}/../build/ql/test/data/metadb/</value>
-  <description> 
+  <description>
   Required by metastore server or if the uris argument below is not supplied
   </description>
 </property>
@@ -125,6 +125,12 @@
 </property>
 
 <property>
+  <name>hive.exec.post.hooks</name>
+  <value>org.apache.hadoop.hive.ql.hooks.PostExecutePrinter</value>
+  <description>Post Execute Hook for Tests</description>
+</property>
+
+<property>
   <name>hive.task.progress</name>
   <value>false</value>
   <description>Track progress of a task</description>



Mime
View raw message