hive-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From gunt...@apache.org
Subject [07/51] [partial] hive git commit: HIVE-15790: Remove unused beeline golden files (Gunther Hagleitner, reviewed by Sergey Shelukhin)
Date Fri, 03 Feb 2017 21:50:20 GMT
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input20.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input20.q.out b/ql/src/test/results/beelinepositive/input20.q.out
deleted file mode 100644
index f1f9c90..0000000
--- a/ql/src/test/results/beelinepositive/input20.q.out
+++ /dev/null
@@ -1,437 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input20.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input20.q
->>>  CREATE TABLE dest1(key INT, value STRING) STORED AS TEXTFILE;
-No rows affected 
->>>  
->>>  ADD FILE ../data/scripts/input20_script.py;
-No rows affected 
->>>  
->>>  EXPLAIN 
-FROM ( 
-FROM src 
-MAP src.key, src.key 
-USING 'cat' 
-DISTRIBUTE BY key 
-SORT BY key, value 
-) tmap 
-INSERT OVERWRITE TABLE dest1 
-REDUCE tmap.key, tmap.value 
-USING 'python input20_script.py' 
-AS key, value;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL src) key)) TOK_SERDE TOK_RECORDWRITER 'cat' TOK_SERDE TOK_RECORDREADER))) (TOK_DISTRIBUTEBY (TOK_TABLE_OR_COL key)) (TOK_SORTBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL value))))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME dest1))) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL tmap) key) (. (TOK_TABLE_OR_COL tmap) value)) TOK_SERDE TOK_RECORDWRITER 'python input20_script.py' TOK_SERDE TOK_RECORDREADER (TOK_ALIASLIST key value))))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 depends on stages: Stage-1'
-'  Stage-2 depends on stages: Stage-0'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        tmap:src '
-'          TableScan'
-'            alias: src'
-'            Select Operator'
-'              expressions:'
-'                    expr: key'
-'                    type: string'
-'                    expr: key'
-'                    type: string'
-'              outputColumnNames: _col0, _col1'
-'              Transform Operator'
-'                command: cat'
-'                output info:'
-'                    input format: org.apache.hadoop.mapred.TextInputFormat'
-'                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'                  sort order: ++'
-'                  Map-reduce partition columns:'
-'                        expr: _col0'
-'                        type: string'
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: string'
-'                  expr: _col1'
-'                  type: string'
-'            outputColumnNames: _col0, _col1'
-'            Transform Operator'
-'              command: python input20_script.py'
-'              output info:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              Select Operator'
-'                expressions:'
-'                      expr: UDFToInteger(_col0)'
-'                      type: int'
-'                      expr: _col1'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                File Output Operator'
-'                  compressed: false'
-'                  GlobalTableId: 1'
-'                  table:'
-'                      input format: org.apache.hadoop.mapred.TextInputFormat'
-'                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                      name: input20.dest1'
-''
-'  Stage: Stage-0'
-'    Move Operator'
-'      tables:'
-'          replace: true'
-'          table:'
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input20.dest1'
-''
-'  Stage: Stage-2'
-'    Stats-Aggr Operator'
-''
-''
-87 rows selected 
->>>  
->>>  FROM ( 
-FROM src 
-MAP src.key, src.key 
-USING 'cat' 
-DISTRIBUTE BY key 
-SORT BY key, value 
-) tmap 
-INSERT OVERWRITE TABLE dest1 
-REDUCE tmap.key, tmap.value 
-USING 'python input20_script.py' 
-AS key, value;
-'_col0','_col1'
-No rows selected 
->>>  
->>>  SELECT * FROM dest1 SORT BY key, value;
-'key','value'
-'1','105_105'
-'1','10_10'
-'1','111_111'
-'1','114_114'
-'1','116_116'
-'1','11_11'
-'1','126_126'
-'1','131_131'
-'1','133_133'
-'1','136_136'
-'1','143_143'
-'1','145_145'
-'1','150_150'
-'1','153_153'
-'1','155_155'
-'1','156_156'
-'1','157_157'
-'1','158_158'
-'1','160_160'
-'1','162_162'
-'1','163_163'
-'1','166_166'
-'1','168_168'
-'1','170_170'
-'1','177_177'
-'1','178_178'
-'1','17_17'
-'1','180_180'
-'1','181_181'
-'1','183_183'
-'1','186_186'
-'1','189_189'
-'1','190_190'
-'1','192_192'
-'1','194_194'
-'1','196_196'
-'1','19_19'
-'1','201_201'
-'1','202_202'
-'1','20_20'
-'1','214_214'
-'1','218_218'
-'1','222_222'
-'1','226_226'
-'1','228_228'
-'1','235_235'
-'1','241_241'
-'1','244_244'
-'1','247_247'
-'1','248_248'
-'1','249_249'
-'1','252_252'
-'1','257_257'
-'1','258_258'
-'1','260_260'
-'1','262_262'
-'1','263_263'
-'1','266_266'
-'1','274_274'
-'1','275_275'
-'1','27_27'
-'1','283_283'
-'1','284_284'
-'1','285_285'
-'1','286_286'
-'1','287_287'
-'1','289_289'
-'1','28_28'
-'1','291_291'
-'1','292_292'
-'1','296_296'
-'1','2_2'
-'1','302_302'
-'1','305_305'
-'1','306_306'
-'1','308_308'
-'1','30_30'
-'1','310_310'
-'1','315_315'
-'1','323_323'
-'1','332_332'
-'1','335_335'
-'1','336_336'
-'1','338_338'
-'1','339_339'
-'1','33_33'
-'1','341_341'
-'1','345_345'
-'1','34_34'
-'1','351_351'
-'1','356_356'
-'1','360_360'
-'1','362_362'
-'1','364_364'
-'1','365_365'
-'1','366_366'
-'1','368_368'
-'1','373_373'
-'1','374_374'
-'1','375_375'
-'1','377_377'
-'1','378_378'
-'1','379_379'
-'1','386_386'
-'1','389_389'
-'1','392_392'
-'1','393_393'
-'1','394_394'
-'1','400_400'
-'1','402_402'
-'1','407_407'
-'1','411_411'
-'1','418_418'
-'1','419_419'
-'1','41_41'
-'1','421_421'
-'1','427_427'
-'1','432_432'
-'1','435_435'
-'1','436_436'
-'1','437_437'
-'1','43_43'
-'1','443_443'
-'1','444_444'
-'1','446_446'
-'1','448_448'
-'1','449_449'
-'1','44_44'
-'1','452_452'
-'1','453_453'
-'1','455_455'
-'1','457_457'
-'1','460_460'
-'1','467_467'
-'1','470_470'
-'1','472_472'
-'1','475_475'
-'1','477_477'
-'1','479_479'
-'1','47_47'
-'1','481_481'
-'1','482_482'
-'1','483_483'
-'1','484_484'
-'1','485_485'
-'1','487_487'
-'1','490_490'
-'1','491_491'
-'1','493_493'
-'1','494_494'
-'1','495_495'
-'1','496_496'
-'1','497_497'
-'1','4_4'
-'1','53_53'
-'1','54_54'
-'1','57_57'
-'1','64_64'
-'1','65_65'
-'1','66_66'
-'1','69_69'
-'1','74_74'
-'1','77_77'
-'1','78_78'
-'1','80_80'
-'1','82_82'
-'1','85_85'
-'1','86_86'
-'1','87_87'
-'1','8_8'
-'1','92_92'
-'1','96_96'
-'1','9_9'
-'2','100_100'
-'2','103_103'
-'2','104_104'
-'2','113_113'
-'2','118_118'
-'2','120_120'
-'2','125_125'
-'2','129_129'
-'2','12_12'
-'2','134_134'
-'2','137_137'
-'2','146_146'
-'2','149_149'
-'2','152_152'
-'2','15_15'
-'2','164_164'
-'2','165_165'
-'2','172_172'
-'2','174_174'
-'2','175_175'
-'2','176_176'
-'2','179_179'
-'2','18_18'
-'2','191_191'
-'2','195_195'
-'2','197_197'
-'2','200_200'
-'2','203_203'
-'2','205_205'
-'2','207_207'
-'2','209_209'
-'2','213_213'
-'2','216_216'
-'2','217_217'
-'2','219_219'
-'2','221_221'
-'2','223_223'
-'2','224_224'
-'2','229_229'
-'2','233_233'
-'2','237_237'
-'2','238_238'
-'2','239_239'
-'2','242_242'
-'2','24_24'
-'2','255_255'
-'2','256_256'
-'2','265_265'
-'2','26_26'
-'2','272_272'
-'2','278_278'
-'2','280_280'
-'2','281_281'
-'2','282_282'
-'2','288_288'
-'2','307_307'
-'2','309_309'
-'2','317_317'
-'2','321_321'
-'2','322_322'
-'2','325_325'
-'2','331_331'
-'2','333_333'
-'2','342_342'
-'2','344_344'
-'2','353_353'
-'2','367_367'
-'2','37_37'
-'2','382_382'
-'2','395_395'
-'2','397_397'
-'2','399_399'
-'2','404_404'
-'2','413_413'
-'2','414_414'
-'2','424_424'
-'2','429_429'
-'2','42_42'
-'2','439_439'
-'2','458_458'
-'2','459_459'
-'2','462_462'
-'2','463_463'
-'2','478_478'
-'2','492_492'
-'2','51_51'
-'2','58_58'
-'2','67_67'
-'2','72_72'
-'2','76_76'
-'2','83_83'
-'2','84_84'
-'2','95_95'
-'2','97_97'
-'2','98_98'
-'3','0_0'
-'3','119_119'
-'3','128_128'
-'3','167_167'
-'3','187_187'
-'3','193_193'
-'3','199_199'
-'3','208_208'
-'3','273_273'
-'3','298_298'
-'3','311_311'
-'3','316_316'
-'3','318_318'
-'3','327_327'
-'3','35_35'
-'3','369_369'
-'3','384_384'
-'3','396_396'
-'3','403_403'
-'3','409_409'
-'3','417_417'
-'3','430_430'
-'3','431_431'
-'3','438_438'
-'3','454_454'
-'3','466_466'
-'3','480_480'
-'3','498_498'
-'3','5_5'
-'3','70_70'
-'3','90_90'
-'4','138_138'
-'4','169_169'
-'4','277_277'
-'4','406_406'
-'4','468_468'
-'4','489_489'
-'5','230_230'
-'5','348_348'
-'5','401_401'
-'5','469_469'
-309 rows selected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input21.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input21.q.out b/ql/src/test/results/beelinepositive/input21.q.out
deleted file mode 100644
index 8431562..0000000
--- a/ql/src/test/results/beelinepositive/input21.q.out
+++ /dev/null
@@ -1,86 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input21.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input21.q
->>>  
->>>  
->>>  CREATE TABLE src_null(a STRING, b STRING, c STRING, d STRING) STORED AS TEXTFILE;
-No rows affected 
->>>  LOAD DATA LOCAL INPATH '../data/files/null.txt' INTO TABLE src_null;
-No rows affected 
->>>  
->>>  EXPLAIN SELECT * FROM src_null DISTRIBUTE BY c SORT BY d;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src_null))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_DISTRIBUTEBY (TOK_TABLE_OR_COL c)) (TOK_SORTBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL d)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src_null '
-'          TableScan'
-'            alias: src_null'
-'            Select Operator'
-'              expressions:'
-'                    expr: a'
-'                    type: string'
-'                    expr: b'
-'                    type: string'
-'                    expr: c'
-'                    type: string'
-'                    expr: d'
-'                    type: string'
-'              outputColumnNames: _col0, _col1, _col2, _col3'
-'              Reduce Output Operator'
-'                key expressions:'
-'                      expr: _col3'
-'                      type: string'
-'                sort order: +'
-'                Map-reduce partition columns:'
-'                      expr: _col2'
-'                      type: string'
-'                tag: -1'
-'                value expressions:'
-'                      expr: _col0'
-'                      type: string'
-'                      expr: _col1'
-'                      type: string'
-'                      expr: _col2'
-'                      type: string'
-'                      expr: _col3'
-'                      type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          File Output Operator'
-'            compressed: false'
-'            GlobalTableId: 0'
-'            table:'
-'                input format: org.apache.hadoop.mapred.TextInputFormat'
-'                output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-57 rows selected 
->>>  
->>>  SELECT * FROM src_null DISTRIBUTE BY c SORT BY d;
-'a','b','c','d'
-'1.0','1','same','0'
-'1.0','1','same','1'
-'1.0','1','same','2'
-'1.0','1','same','3'
-'1.0','1','same','4'
-'','1','same','5'
-'','','same','6'
-'1.0','','same','7'
-'1.0','1','same','8'
-'1.0','1','same','9'
-10 rows selected 
->>>  
->>>  
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input22.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input22.q.out b/ql/src/test/results/beelinepositive/input22.q.out
deleted file mode 100644
index af72a44..0000000
--- a/ql/src/test/results/beelinepositive/input22.q.out
+++ /dev/null
@@ -1,82 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input22.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input22.q
->>>  CREATE TABLE INPUT4(KEY STRING, VALUE STRING) STORED AS TEXTFILE;
-No rows affected 
->>>  LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE INPUT4;
-No rows affected 
->>>  
->>>  EXPLAIN 
-SELECT a.KEY2 
-FROM (SELECT INPUT4.*, INPUT4.KEY as KEY2 
-FROM INPUT4) a 
-ORDER BY KEY2 LIMIT 10;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME INPUT4))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME INPUT4))) (TOK_SELEXPR (. (TOK_TABLE_OR_COL INPUT4) KEY) KEY2)))) a)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL a) KEY2))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL KEY2))) (TOK_LIMIT 10)))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        a:input4 '
-'          TableScan'
-'            alias: input4'
-'            Select Operator'
-'              expressions:'
-'                    expr: key'
-'                    type: string'
-'              outputColumnNames: _col2'
-'              Select Operator'
-'                expressions:'
-'                      expr: _col2'
-'                      type: string'
-'                outputColumnNames: _col0'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                  sort order: +'
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          Limit'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: 10'
-''
-''
-48 rows selected 
->>>  
->>>  SELECT a.KEY2 
-FROM (SELECT INPUT4.*, INPUT4.KEY as KEY2 
-FROM INPUT4) a 
-ORDER BY KEY2 LIMIT 10;
-'key2'
-'0'
-'0'
-'0'
-'10'
-'100'
-'100'
-'103'
-'103'
-'104'
-'104'
-10 rows selected 
->>>  
->>>  
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input23.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input23.q.out b/ql/src/test/results/beelinepositive/input23.q.out
deleted file mode 100644
index ece4859..0000000
--- a/ql/src/test/results/beelinepositive/input23.q.out
+++ /dev/null
@@ -1,167 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input23.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input23.q
->>>  explain extended 
-select * from srcpart a join srcpart b where a.ds = '2008-04-08' and a.hr = '11' and b.ds = '2008-04-08' and b.hr = '14' limit 5;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME srcpart) a) (TOK_TABREF (TOK_TABNAME srcpart) b))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (and (and (and (= (. (TOK_TABLE_OR_COL a) ds) '2008-04-08') (= (. (TOK_TABLE_OR_COL a) hr) '11')) (= (. (TOK_TABLE_OR_COL b) ds) '2008-04-08')) (= (. (TOK_TABLE_OR_COL b) hr) '14'))) (TOK_LIMIT 5)))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        a '
-'          TableScan'
-'            alias: a'
-'            GatherStats: false'
-'            Reduce Output Operator'
-'              sort order: '
-'              tag: 0'
-'              value expressions:'
-'                    expr: key'
-'                    type: string'
-'                    expr: value'
-'                    type: string'
-'                    expr: ds'
-'                    type: string'
-'                    expr: hr'
-'                    type: string'
-'        b '
-'          TableScan'
-'            alias: b'
-'            GatherStats: false'
-'            Filter Operator'
-'              isSamplingPred: false'
-'              predicate:'
-'                  expr: ((ds = '2008-04-08') and (hr = '14'))'
-'                  type: boolean'
-'              Reduce Output Operator'
-'                sort order: '
-'                tag: 1'
-'                value expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                      expr: ds'
-'                      type: string'
-'                      expr: hr'
-'                      type: string'
-'      Needs Tagging: true'
-'      Path -> Alias:'
-'        !!{hive.metastore.warehouse.dir}!!/input23.db/srcpart/ds=2008-04-08/hr=11 [a]'
-'      Path -> Partition:'
-'        !!{hive.metastore.warehouse.dir}!!/input23.db/srcpart/ds=2008-04-08/hr=11 '
-'          Partition'
-'            base file name: hr=11'
-'            input format: org.apache.hadoop.mapred.TextInputFormat'
-'            output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'            partition values:'
-'              ds 2008-04-08'
-'              hr 11'
-'            properties:'
-'              bucket_count -1'
-'              columns key,value'
-'              columns.types string:string'
-'              file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              location !!{hive.metastore.warehouse.dir}!!/input23.db/srcpart/ds=2008-04-08/hr=11'
-'              name input23.srcpart'
-'              numFiles 1'
-'              numPartitions 4'
-'              numRows 0'
-'              partition_columns ds/hr'
-'              rawDataSize 0'
-'              serialization.ddl struct srcpart { string key, string value}'
-'              serialization.format 1'
-'              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              totalSize 5812'
-'              transient_lastDdlTime !!UNIXTIME!!'
-'            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'          '
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              properties:'
-'                bucket_count -1'
-'                columns key,value'
-'                columns.types string:string'
-'                file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                location !!{hive.metastore.warehouse.dir}!!/input23.db/srcpart'
-'                name input23.srcpart'
-'                numFiles 4'
-'                numPartitions 4'
-'                numRows 0'
-'                partition_columns ds/hr'
-'                rawDataSize 0'
-'                serialization.ddl struct srcpart { string key, string value}'
-'                serialization.format 1'
-'                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                totalSize 23248'
-'                transient_lastDdlTime !!UNIXTIME!!'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input23.srcpart'
-'            name: input23.srcpart'
-'      Reduce Operator Tree:'
-'        Join Operator'
-'          condition map:'
-'               Inner Join 0 to 1'
-'          condition expressions:'
-'            0 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3}'
-'            1 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3}'
-'          handleSkewJoin: false'
-'          outputColumnNames: _col0, _col1, _col2, _col3, _col6, _col7, _col8, _col9'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: string'
-'                  expr: _col1'
-'                  type: string'
-'                  expr: _col2'
-'                  type: string'
-'                  expr: _col3'
-'                  type: string'
-'                  expr: _col6'
-'                  type: string'
-'                  expr: _col7'
-'                  type: string'
-'                  expr: _col8'
-'                  type: string'
-'                  expr: _col9'
-'                  type: string'
-'            outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7'
-'            Limit'
-'              File Output Operator'
-'                compressed: false'
-'                GlobalTableId: 0'
-'                directory: file:!!{hive.exec.scratchdir}!!'
-'                NumFilesPerFileSink: 1'
-'                Stats Publishing Key Prefix: file:!!{hive.exec.scratchdir}!!'
-'                table:'
-'                    input format: org.apache.hadoop.mapred.TextInputFormat'
-'                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                    properties:'
-'                      columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7'
-'                      columns.types string:string:string:string:string:string:string:string'
-'                      escape.delim \'
-'                      serialization.format 1'
-'                TotalFiles: 1'
-'                GatherStats: false'
-'                MultiFileSpray: false'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: 5'
-''
-''
-155 rows selected 
->>>  
->>>  select * from srcpart a join srcpart b where a.ds = '2008-04-08' and a.hr = '11' and b.ds = '2008-04-08' and b.hr = '14' limit 5;
-'key','value','ds','hr','key','value','ds','hr'
-No rows selected 
->>>  
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input24.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input24.q.out b/ql/src/test/results/beelinepositive/input24.q.out
deleted file mode 100644
index e8e6140..0000000
--- a/ql/src/test/results/beelinepositive/input24.q.out
+++ /dev/null
@@ -1,69 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input24.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input24.q
->>>  
->>>  create table tst(a int, b int) partitioned by (d string);
-No rows affected 
->>>  alter table tst add partition (d='2009-01-01');
-No rows affected 
->>>  explain 
-select count(1) from tst x where x.d='2009-01-01';
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME tst) x)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION count 1))) (TOK_WHERE (= (. (TOK_TABLE_OR_COL x) d) '2009-01-01'))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        x '
-'          TableScan'
-'            alias: x'
-'            Select Operator'
-'              Group By Operator'
-'                aggregations:'
-'                      expr: count(1)'
-'                bucketGroup: false'
-'                mode: hash'
-'                outputColumnNames: _col0'
-'                Reduce Output Operator'
-'                  sort order: '
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: bigint'
-'      Reduce Operator Tree:'
-'        Group By Operator'
-'          aggregations:'
-'                expr: count(VALUE._col0)'
-'          bucketGroup: false'
-'          mode: mergepartial'
-'          outputColumnNames: _col0'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: bigint'
-'            outputColumnNames: _col0'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-50 rows selected 
->>>  
->>>  select count(1) from tst x where x.d='2009-01-01';
-'_c0'
-'0'
-1 row selected 
->>>  
->>>  
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input25.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input25.q.out b/ql/src/test/results/beelinepositive/input25.q.out
deleted file mode 100644
index 0f3e882..0000000
--- a/ql/src/test/results/beelinepositive/input25.q.out
+++ /dev/null
@@ -1,156 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input25.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input25.q
->>>  
->>>  create table tst(a int, b int) partitioned by (d string);
-No rows affected 
->>>  alter table tst add partition (d='2009-01-01');
-No rows affected 
->>>  alter table tst add partition (d='2009-02-02');
-No rows affected 
->>>  
->>>  explain 
-select * from ( 
-select * from tst x where x.d='2009-01-01' limit 10 
-union all 
-select * from tst x where x.d='2009-02-02' limit 10 
-) subq;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_UNION (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME tst) x)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (= (. (TOK_TABLE_OR_COL x) d) '2009-01-01')) (TOK_LIMIT 10))) (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME tst) x)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (= (. (TOK_TABLE_OR_COL x) d) '2009-02-02')) (TOK_LIMIT 10)))) subq)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-2 depends on stages: Stage-1, Stage-3'
-'  Stage-3 is a root stage'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        null-subquery1:subq-subquery1:x '
-'          TableScan'
-'            alias: x'
-'            Select Operator'
-'              expressions:'
-'                    expr: a'
-'                    type: int'
-'                    expr: b'
-'                    type: int'
-'                    expr: d'
-'                    type: string'
-'              outputColumnNames: _col0, _col1, _col2'
-'              Limit'
-'                Reduce Output Operator'
-'                  sort order: '
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: int'
-'                        expr: _col1'
-'                        type: int'
-'                        expr: _col2'
-'                        type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          Limit'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-'  Stage: Stage-2'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        file:!!{hive.exec.scratchdir}!! '
-'          TableScan'
-'            Union'
-'              Select Operator'
-'                expressions:'
-'                      expr: _col0'
-'                      type: int'
-'                      expr: _col1'
-'                      type: int'
-'                      expr: _col2'
-'                      type: string'
-'                outputColumnNames: _col0, _col1, _col2'
-'                File Output Operator'
-'                  compressed: false'
-'                  GlobalTableId: 0'
-'                  table:'
-'                      input format: org.apache.hadoop.mapred.TextInputFormat'
-'                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'        file:!!{hive.exec.scratchdir}!! '
-'          TableScan'
-'            Union'
-'              Select Operator'
-'                expressions:'
-'                      expr: _col0'
-'                      type: int'
-'                      expr: _col1'
-'                      type: int'
-'                      expr: _col2'
-'                      type: string'
-'                outputColumnNames: _col0, _col1, _col2'
-'                File Output Operator'
-'                  compressed: false'
-'                  GlobalTableId: 0'
-'                  table:'
-'                      input format: org.apache.hadoop.mapred.TextInputFormat'
-'                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-3'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        null-subquery2:subq-subquery2:x '
-'          TableScan'
-'            alias: x'
-'            Select Operator'
-'              expressions:'
-'                    expr: a'
-'                    type: int'
-'                    expr: b'
-'                    type: int'
-'                    expr: d'
-'                    type: string'
-'              outputColumnNames: _col0, _col1, _col2'
-'              Limit'
-'                Reduce Output Operator'
-'                  sort order: '
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: int'
-'                        expr: _col1'
-'                        type: int'
-'                        expr: _col2'
-'                        type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          Limit'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-127 rows selected 
->>>  
->>>  select * from ( 
-select * from tst x where x.d='2009-01-01' limit 10 
-union all 
-select * from tst x where x.d='2009-02-02' limit 10 
-) subq;
-'a','b','d'
-No rows selected 
->>>  
->>>  
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input26.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input26.q.out b/ql/src/test/results/beelinepositive/input26.q.out
deleted file mode 100644
index 0a99de2..0000000
--- a/ql/src/test/results/beelinepositive/input26.q.out
+++ /dev/null
@@ -1,169 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input26.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input26.q
->>>  explain 
-select * from ( 
-select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5 
-union all 
-select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5 
-)subq;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_UNION (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME srcpart) a)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (and (= (. (TOK_TABLE_OR_COL a) ds) '2008-04-08') (= (. (TOK_TABLE_OR_COL a) hr) '11'))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (. (TOK_TABLE_OR_COL a) key))) (TOK_LIMIT 5))) (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME srcpart) b)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (and (= (. (TOK_TABLE_OR_COL b) ds) '2008-04-08') (= (. (TOK_TABLE_OR_COL b) hr) '14'))) (TOK_LIMIT 5)))) subq)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-2 depends on stages: Stage-1, Stage-3'
-'  Stage-3 is a root stage'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        null-subquery1:subq-subquery1:a '
-'          TableScan'
-'            alias: a'
-'            Select Operator'
-'              expressions:'
-'                    expr: key'
-'                    type: string'
-'                    expr: value'
-'                    type: string'
-'                    expr: ds'
-'                    type: string'
-'                    expr: hr'
-'                    type: string'
-'              outputColumnNames: _col0, _col1, _col2, _col3'
-'              Reduce Output Operator'
-'                key expressions:'
-'                      expr: _col0'
-'                      type: string'
-'                sort order: +'
-'                tag: -1'
-'                value expressions:'
-'                      expr: _col0'
-'                      type: string'
-'                      expr: _col1'
-'                      type: string'
-'                      expr: _col2'
-'                      type: string'
-'                      expr: _col3'
-'                      type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          Limit'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-'  Stage: Stage-2'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        file:!!{hive.exec.scratchdir}!! '
-'          TableScan'
-'            Union'
-'              Select Operator'
-'                expressions:'
-'                      expr: _col0'
-'                      type: string'
-'                      expr: _col1'
-'                      type: string'
-'                      expr: _col2'
-'                      type: string'
-'                      expr: _col3'
-'                      type: string'
-'                outputColumnNames: _col0, _col1, _col2, _col3'
-'                File Output Operator'
-'                  compressed: false'
-'                  GlobalTableId: 0'
-'                  table:'
-'                      input format: org.apache.hadoop.mapred.TextInputFormat'
-'                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'        file:!!{hive.exec.scratchdir}!! '
-'          TableScan'
-'            Union'
-'              Select Operator'
-'                expressions:'
-'                      expr: _col0'
-'                      type: string'
-'                      expr: _col1'
-'                      type: string'
-'                      expr: _col2'
-'                      type: string'
-'                      expr: _col3'
-'                      type: string'
-'                outputColumnNames: _col0, _col1, _col2, _col3'
-'                File Output Operator'
-'                  compressed: false'
-'                  GlobalTableId: 0'
-'                  table:'
-'                      input format: org.apache.hadoop.mapred.TextInputFormat'
-'                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-3'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        null-subquery2:subq-subquery2:b '
-'          TableScan'
-'            alias: b'
-'            Filter Operator'
-'              predicate:'
-'                  expr: ((ds = '2008-04-08') and (hr = '14'))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                      expr: ds'
-'                      type: string'
-'                      expr: hr'
-'                      type: string'
-'                outputColumnNames: _col0, _col1, _col2, _col3'
-'                Limit'
-'                  Reduce Output Operator'
-'                    sort order: '
-'                    tag: -1'
-'                    value expressions:'
-'                          expr: _col0'
-'                          type: string'
-'                          expr: _col1'
-'                          type: string'
-'                          expr: _col2'
-'                          type: string'
-'                          expr: _col3'
-'                          type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          Limit'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.SequenceFileInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-''
-''
-145 rows selected 
->>>  
->>>  select * from ( 
-select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5 
-union all 
-select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5 
-)subq;
-'key','value','ds','hr'
-'0','val_0','2008-04-08','11'
-'0','val_0','2008-04-08','11'
-'0','val_0','2008-04-08','11'
-'10','val_10','2008-04-08','11'
-'100','val_100','2008-04-08','11'
-5 rows selected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input28.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input28.q.out b/ql/src/test/results/beelinepositive/input28.q.out
deleted file mode 100644
index 38c07fe..0000000
--- a/ql/src/test/results/beelinepositive/input28.q.out
+++ /dev/null
@@ -1,19 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input28.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input28.q
->>>  
->>>  create table tst(a string, b string) partitioned by (d string);
-No rows affected 
->>>  alter table tst add partition (d='2009-01-01');
-No rows affected 
->>>  
->>>  insert overwrite table tst partition(d='2009-01-01') 
-select tst.a, src.value from tst join src ON (tst.a = src.key);
-'a','value'
-No rows selected 
->>>  
->>>  select * from tst where tst.d='2009-01-01';
-'a','b','d'
-No rows selected 
->>>  
->>>  
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input2_limit.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input2_limit.q.out b/ql/src/test/results/beelinepositive/input2_limit.q.out
deleted file mode 100644
index 67ae187..0000000
--- a/ql/src/test/results/beelinepositive/input2_limit.q.out
+++ /dev/null
@@ -1,54 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input2_limit.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input2_limit.q
->>>  EXPLAIN 
-SELECT x.* FROM SRC x WHERE x.key < 300 LIMIT 5;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME SRC) x)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME x)))) (TOK_WHERE (< (. (TOK_TABLE_OR_COL x) key) 300)) (TOK_LIMIT 5)))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        x '
-'          TableScan'
-'            alias: x'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (key < 300.0)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                Limit'
-'                  File Output Operator'
-'                    compressed: false'
-'                    GlobalTableId: 0'
-'                    table:'
-'                        input format: org.apache.hadoop.mapred.TextInputFormat'
-'                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-''
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: 5'
-''
-''
-38 rows selected 
->>>  
->>>  SELECT x.* FROM SRC x WHERE x.key < 300 LIMIT 5;
-'key','value'
-'238','val_238'
-'86','val_86'
-'27','val_27'
-'165','val_165'
-'255','val_255'
-5 rows selected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input3.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input3.q.out b/ql/src/test/results/beelinepositive/input3.q.out
deleted file mode 100644
index 80ba8b6..0000000
--- a/ql/src/test/results/beelinepositive/input3.q.out
+++ /dev/null
@@ -1,138 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input3.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input3.q
->>>  
->>>  
->>>  
->>>  
->>>  CREATE TABLE TEST3a(A INT, B DOUBLE) STORED AS TEXTFILE;
-No rows affected 
->>>  DESCRIBE TEST3a;
-'col_name','data_type','comment'
-'a','int',''
-'b','double',''
-2 rows selected 
->>>  CREATE TABLE TEST3b(A ARRAY<INT>, B DOUBLE, C MAP<DOUBLE, INT>) STORED AS TEXTFILE;
-No rows affected 
->>>  DESCRIBE TEST3b;
-'col_name','data_type','comment'
-'a','array<int>',''
-'b','double',''
-'c','map<double,int>',''
-3 rows selected 
->>>  SHOW TABLES;
-'tab_name'
-'primitives'
-'src'
-'src1'
-'src_json'
-'src_sequencefile'
-'src_thrift'
-'srcbucket'
-'srcbucket2'
-'srcpart'
-'test3a'
-'test3b'
-11 rows selected 
->>>  EXPLAIN 
-ALTER TABLE TEST3b ADD COLUMNS (X DOUBLE);
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_ALTERTABLE_ADDCOLS TEST3b (TOK_TABCOLLIST (TOK_TABCOL X TOK_DOUBLE)))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-0'
-'      Alter Table Operator:'
-'        Alter Table'
-'          type: add columns'
-'          new columns: x double'
-'          old name: TEST3b'
-''
-''
-15 rows selected 
->>>  ALTER TABLE TEST3b ADD COLUMNS (X DOUBLE);
-No rows affected 
->>>  DESCRIBE TEST3b;
-'col_name','data_type','comment'
-'a','array<int>',''
-'b','double',''
-'c','map<double,int>',''
-'x','double',''
-4 rows selected 
->>>  EXPLAIN 
-ALTER TABLE TEST3b RENAME TO TEST3c;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_ALTERTABLE_RENAME TEST3b TEST3c)'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-0'
-'      Alter Table Operator:'
-'        Alter Table'
-'          type: rename'
-'          new name: TEST3c'
-'          old name: TEST3b'
-''
-''
-15 rows selected 
->>>  ALTER TABLE TEST3b RENAME TO TEST3c;
-No rows affected 
->>>  DESCRIBE TEST3c;
-'col_name','data_type','comment'
-'a','array<int>',''
-'b','double',''
-'c','map<double,int>',''
-'x','double',''
-4 rows selected 
->>>  SHOW TABLES;
-'tab_name'
-'primitives'
-'src'
-'src1'
-'src_json'
-'src_sequencefile'
-'src_thrift'
-'srcbucket'
-'srcbucket2'
-'srcpart'
-'test3a'
-'test3c'
-11 rows selected 
->>>  EXPLAIN 
-ALTER TABLE TEST3c REPLACE COLUMNS (R1 INT, R2 DOUBLE);
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_ALTERTABLE_REPLACECOLS TEST3c (TOK_TABCOLLIST (TOK_TABCOL R1 TOK_INT) (TOK_TABCOL R2 TOK_DOUBLE)))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-0'
-'      Alter Table Operator:'
-'        Alter Table'
-'          type: replace columns'
-'          new columns: r1 int, r2 double'
-'          old name: TEST3c'
-''
-''
-15 rows selected 
->>>  ALTER TABLE TEST3c REPLACE COLUMNS (R1 INT, R2 DOUBLE);
-No rows affected 
->>>  DESCRIBE EXTENDED TEST3c;
-'col_name','data_type','comment'
-'r1','int',''
-'r2','double',''
-'','',''
-'Detailed Table Information','Table(tableName:test3c, dbName:input3, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:r1, type:int, comment:null), FieldSchema(name:r2, type:double, comment:null)], location:!!{hive.metastore.warehouse.dir}!!/input3.db/test3c, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{last_modified_by=!!ELIDED!!, last_modified_time=!!UNIXTIME!!, transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)',''
-4 rows selected 
->>>  
->>>  
->>>  
->>>  
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input30.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input30.q.out b/ql/src/test/results/beelinepositive/input30.q.out
deleted file mode 100644
index 5cfa26d..0000000
--- a/ql/src/test/results/beelinepositive/input30.q.out
+++ /dev/null
@@ -1,110 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input30.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input30.q
->>>  
->>>  
->>>  
->>>  
->>>  create table dest30(a int);
-No rows affected 
->>>  create table tst_dest30(a int);
-No rows affected 
->>>  
->>>  set hive.test.mode=true;
-No rows affected 
->>>  set hive.test.mode.prefix=tst_;
-No rows affected 
->>>  
->>>  explain 
-insert overwrite table dest30 
-select count(1) from src;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME dest30))) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION count 1)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 depends on stages: Stage-1'
-'  Stage-2 depends on stages: Stage-0'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src '
-'          TableScan'
-'            alias: src'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (((hash(rand(460476415)) & 2147483647) % 32) = 0)'
-'                  type: boolean'
-'              Select Operator'
-'                Group By Operator'
-'                  aggregations:'
-'                        expr: count(1)'
-'                  bucketGroup: false'
-'                  mode: hash'
-'                  outputColumnNames: _col0'
-'                  Reduce Output Operator'
-'                    sort order: '
-'                    tag: -1'
-'                    value expressions:'
-'                          expr: _col0'
-'                          type: bigint'
-'      Reduce Operator Tree:'
-'        Group By Operator'
-'          aggregations:'
-'                expr: count(VALUE._col0)'
-'          bucketGroup: false'
-'          mode: mergepartial'
-'          outputColumnNames: _col0'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: bigint'
-'            outputColumnNames: _col0'
-'            Select Operator'
-'              expressions:'
-'                    expr: UDFToInteger(_col0)'
-'                    type: int'
-'              outputColumnNames: _col0'
-'              File Output Operator'
-'                compressed: false'
-'                GlobalTableId: 1'
-'                table:'
-'                    input format: org.apache.hadoop.mapred.TextInputFormat'
-'                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                    name: input30.tst_dest30'
-''
-'  Stage: Stage-0'
-'    Move Operator'
-'      tables:'
-'          replace: true'
-'          table:'
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input30.tst_dest30'
-''
-'  Stage: Stage-2'
-'    Stats-Aggr Operator'
-''
-''
-72 rows selected 
->>>  
->>>  insert overwrite table dest30 
-select count(1) from src;
-'_col0'
-No rows selected 
->>>  
->>>  set hive.test.mode=false;
-No rows affected 
->>>  
->>>  select * from tst_dest30;
-'a'
-'18'
-1 row selected 
->>>  
->>>  
->>>  
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input31.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input31.q.out b/ql/src/test/results/beelinepositive/input31.q.out
deleted file mode 100644
index 229fd8b..0000000
--- a/ql/src/test/results/beelinepositive/input31.q.out
+++ /dev/null
@@ -1,111 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input31.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input31.q
->>>  
->>>  
->>>  
->>>  set hive.test.mode=true;
-No rows affected 
->>>  set hive.test.mode.prefix=tst_;
-No rows affected 
->>>  
->>>  create table tst_dest31(a int);
-No rows affected 
->>>  create table dest31(a int);
-No rows affected 
->>>  
->>>  explain 
-insert overwrite table dest31 
-select count(1) from srcbucket;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME srcbucket))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME dest31))) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION count 1)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 depends on stages: Stage-1'
-'  Stage-2 depends on stages: Stage-0'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        srcbucket '
-'          TableScan'
-'            alias: srcbucket'
-'            Filter Operator'
-'              predicate:'
-'                  expr: (((hash(key) & 2147483647) % 2) = 0)'
-'                  type: boolean'
-'              Select Operator'
-'                Group By Operator'
-'                  aggregations:'
-'                        expr: count(1)'
-'                  bucketGroup: false'
-'                  mode: hash'
-'                  outputColumnNames: _col0'
-'                  Reduce Output Operator'
-'                    sort order: '
-'                    tag: -1'
-'                    value expressions:'
-'                          expr: _col0'
-'                          type: bigint'
-'      Reduce Operator Tree:'
-'        Group By Operator'
-'          aggregations:'
-'                expr: count(VALUE._col0)'
-'          bucketGroup: false'
-'          mode: mergepartial'
-'          outputColumnNames: _col0'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: bigint'
-'            outputColumnNames: _col0'
-'            Select Operator'
-'              expressions:'
-'                    expr: UDFToInteger(_col0)'
-'                    type: int'
-'              outputColumnNames: _col0'
-'              File Output Operator'
-'                compressed: false'
-'                GlobalTableId: 1'
-'                table:'
-'                    input format: org.apache.hadoop.mapred.TextInputFormat'
-'                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                    name: input31.tst_dest31'
-''
-'  Stage: Stage-0'
-'    Move Operator'
-'      tables:'
-'          replace: true'
-'          table:'
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input31.tst_dest31'
-''
-'  Stage: Stage-2'
-'    Stats-Aggr Operator'
-''
-''
-72 rows selected 
->>>  
->>>  insert overwrite table dest31 
-select count(1) from srcbucket;
-'_col0'
-No rows selected 
->>>  
->>>  set hive.test.mode=false;
-No rows affected 
->>>  
->>>  select * from tst_dest31;
-'a'
-'493'
-1 row selected 
->>>  
->>>  
->>>  
->>>  
->>>  
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input32.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input32.q.out b/ql/src/test/results/beelinepositive/input32.q.out
deleted file mode 100644
index c9ba3c8a..0000000
--- a/ql/src/test/results/beelinepositive/input32.q.out
+++ /dev/null
@@ -1,109 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input32.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input32.q
->>>  
->>>  
->>>  
->>>  set hive.test.mode=true;
-No rows affected 
->>>  set hive.test.mode.prefix=tst_;
-No rows affected 
->>>  set hive.test.mode.nosamplelist=src,srcbucket;
-No rows affected 
->>>  
->>>  create table dest32(a int);
-No rows affected 
->>>  create table tst_dest32(a int);
-No rows affected 
->>>  
->>>  explain 
-insert overwrite table dest32 
-select count(1) from srcbucket;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME srcbucket))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME dest32))) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION count 1)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 depends on stages: Stage-1'
-'  Stage-2 depends on stages: Stage-0'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        srcbucket '
-'          TableScan'
-'            alias: srcbucket'
-'            Select Operator'
-'              Group By Operator'
-'                aggregations:'
-'                      expr: count(1)'
-'                bucketGroup: false'
-'                mode: hash'
-'                outputColumnNames: _col0'
-'                Reduce Output Operator'
-'                  sort order: '
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: bigint'
-'      Reduce Operator Tree:'
-'        Group By Operator'
-'          aggregations:'
-'                expr: count(VALUE._col0)'
-'          bucketGroup: false'
-'          mode: mergepartial'
-'          outputColumnNames: _col0'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: bigint'
-'            outputColumnNames: _col0'
-'            Select Operator'
-'              expressions:'
-'                    expr: UDFToInteger(_col0)'
-'                    type: int'
-'              outputColumnNames: _col0'
-'              File Output Operator'
-'                compressed: false'
-'                GlobalTableId: 1'
-'                table:'
-'                    input format: org.apache.hadoop.mapred.TextInputFormat'
-'                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                    name: input32.tst_dest32'
-''
-'  Stage: Stage-0'
-'    Move Operator'
-'      tables:'
-'          replace: true'
-'          table:'
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input32.tst_dest32'
-''
-'  Stage: Stage-2'
-'    Stats-Aggr Operator'
-''
-''
-68 rows selected 
->>>  
->>>  insert overwrite table dest32 
-select count(1) from srcbucket;
-'_col0'
-No rows selected 
->>>  
->>>  set hive.test.mode=false;
-No rows affected 
->>>  
->>>  select * from tst_dest32;
-'a'
-'1000'
-1 row selected 
->>>  
->>>  
->>>  
->>>  
->>>  
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input33.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input33.q.out b/ql/src/test/results/beelinepositive/input33.q.out
deleted file mode 100644
index b7ad369..0000000
--- a/ql/src/test/results/beelinepositive/input33.q.out
+++ /dev/null
@@ -1,437 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input33.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input33.q
->>>  CREATE TABLE dest1(key INT, value STRING) STORED AS TEXTFILE;
-No rows affected 
->>>  
->>>  ADD FILE ../data/scripts/input20_script.py;
-No rows affected 
->>>  
->>>  EXPLAIN 
-FROM ( 
-FROM src 
-MAP src.key, src.key 
-USING 'cat' 
-DISTRIBUTE BY key 
-SORT BY key, value 
-) tmap 
-INSERT OVERWRITE TABLE dest1 
-REDUCE tmap.key, tmap.value 
-USING 'python input20_script.py' 
-AS (key STRING, value STRING);
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL src) key)) TOK_SERDE TOK_RECORDWRITER 'cat' TOK_SERDE TOK_RECORDREADER))) (TOK_DISTRIBUTEBY (TOK_TABLE_OR_COL key)) (TOK_SORTBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL value))))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME dest1))) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL tmap) key) (. (TOK_TABLE_OR_COL tmap) value)) TOK_SERDE TOK_RECORDWRITER 'python input20_script.py' TOK_SERDE TOK_RECORDREADER (TOK_TABCOLLIST (TOK_TABCOL key TOK_STRING) (TOK_TABCOL value TOK_STRING)))))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 depends on stages: Stage-1'
-'  Stage-2 depends on stages: Stage-0'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        tmap:src '
-'          TableScan'
-'            alias: src'
-'            Select Operator'
-'              expressions:'
-'                    expr: key'
-'                    type: string'
-'                    expr: key'
-'                    type: string'
-'              outputColumnNames: _col0, _col1'
-'              Transform Operator'
-'                command: cat'
-'                output info:'
-'                    input format: org.apache.hadoop.mapred.TextInputFormat'
-'                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                Reduce Output Operator'
-'                  key expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'                  sort order: ++'
-'                  Map-reduce partition columns:'
-'                        expr: _col0'
-'                        type: string'
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'      Reduce Operator Tree:'
-'        Extract'
-'          Select Operator'
-'            expressions:'
-'                  expr: _col0'
-'                  type: string'
-'                  expr: _col1'
-'                  type: string'
-'            outputColumnNames: _col0, _col1'
-'            Transform Operator'
-'              command: python input20_script.py'
-'              output info:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              Select Operator'
-'                expressions:'
-'                      expr: UDFToInteger(_col0)'
-'                      type: int'
-'                      expr: _col1'
-'                      type: string'
-'                outputColumnNames: _col0, _col1'
-'                File Output Operator'
-'                  compressed: false'
-'                  GlobalTableId: 1'
-'                  table:'
-'                      input format: org.apache.hadoop.mapred.TextInputFormat'
-'                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                      name: input33.dest1'
-''
-'  Stage: Stage-0'
-'    Move Operator'
-'      tables:'
-'          replace: true'
-'          table:'
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input33.dest1'
-''
-'  Stage: Stage-2'
-'    Stats-Aggr Operator'
-''
-''
-87 rows selected 
->>>  
->>>  FROM ( 
-FROM src 
-MAP src.key, src.key 
-USING 'cat' 
-DISTRIBUTE BY key 
-SORT BY key, value 
-) tmap 
-INSERT OVERWRITE TABLE dest1 
-REDUCE tmap.key, tmap.value 
-USING 'python input20_script.py' 
-AS (key STRING, value STRING);
-'_col0','_col1'
-No rows selected 
->>>  
->>>  SELECT * FROM dest1 SORT BY key, value;
-'key','value'
-'1','105_105'
-'1','10_10'
-'1','111_111'
-'1','114_114'
-'1','116_116'
-'1','11_11'
-'1','126_126'
-'1','131_131'
-'1','133_133'
-'1','136_136'
-'1','143_143'
-'1','145_145'
-'1','150_150'
-'1','153_153'
-'1','155_155'
-'1','156_156'
-'1','157_157'
-'1','158_158'
-'1','160_160'
-'1','162_162'
-'1','163_163'
-'1','166_166'
-'1','168_168'
-'1','170_170'
-'1','177_177'
-'1','178_178'
-'1','17_17'
-'1','180_180'
-'1','181_181'
-'1','183_183'
-'1','186_186'
-'1','189_189'
-'1','190_190'
-'1','192_192'
-'1','194_194'
-'1','196_196'
-'1','19_19'
-'1','201_201'
-'1','202_202'
-'1','20_20'
-'1','214_214'
-'1','218_218'
-'1','222_222'
-'1','226_226'
-'1','228_228'
-'1','235_235'
-'1','241_241'
-'1','244_244'
-'1','247_247'
-'1','248_248'
-'1','249_249'
-'1','252_252'
-'1','257_257'
-'1','258_258'
-'1','260_260'
-'1','262_262'
-'1','263_263'
-'1','266_266'
-'1','274_274'
-'1','275_275'
-'1','27_27'
-'1','283_283'
-'1','284_284'
-'1','285_285'
-'1','286_286'
-'1','287_287'
-'1','289_289'
-'1','28_28'
-'1','291_291'
-'1','292_292'
-'1','296_296'
-'1','2_2'
-'1','302_302'
-'1','305_305'
-'1','306_306'
-'1','308_308'
-'1','30_30'
-'1','310_310'
-'1','315_315'
-'1','323_323'
-'1','332_332'
-'1','335_335'
-'1','336_336'
-'1','338_338'
-'1','339_339'
-'1','33_33'
-'1','341_341'
-'1','345_345'
-'1','34_34'
-'1','351_351'
-'1','356_356'
-'1','360_360'
-'1','362_362'
-'1','364_364'
-'1','365_365'
-'1','366_366'
-'1','368_368'
-'1','373_373'
-'1','374_374'
-'1','375_375'
-'1','377_377'
-'1','378_378'
-'1','379_379'
-'1','386_386'
-'1','389_389'
-'1','392_392'
-'1','393_393'
-'1','394_394'
-'1','400_400'
-'1','402_402'
-'1','407_407'
-'1','411_411'
-'1','418_418'
-'1','419_419'
-'1','41_41'
-'1','421_421'
-'1','427_427'
-'1','432_432'
-'1','435_435'
-'1','436_436'
-'1','437_437'
-'1','43_43'
-'1','443_443'
-'1','444_444'
-'1','446_446'
-'1','448_448'
-'1','449_449'
-'1','44_44'
-'1','452_452'
-'1','453_453'
-'1','455_455'
-'1','457_457'
-'1','460_460'
-'1','467_467'
-'1','470_470'
-'1','472_472'
-'1','475_475'
-'1','477_477'
-'1','479_479'
-'1','47_47'
-'1','481_481'
-'1','482_482'
-'1','483_483'
-'1','484_484'
-'1','485_485'
-'1','487_487'
-'1','490_490'
-'1','491_491'
-'1','493_493'
-'1','494_494'
-'1','495_495'
-'1','496_496'
-'1','497_497'
-'1','4_4'
-'1','53_53'
-'1','54_54'
-'1','57_57'
-'1','64_64'
-'1','65_65'
-'1','66_66'
-'1','69_69'
-'1','74_74'
-'1','77_77'
-'1','78_78'
-'1','80_80'
-'1','82_82'
-'1','85_85'
-'1','86_86'
-'1','87_87'
-'1','8_8'
-'1','92_92'
-'1','96_96'
-'1','9_9'
-'2','100_100'
-'2','103_103'
-'2','104_104'
-'2','113_113'
-'2','118_118'
-'2','120_120'
-'2','125_125'
-'2','129_129'
-'2','12_12'
-'2','134_134'
-'2','137_137'
-'2','146_146'
-'2','149_149'
-'2','152_152'
-'2','15_15'
-'2','164_164'
-'2','165_165'
-'2','172_172'
-'2','174_174'
-'2','175_175'
-'2','176_176'
-'2','179_179'
-'2','18_18'
-'2','191_191'
-'2','195_195'
-'2','197_197'
-'2','200_200'
-'2','203_203'
-'2','205_205'
-'2','207_207'
-'2','209_209'
-'2','213_213'
-'2','216_216'
-'2','217_217'
-'2','219_219'
-'2','221_221'
-'2','223_223'
-'2','224_224'
-'2','229_229'
-'2','233_233'
-'2','237_237'
-'2','238_238'
-'2','239_239'
-'2','242_242'
-'2','24_24'
-'2','255_255'
-'2','256_256'
-'2','265_265'
-'2','26_26'
-'2','272_272'
-'2','278_278'
-'2','280_280'
-'2','281_281'
-'2','282_282'
-'2','288_288'
-'2','307_307'
-'2','309_309'
-'2','317_317'
-'2','321_321'
-'2','322_322'
-'2','325_325'
-'2','331_331'
-'2','333_333'
-'2','342_342'
-'2','344_344'
-'2','353_353'
-'2','367_367'
-'2','37_37'
-'2','382_382'
-'2','395_395'
-'2','397_397'
-'2','399_399'
-'2','404_404'
-'2','413_413'
-'2','414_414'
-'2','424_424'
-'2','429_429'
-'2','42_42'
-'2','439_439'
-'2','458_458'
-'2','459_459'
-'2','462_462'
-'2','463_463'
-'2','478_478'
-'2','492_492'
-'2','51_51'
-'2','58_58'
-'2','67_67'
-'2','72_72'
-'2','76_76'
-'2','83_83'
-'2','84_84'
-'2','95_95'
-'2','97_97'
-'2','98_98'
-'3','0_0'
-'3','119_119'
-'3','128_128'
-'3','167_167'
-'3','187_187'
-'3','193_193'
-'3','199_199'
-'3','208_208'
-'3','273_273'
-'3','298_298'
-'3','311_311'
-'3','316_316'
-'3','318_318'
-'3','327_327'
-'3','35_35'
-'3','369_369'
-'3','384_384'
-'3','396_396'
-'3','403_403'
-'3','409_409'
-'3','417_417'
-'3','430_430'
-'3','431_431'
-'3','438_438'
-'3','454_454'
-'3','466_466'
-'3','480_480'
-'3','498_498'
-'3','5_5'
-'3','70_70'
-'3','90_90'
-'4','138_138'
-'4','169_169'
-'4','277_277'
-'4','406_406'
-'4','468_468'
-'4','489_489'
-'5','230_230'
-'5','348_348'
-'5','401_401'
-'5','469_469'
-309 rows selected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input34.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input34.q.out b/ql/src/test/results/beelinepositive/input34.q.out
deleted file mode 100644
index 82620b6..0000000
--- a/ql/src/test/results/beelinepositive/input34.q.out
+++ /dev/null
@@ -1,640 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input34.q.raw". Enter "record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input34.q
->>>  CREATE TABLE dest1(key INT, value STRING) STORED AS TEXTFILE;
-No rows affected 
->>>  
->>>  EXPLAIN 
-FROM ( 
-FROM src 
-SELECT TRANSFORM(src.key, src.value) ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' 
-USING 'cat' 
-AS (tkey, tvalue) ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' 
-) tmap 
-INSERT OVERWRITE TABLE dest1 SELECT tkey, tvalue;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (. (TOK_TABLE_OR_COL src) key) (. (TOK_TABLE_OR_COL src) value)) (TOK_SERDE (TOK_SERDENAME 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe')) TOK_RECORDWRITER 'cat' (TOK_SERDE (TOK_SERDENAME 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe')) TOK_RECORDREADER (TOK_ALIASLIST tkey tvalue)))))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME dest1))) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL tkey)) (TOK_SELEXPR (TOK_TABLE_OR_COL tvalue)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5'
-'  Stage-4'
-'  Stage-0 depends on stages: Stage-4, Stage-3, Stage-6'
-'  Stage-2 depends on stages: Stage-0'
-'  Stage-3'
-'  Stage-5'
-'  Stage-6 depends on stages: Stage-5'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        tmap:src '
-'          TableScan'
-'            alias: src'
-'            Select Operator'
-'              expressions:'
-'                    expr: key'
-'                    type: string'
-'                    expr: value'
-'                    type: string'
-'              outputColumnNames: _col0, _col1'
-'              Transform Operator'
-'                command: cat'
-'                output info:'
-'                    input format: org.apache.hadoop.mapred.TextInputFormat'
-'                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                Select Operator'
-'                  expressions:'
-'                        expr: _col0'
-'                        type: string'
-'                        expr: _col1'
-'                        type: string'
-'                  outputColumnNames: _col0, _col1'
-'                  Select Operator'
-'                    expressions:'
-'                          expr: UDFToInteger(_col0)'
-'                          type: int'
-'                          expr: _col1'
-'                          type: string'
-'                    outputColumnNames: _col0, _col1'
-'                    File Output Operator'
-'                      compressed: false'
-'                      GlobalTableId: 1'
-'                      table:'
-'                          input format: org.apache.hadoop.mapred.TextInputFormat'
-'                          output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                          serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                          name: input34.dest1'
-''
-'  Stage: Stage-7'
-'    Conditional Operator'
-''
-'  Stage: Stage-4'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: pfile:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-0'
-'    Move Operator'
-'      tables:'
-'          replace: true'
-'          table:'
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input34.dest1'
-''
-'  Stage: Stage-2'
-'    Stats-Aggr Operator'
-''
-'  Stage: Stage-3'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        pfile:!!{hive.exec.scratchdir}!! '
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                  name: input34.dest1'
-''
-'  Stage: Stage-5'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        pfile:!!{hive.exec.scratchdir}!! '
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                  name: input34.dest1'
-''
-'  Stage: Stage-6'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          destination: pfile:!!{hive.exec.scratchdir}!!'
-''
-''
-110 rows selected 
->>>  
->>>  FROM ( 
-FROM src 
-SELECT TRANSFORM(src.key, src.value) ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' 
-USING 'cat' 
-AS (tkey, tvalue) ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' 
-) tmap 
-INSERT OVERWRITE TABLE dest1 SELECT tkey, tvalue;
-'_col0','_col1'
-No rows selected 
->>>  
->>>  SELECT dest1.* FROM dest1;
-'key','value'
-'238','val_238'
-'86','val_86'
-'311','val_311'
-'27','val_27'
-'165','val_165'
-'409','val_409'
-'255','val_255'
-'278','val_278'
-'98','val_98'
-'484','val_484'
-'265','val_265'
-'193','val_193'
-'401','val_401'
-'150','val_150'
-'273','val_273'
-'224','val_224'
-'369','val_369'
-'66','val_66'
-'128','val_128'
-'213','val_213'
-'146','val_146'
-'406','val_406'
-'429','val_429'
-'374','val_374'
-'152','val_152'
-'469','val_469'
-'145','val_145'
-'495','val_495'
-'37','val_37'
-'327','val_327'
-'281','val_281'
-'277','val_277'
-'209','val_209'
-'15','val_15'
-'82','val_82'
-'403','val_403'
-'166','val_166'
-'417','val_417'
-'430','val_430'
-'252','val_252'
-'292','val_292'
-'219','val_219'
-'287','val_287'
-'153','val_153'
-'193','val_193'
-'338','val_338'
-'446','val_446'
-'459','val_459'
-'394','val_394'
-'237','val_237'
-'482','val_482'
-'174','val_174'
-'413','val_413'
-'494','val_494'
-'207','val_207'
-'199','val_199'
-'466','val_466'
-'208','val_208'
-'174','val_174'
-'399','val_399'
-'396','val_396'
-'247','val_247'
-'417','val_417'
-'489','val_489'
-'162','val_162'
-'377','val_377'
-'397','val_397'
-'309','val_309'
-'365','val_365'
-'266','val_266'
-'439','val_439'
-'342','val_342'
-'367','val_367'
-'325','val_325'
-'167','val_167'
-'195','val_195'
-'475','val_475'
-'17','val_17'
-'113','val_113'
-'155','val_155'
-'203','val_203'
-'339','val_339'
-'0','val_0'
-'455','val_455'
-'128','val_128'
-'311','val_311'
-'316','val_316'
-'57','val_57'
-'302','val_302'
-'205','val_205'
-'149','val_149'
-'438','val_438'
-'345','val_345'
-'129','val_129'
-'170','val_170'
-'20','val_20'
-'489','val_489'
-'157','val_157'
-'378','val_378'
-'221','val_221'
-'92','val_92'
-'111','val_111'
-'47','val_47'
-'72','val_72'
-'4','val_4'
-'280','val_280'
-'35','val_35'
-'427','val_427'
-'277','val_277'
-'208','val_208'
-'356','val_356'
-'399','val_399'
-'169','val_169'
-'382','val_382'
-'498','val_498'
-'125','val_125'
-'386','val_386'
-'437','val_437'
-'469','val_469'
-'192','val_192'
-'286','val_286'
-'187','val_187'
-'176','val_176'
-'54','val_54'
-'459','val_459'
-'51','val_51'
-'138','val_138'
-'103','val_103'
-'239','val_239'
-'213','val_213'
-'216','val_216'
-'430','val_430'
-'278','val_278'
-'176','val_176'
-'289','val_289'
-'221','val_221'
-'65','val_65'
-'318','val_318'
-'332','val_332'
-'311','val_311'
-'275','val_275'
-'137','val_137'
-'241','val_241'
-'83','val_83'
-'333','val_333'
-'180','val_180'
-'284','val_284'
-'12','val_12'
-'230','val_230'
-'181','val_181'
-'67','val_67'
-'260','val_260'
-'404','val_404'
-'384','val_384'
-'489','val_489'
-'353','val_353'
-'373','val_373'
-'272','val_272'
-'138','val_138'
-'217','val_217'
-'84','val_84'
-'348','val_348'
-'466','val_466'
-'58','val_58'
-'8','val_8'
-'411','val_411'
-'230','val_230'
-'208','val_208'
-'348','val_348'
-'24','val_24'
-'463','val_463'
-'431','val_431'
-'179','val_179'
-'172','val_172'
-'42','val_42'
-'129','val_129'
-'158','val_158'
-'119','val_119'
-'496','val_496'
-'0','val_0'
-'322','val_322'
-'197','val_197'
-'468','val_468'
-'393','val_393'
-'454','val_454'
-'100','val_100'
-'298','val_298'
-'199','val_199'
-'191','val_191'
-'418','val_418'
-'96','val_96'
-'26','val_26'
-'165','val_165'
-'327','val_327'
-'230','val_230'
-'205','val_205'
-'120','val_120'
-'131','val_131'
-'51','val_51'
-'404','val_404'
-'43','val_43'
-'436','val_436'
-'156','val_156'
-'469','val_469'
-'468','val_468'
-'308','val_308'
-'95','val_95'
-'196','val_196'
-'288','val_288'
-'481','val_481'
-'457','val_457'
-'98','val_98'
-'282','val_282'
-'197','val_197'
-'187','val_187'
-'318','val_318'
-'318','val_318'
-'409','val_409'
-'470','val_470'
-'137','val_137'
-'369','val_369'
-'316','val_316'
-'169','val_169'
-'413','val_413'
-'85','val_85'
-'77','val_77'
-'0','val_0'
-'490','val_490'
-'87','val_87'
-'364','val_364'
-'179','val_179'
-'118','val_118'
-'134','val_134'
-'395','val_395'
-'282','val_282'
-'138','val_138'
-'238','val_238'
-'419','val_419'
-'15','val_15'
-'118','val_118'
-'72','val_72'
-'90','val_90'
-'307','val_307'
-'19','val_19'
-'435','val_435'
-'10','val_10'
-'277','val_277'
-'273','val_273'
-'306','val_306'
-'224','val_224'
-'309','val_309'
-'389','val_389'
-'327','val_327'
-'242','val_242'
-'369','val_369'
-'392','val_392'
-'272','val_272'
-'331','val_331'
-'401','val_401'
-'242','val_242'
-'452','val_452'
-'177','val_177'
-'226','val_226'
-'5','val_5'
-'497','val_497'
-'402','val_402'
-'396','val_396'
-'317','val_317'
-'395','val_395'
-'58','val_58'
-'35','val_35'
-'336','val_336'
-'95','val_95'
-'11','val_11'
-'168','val_168'
-'34','val_34'
-'229','val_229'
-'233','val_233'
-'143','val_143'
-'472','val_472'
-'322','val_322'
-'498','val_498'
-'160','val_160'
-'195','val_195'
-'42','val_42'
-'321','val_321'
-'430','val_430'
-'119','val_119'
-'489','val_489'
-'458','val_458'
-'78','val_78'
-'76','val_76'
-'41','val_41'
-'223','val_223'
-'492','val_492'
-'149','val_149'
-'449','val_449'
-'218','val_218'
-'228','val_228'
-'138','val_138'
-'453','val_453'
-'30','val_30'
-'209','val_209'
-'64','val_64'
-'468','val_468'
-'76','val_76'
-'74','val_74'
-'342','val_342'
-'69','val_69'
-'230','val_230'
-'33','val_33'
-'368','val_368'
-'103','val_103'
-'296','val_296'
-'113','val_113'
-'216','val_216'
-'367','val_367'
-'344','val_344'
-'167','val_167'
-'274','val_274'
-'219','val_219'
-'239','val_239'
-'485','val_485'
-'116','val_116'
-'223','val_223'
-'256','val_256'
-'263','val_263'
-'70','val_70'
-'487','val_487'
-'480','val_480'
-'401','val_401'
-'288','val_288'
-'191','val_191'
-'5','val_5'
-'244','val_244'
-'438','val_438'
-'128','val_128'
-'467','val_467'
-'432','val_432'
-'202','val_202'
-'316','val_316'
-'229','val_229'
-'469','val_469'
-'463','val_463'
-'280','val_280'
-'2','val_2'
-'35','val_35'
-'283','val_283'
-'331','val_331'
-'235','val_235'
-'80','val_80'
-'44','val_44'
-'193','val_193'
-'321','val_321'
-'335','val_335'
-'104','val_104'
-'466','val_466'
-'366','val_366'
-'175','val_175'
-'403','val_403'
-'483','val_483'
-'53','val_53'
-'105','val_105'
-'257','val_257'
-'406','val_406'
-'409','val_409'
-'190','val_190'
-'406','val_406'
-'401','val_401'
-'114','val_114'
-'258','val_258'
-'90','val_90'
-'203','val_203'
-'262','val_262'
-'348','val_348'
-'424','val_424'
-'12','val_12'
-'396','val_396'
-'201','val_201'
-'217','val_217'
-'164','val_164'
-'431','val_431'
-'454','val_454'
-'478','val_478'
-'298','val_298'
-'125','val_125'
-'431','val_431'
-'164','val_164'
-'424','val_424'
-'187','val_187'
-'382','val_382'
-'5','val_5'
-'70','val_70'
-'397','val_397'
-'480','val_480'
-'291','val_291'
-'24','val_24'
-'351','val_351'
-'255','val_255'
-'104','val_104'
-'70','val_70'
-'163','val_163'
-'438','val_438'
-'119','val_119'
-'414','val_414'
-'200','val_200'
-'491','val_491'
-'237','val_237'
-'439','val_439'
-'360','val_360'
-'248','val_248'
-'479','val_479'
-'305','val_305'
-'417','val_417'
-'199','val_199'
-'444','val_444'
-'120','val_120'
-'429','val_429'
-'169','val_169'
-'443','val_443'
-'323','val_323'
-'325','val_325'
-'277','val_277'
-'230','val_230'
-'478','val_478'
-'178','val_178'
-'468','val_468'
-'310','val_310'
-'317','val_317'
-'333','val_333'
-'493','val_493'
-'460','val_460'
-'207','val_207'
-'249','val_249'
-'265','val_265'
-'480','val_480'
-'83','val_83'
-'136','val_136'
-'353','val_353'
-'172','val_172'
-'214','val_214'
-'462','val_462'
-'233','val_233'
-'406','val_406'
-'133','val_133'
-'175','val_175'
-'189','val_189'
-'454','val_454'
-'375','val_375'
-'401','val_401'
-'421','val_421'
-'407','val_407'
-'384','val_384'
-'256','val_256'
-'26','val_26'
-'134','val_134'
-'67','val_67'
-'384','val_384'
-'379','val_379'
-'18','val_18'
-'462','val_462'
-'492','val_492'
-'100','val_100'
-'298','val_298'
-'9','val_9'
-'341','val_341'
-'498','val_498'
-'146','val_146'
-'458','val_458'
-'362','val_362'
-'186','val_186'
-'285','val_285'
-'348','val_348'
-'167','val_167'
-'18','val_18'
-'273','val_273'
-'183','val_183'
-'281','val_281'
-'344','val_344'
-'97','val_97'
-'469','val_469'
-'315','val_315'
-'84','val_84'
-'28','val_28'
-'37','val_37'
-'448','val_448'
-'152','val_152'
-'348','val_348'
-'307','val_307'
-'194','val_194'
-'414','val_414'
-'477','val_477'
-'222','val_222'
-'126','val_126'
-'90','val_90'
-'169','val_169'
-'403','val_403'
-'400','val_400'
-'200','val_200'
-'97','val_97'
-500 rows selected 
->>>  !record


Mime
View raw message