hawq-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From ztao1...@apache.org
Subject [2/2] incubator-hawq git commit: HAWQ-791. remove parquet related test from installcheck-good.
Date Wed, 08 Jun 2016 03:03:03 GMT
HAWQ-791. remove parquet related test from installcheck-good.


Project: http://git-wip-us.apache.org/repos/asf/incubator-hawq/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-hawq/commit/62f2dfb3
Tree: http://git-wip-us.apache.org/repos/asf/incubator-hawq/tree/62f2dfb3
Diff: http://git-wip-us.apache.org/repos/asf/incubator-hawq/diff/62f2dfb3

Branch: refs/heads/master
Commit: 62f2dfb3c903589449c87e46065d3808f3d80dbf
Parents: 3d11a58
Author: ztao1987 <zhenglin.taozl@gmail.com>
Authored: Wed Jun 8 10:58:14 2016 +0800
Committer: ztao1987 <zhenglin.taozl@gmail.com>
Committed: Wed Jun 8 11:02:54 2016 +0800

----------------------------------------------------------------------
 src/test/feature/parquet/test_parquet.cpp       |  46 ++--
 .../regress/expected/parquet_compression.out    | 147 -----------
 src/test/regress/expected/parquet_ddl.out       | 112 --------
 .../regress/expected/parquet_multipletype.out   |  14 -
 .../expected/parquet_pagerowgroup_size.out      |  74 ------
 .../regress/expected/parquet_subpartition.out   | 261 -------------------
 src/test/regress/known_good_schedule            |   5 -
 src/test/regress/sql/parquet_compression.sql    | 124 ---------
 src/test/regress/sql/parquet_ddl.sql            |  75 ------
 src/test/regress/sql/parquet_multipletype.sql   |   5 -
 .../regress/sql/parquet_pagerowgroup_size.sql   |  45 ----
 src/test/regress/sql/parquet_subpartition.sql   | 109 --------
 12 files changed, 25 insertions(+), 992 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/62f2dfb3/src/test/feature/parquet/test_parquet.cpp
----------------------------------------------------------------------
diff --git a/src/test/feature/parquet/test_parquet.cpp b/src/test/feature/parquet/test_parquet.cpp
index cf27870..52d9870 100644
--- a/src/test/feature/parquet/test_parquet.cpp
+++ b/src/test/feature/parquet/test_parquet.cpp
@@ -17,32 +17,36 @@ TEST_F(TestParquet, TestMultipleType) {
   hawq::test::DataGenerator dGen(&util);
   dGen.genTableWithFullTypes("t1", true, "parquet");
   util.query(
-      "select * from t1",
-      "2147483647|||00:00:00||ff:89:71:45:ae:01|2000-01-01 "
-      "07:00:00+08||32767|t|192.168.1.255/"
-      "32|<(1,2),3>|[(0,0),(6,6)]|-178000000 "
-      "years|0|-$21,474,836.48|(200,400),(100,200)||<aa>bb</"
-      "aa>||123456789a|2001:db8:85a3:8d3:1319:8a2e:370:7344/"
-      "64|||1|0|(1,2)|4277-12-31|128|\n0|((100,123),(5,10),(7,2),(4,5))|hello "
+      "select * from t1 order by "
+      "c0,c2,c7,c8,c22,c25,c28",
+      "0|((100,123),(5,10),(7,2),(4,5))|hello "
+      "world||04:45:05.0012+08:40|||bbccddeeff|128||2001:db8:85a3:8d3:1319:"
+      "8a2e:370:7344/128|<(1,2),3>||||$0.00||hello "
+      "world|||||0||2147483647|-Infinity|(1,2)|4277-12-31|Infinity|\n0|((100,"
+      "123),(5,10),(7,2),(4,5))|hello "
       "world||04:45:05.0012+08:40|||bbccddeeff|128||2001:db8:85a3:8d3:1319:"
       "8a2e:370:7344/128|<(1,2),3>|[(0,0),(6,6)]||1|$0.00|(2,3),(0,1)|hello "
       "world|<aa>bb</aa>||aaaa|2001:db8:85a3:8d3:1319:8a2e:370:7344/"
-      "64|0||2147483647|-Infinity|(1,2)|4277-12-31|Infinity|\n||abcd|15:01:03||"
-      "|2000-01-01 07:00:00+08|||t|||[(0,0),(6,6)]|-178000000 "
-      "years|0|-$21,474,836.48|(200,400),(100,200)||<aa>bb</"
-      "aa>||123456789a|2001:db8:85a3:8d3:1319:8a2e:370:7344/"
-      "64|||1|0|(1,2)|4277-12-31|128|\n0|((100,123),(5,10),(7,2),(4,5))|hello "
+      "64|0||2147483647|-Infinity|(1,2)|4277-12-31|Infinity|\n0|((100,123),(5,"
+      "10),(7,2),(4,5))|hello "
       "world||04:45:05.0012+08:40|||bbccddeeff|128||2001:db8:85a3:8d3:1319:"
       "8a2e:370:7344/128|<(1,2),3>||||$0.00||hello "
       "world|||aaaa|2001:db8:85a3:8d3:1319:8a2e:370:7344/"
       "64|0||2147483647|-Infinity|(1,2)|4277-12-31|Infinity|\n0|((100,123),(5,"
       "10),(7,2),(4,5))|hello "
       "world||04:45:05.0012+08:40|||bbccddeeff|128||2001:db8:85a3:8d3:1319:"
-      "8a2e:370:7344/128|<(1,2),3>||||$0.00||hello "
-      "world|||||0||2147483647|-Infinity|(1,2)|4277-12-31|Infinity|\n0|((100,"
-      "123),(5,10),(7,2),(4,5))|hello "
-      "world||04:45:05.0012+08:40|||bbccddeeff|128||2001:db8:85a3:8d3:1319:"
-      "8a2e:370:7344/128|<(1,2),3>||||||hello world|||||0||34|||||\n");
+      "8a2e:370:7344/128|<(1,2),3>||||||hello "
+      "world|||||0||34|||||\n2147483647|||00:00:00||ff:89:71:45:ae:01|2000-01-"
+      "01 "
+      "07:00:00+08||32767|t|192.168.1.255/"
+      "32|<(1,2),3>|[(0,0),(6,6)]|-178000000 "
+      "years|0|-$21,474,836.48|(200,400),(100,200)||<aa>bb</"
+      "aa>||123456789a|2001:db8:85a3:8d3:1319:8a2e:370:7344/"
+      "64|||1|0|(1,2)|4277-12-31|128|\n||abcd|15:01:03|||2000-01-01 "
+      "07:00:00+08|||t|||[(0,0),(6,6)]|-178000000 "
+      "years|0|-$21,474,836.48|(200,400),(100,200)||<aa>bb</"
+      "aa>||123456789a|2001:db8:85a3:8d3:1319:8a2e:370:7344/"
+      "64|||1|0|(1,2)|4277-12-31|128|\n");
 }
 
 TEST_F(TestParquet, TestCompression) {
@@ -52,10 +56,10 @@ TEST_F(TestParquet, TestCompression) {
   hawq::test::DataGenerator dGen(&util);
   dGen.genTableWithNull("t21", true, "parquet", "gzip", 9);
   dGen.genTableWithNull("t22", true, "parquet", "snappy");
-  util.query("select * from t21,t22",
-             "15||aa|15||aa|\n15||aa|||WET|\n15||aa||51||\n||WET|15||aa|\n||"
-             "WET|||WET|\n||WET||51||\n|51||15||aa|\n|51||||WET|\n|51|||51||"
-             "\n");
+  util.query(
+      "select * from t21,t22 order by t21.a,t21.b,t21.c,t22.a,t22.b,t22.c",
+      "15||aa|15||aa|\n15||aa||51||\n15||aa|||WET|\n|51||15||aa|\n|51|||51||\n|"
+      "51||||WET|\n||WET|15||aa|\n||WET||51||\n||WET|||WET|\n");
 }
 
 TEST_F(TestParquet, TestSize) {

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/62f2dfb3/src/test/regress/expected/parquet_compression.out
----------------------------------------------------------------------
diff --git a/src/test/regress/expected/parquet_compression.out b/src/test/regress/expected/parquet_compression.out
deleted file mode 100644
index b214aec..0000000
--- a/src/test/regress/expected/parquet_compression.out
+++ /dev/null
@@ -1,147 +0,0 @@
--- @description parquet compression test
--- @created 2013-08-09 19:57:38
--- @modified 2013-08-09 19:57:38
--- @tags HAWQ parquet
---start_ignore
-drop schema parquet cascade;
-ERROR:  schema "parquet" does not exist
-create schema parquet;
-set search_path=parquet,"$user",public;
-set default_segment_num = 4;
-drop table parquet_gzip;
-ERROR:  table "parquet_gzip" does not exist
-drop table parquet_gzip_uncompr;
-ERROR:  table "parquet_gzip_uncompr" does not exist
-drop table parquet_gzip_part;
-ERROR:  table "parquet_gzip_part" does not exist
-drop table parquet_gzip_part_unc;
-ERROR:  table "parquet_gzip_part_unc" does not exist
-drop table parquet_snappy;
-ERROR:  table "parquet_snappy" does not exist
-drop table parquet_snappy_uncompr;
-ERROR:  table "parquet_snappy_uncompr" does not exist
-drop table parquet_snappy_part;
-ERROR:  table "parquet_snappy_part" does not exist
-drop table parquet_snappy_part_unc;
-ERROR:  table "parquet_snappy_part_unc" does not exist
-drop table parquet_gzip_2;
-ERROR:  table "parquet_gzip_2" does not exist
-alter resource queue pg_default with ( vseg_resource_quota='mem:4gb');
---end_ignore
---Datatypes covered: text,bytea,varchar,bit varying
--- parquet table ,compresstype = gzip
-create table parquet_gzip (i int,document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, compresstype=gzip, pagesize=150000000, rowgroupsize= 500000000);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'i' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-insert into parquet_gzip  values(1001,repeat('large data value for text data type',32732),repeat('large data value for text data type',32732),pg_catalog.decode(repeat('large data value for text data type',32732), 'escape'),1000::int::bit(3244447)::varbit);
--- Uncompressed table and join with compressed to see if select works
-create table parquet_gzip_uncompr(i int,document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, pagesize=150000000, rowgroupsize= 500000000);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'i' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-insert into parquet_gzip_uncompr values(1001,repeat('large data value for text data type',32732),repeat('large data value for text data type',32732),pg_catalog.decode(repeat('large data value for text data type',32732), 'escape'),1000::int::bit(3244447)::varbit);
-select count(*) from parquet_gzip c1 full outer join parquet_gzip_uncompr c2 on c1.document=c2.document and c1.vch1=c2.vch1 and c1.bta1=c2.bta1 and c1.bitv1=c2.bitv1;
- count 
--------
-     1
-(1 row)
-
---select get_ao_compression_ratio('parquet_gzip');
--- parquet partition table with hybrid partitions
-create table parquet_gzip_part (p1 int,document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, compresstype=gzip, pagesize=150000000, rowgroupsize= 500000000)
-      Partition by range(p1)
-      (Partition p1 start(1) end(4) with (appendonly=true),
-       Partition p2 start(4) end(7) with (appendonly=true),
-       Partition p3 start(7) end(11) with (appendonly=true,orientation=parquet,compresstype=gzip, pagesize=150000000, rowgroupsize= 500000000));
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'p1' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-NOTICE:  CREATE TABLE will create partition "parquet_gzip_part_1_prt_p1" for table "parquet_gzip_part"
-NOTICE:  CREATE TABLE will create partition "parquet_gzip_part_1_prt_p2" for table "parquet_gzip_part"
-NOTICE:  CREATE TABLE will create partition "parquet_gzip_part_1_prt_p3" for table "parquet_gzip_part"
-insert into parquet_gzip_part values(generate_series(1,5), repeat('large data value for text data type',32732),repeat('large data value for text data type',32732),pg_catalog.decode(repeat('large data value for text data type',32732), 'escape'),1000::int::bit(3244447)::varbit);
--- Uncompressed table and join with compressed to see if select works
-create table parquet_gzip_part_unc (p1 int,document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, pagesize=150000000, rowgroupsize= 500000000)
-      Partition by range(p1)
-      (Partition p1 start(1) end(4) with (appendonly=true),
-       Partition p2 start(4) end(7) with (appendonly=true),
-       Partition p3 start(7) end(11) with (appendonly=true,orientation=parquet, pagesize=150000000, rowgroupsize= 500000000));
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'p1' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-NOTICE:  CREATE TABLE will create partition "parquet_gzip_part_unc_1_prt_p1" for table "parquet_gzip_part_unc"
-NOTICE:  CREATE TABLE will create partition "parquet_gzip_part_unc_1_prt_p2" for table "parquet_gzip_part_unc"
-NOTICE:  CREATE TABLE will create partition "parquet_gzip_part_unc_1_prt_p3" for table "parquet_gzip_part_unc"
-insert into parquet_gzip_part_unc values(generate_series(1,5), repeat('large data value for text data type',32732),repeat('large data value for text data type',32732),pg_catalog.decode(repeat('large data value for text data type',32732), 'escape'),1000::int::bit(3244447)::varbit);
-select count(*) from parquet_gzip_part c1 full outer join parquet_gzip_part_unc c2 on c1.p1=c2.p1 and c1.document=c2.document and c1.vch1=c2.vch1 and c1.bta1=c2.bta1 and c1.bitv1=c2.bitv1;
- count 
--------
-     5
-(1 row)
-
---select get_ao_compression_ratio('parquet_gzip_part_1_prt_p3');
--- parquet table , compresstype = snappy
-create table parquet_snappy (p1 int, document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, compresstype=snappy, pagesize=150000000, rowgroupsize= 500000000);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'p1' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-insert into parquet_snappy values(1001, repeat('large data value for text data type',32732),repeat('large data value for text data type',32732),pg_catalog.decode(repeat('large data value for text data type',32732), 'escape'),1000::int::bit(3244447)::varbit);
--- Uncompressed table and join with compressed to see if select works
-create table parquet_snappy_uncompr(p1 int, document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, pagesize=150000000, rowgroupsize= 500000000);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'p1' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-insert into parquet_snappy_uncompr values(1001,repeat('large data value for text data type',32732),repeat('large data value for text data type',32732),pg_catalog.decode(repeat('large data value for text data type',32732), 'escape'),1000::int::bit(3244447)::varbit);
-select count(*) from parquet_snappy c1 full outer join parquet_snappy_uncompr c2 on c1.p1=c2.p1 and c1.document=c2.document and c1.vch1=c2.vch1 and c1.bta1=c2.bta1 and c1.bitv1=c2.bitv1;
- count 
--------
-     1
-(1 row)
-
---select get_ao_compression_ratio('parquet_snappy');
--- parquet partition table with hybrid partitions
-create table parquet_snappy_part (p1 int,document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, compresstype=snappy, pagesize=150000000, rowgroupsize= 500000000)
-      Partition by range(p1)
-      (Partition p1 start(1) end(4) with (appendonly=true),
-       Partition p2 start(4) end(7) with (appendonly=true),
-       Partition p3 start(7) end(11) with (appendonly=true,orientation=parquet,compresstype=snappy, pagesize=150000000, rowgroupsize= 500000000));
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'p1' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-NOTICE:  CREATE TABLE will create partition "parquet_snappy_part_1_prt_p1" for table "parquet_snappy_part"
-NOTICE:  CREATE TABLE will create partition "parquet_snappy_part_1_prt_p2" for table "parquet_snappy_part"
-NOTICE:  CREATE TABLE will create partition "parquet_snappy_part_1_prt_p3" for table "parquet_snappy_part"
-insert into parquet_snappy_part values(generate_series(1,5), repeat('large data value for text data type',32732),repeat('large data value for text data type',32732),pg_catalog.decode(repeat('large data value for text data type',32732), 'escape'),1000::int::bit(3244447)::varbit);
--- Uncompressed table and join with compressed to see if select works
-create table parquet_snappy_part_unc (p1 int,document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, pagesize=150000000, rowgroupsize= 500000000)
-      Partition by range(p1)
-      (Partition p1 start(1) end(4) with (appendonly=true),
-       Partition p2 start(4) end(7) with (appendonly=true),
-       Partition p3 start(7) end(11) with (appendonly=true,orientation=parquet, pagesize=150000000, rowgroupsize= 500000000));
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'p1' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-NOTICE:  CREATE TABLE will create partition "parquet_snappy_part_unc_1_prt_p1" for table "parquet_snappy_part_unc"
-NOTICE:  CREATE TABLE will create partition "parquet_snappy_part_unc_1_prt_p2" for table "parquet_snappy_part_unc"
-NOTICE:  CREATE TABLE will create partition "parquet_snappy_part_unc_1_prt_p3" for table "parquet_snappy_part_unc"
-insert into parquet_snappy_part_unc values(generate_series(1,5), repeat('large data value for text data type',32732),repeat('large data value for text data type',32732),pg_catalog.decode(repeat('large data value for text data type',32732), 'escape'),1000::int::bit(3244447)::varbit);
-select count(*) from parquet_snappy_part c1 full outer join parquet_snappy_part_unc c2 on c1.p1=c2.p1 and c1.document=c2.document and c1.vch1=c2.vch1 and c1.bta1=c2.bta1 and c1.bitv1=c2.bitv1;
- count 
--------
-     5
-(1 row)
-
---select get_ao_compression_ratio('parquet_snappy_part_1_prt_p3');
-create table parquet_gzip_2 (i int,document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, compresstype=gzip, pagesize=150000000, rowgroupsize= 500000000);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'i' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-insert into parquet_gzip_2 values(12,array_to_string(ARRAY(SELECT chr((65 + round(random() * 25)) :: integer) FROM generate_series(1, 32732)), ''),array_to_string(ARRAY(SELECT chr((65 + round(random() * 25)) :: integer) FROM generate_series(1, 32732)), ''),pg_catalog.decode(array_to_string(ARRAY(SELECT chr((65 + round(random() * 25)) :: integer) FROM generate_series(1, 32732)), ''),'escape'),1010110::int::bit(3244447)::varbit);
-Select count(*) from parquet_gzip_2;
- count 
--------
-     1
-(1 row)
-
-alter resource queue pg_default with ( vseg_resource_quota='mem:256mb');

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/62f2dfb3/src/test/regress/expected/parquet_ddl.out
----------------------------------------------------------------------
diff --git a/src/test/regress/expected/parquet_ddl.out b/src/test/regress/expected/parquet_ddl.out
deleted file mode 100644
index 9c2b440..0000000
--- a/src/test/regress/expected/parquet_ddl.out
+++ /dev/null
@@ -1,112 +0,0 @@
---start_ignore
-drop table if exists parquet_table_tobe_ctas;
-NOTICE:  table "parquet_table_tobe_ctas" does not exist, skipping
-drop table if exists parquet_ctas_table1;
-NOTICE:  table "parquet_ctas_table1" does not exist, skipping
-DROP TABLE if exists parquet_alter cascade;
-NOTICE:  table "parquet_alter" does not exist, skipping
-DROP TABLE if exists parquet_alter_table cascade;
-NOTICE:  table "parquet_alter_table" does not exist, skipping
-Drop Table if exists parquet_alter_table_exch;
-NOTICE:  table "parquet_alter_table_exch" does not exist, skipping
-DROP SCHEMA if exists mytest cascade;
-NOTICE:  schema "mytest" does not exist, skipping
---end_ignore
-CREATE TABLE parquet_table_tobe_ctas(text_col text, bigint_col bigint, char_vary_col character varying(30), numeric_col numeric, int_col int4, float_col float4, before_rename_col int4, change_datatype_col numeric, a_ts_without timestamp without time zone, b_ts_with timestamp with time zone, date_column date, col_set_default numeric) with (appendonly=true, orientation=parquet) DISTRIBUTED RANDOMLY;
-CREATE TABLE parquet_ctas_table1 with (appendonly=true, orientation=parquet) AS SELECT text_col,bigint_col,char_vary_col,numeric_col FROM parquet_table_tobe_ctas;
-CREATE TABLE parquet_alter
-    (id SERIAL,a1 int,a2 char(5),a3 numeric,a4 boolean DEFAULT false ,a5 char DEFAULT 'd',a6 text,a7 timestamp,a8 character varying(705),a9 bigint,a10 date)
-    WITH (appendonly=true, orientation=parquet,compresstype=gzip,compresslevel=1) distributed randomly Partition by range(a1) (start(1)  end(16) every(8)
-    WITH (appendonly=true, orientation=parquet,compresstype=snappy));
-NOTICE:  CREATE TABLE will create implicit sequence "parquet_alter_id_seq" for serial column "parquet_alter.id"
-NOTICE:  CREATE TABLE will create partition "parquet_alter_1_prt_1" for table "parquet_alter"
-NOTICE:  CREATE TABLE will create partition "parquet_alter_1_prt_2" for table "parquet_alter"
-INSERT INTO parquet_alter(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(generate_series(1,5),'M',2011,'t','a','This is news of today: Deadlock between Republicans and Democrats over how best to reduce the U.S. deficit, and over what period, has blocked an agreement to allow the raising of the $14.3 trillion debt ceiling','2001-12-24 02:26:11','U.S. House of Representatives Speaker John Boehner, the top Republican in Congress who has put forward a deficit reduction plan to be voted on later on Thursday said he had no control over whether his bill would avert a credit downgrade.',generate_series(2490,2495),'2011-10-11');
-alter table parquet_alter add partition new_p start(17) end (20) WITH (appendonly=true, orientation=parquet, compresstype=snappy);
-NOTICE:  CREATE TABLE will create partition "parquet_alter_1_prt_new_p" for table "parquet_alter"
-alter table parquet_alter add default partition df_p WITH (appendonly=true, orientation=parquet, compresstype=gzip, compresslevel=3);
-NOTICE:  CREATE TABLE will create partition "parquet_alter_1_prt_df_p" for table "parquet_alter"
-INSERT INTO parquet_alter(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(generate_series(6,25),'F',2010,'f','b','Some students may need time to adjust to school.For most children, the adjustment is quick. Tears will usually disappear after Mommy and  Daddy leave the classroom. Do not plead with your child','2001-12-25 02:22:11','Some students may need time to adjust to school.For most children, the adjustment is quick. Tears will usually disappear after Mommy and  Daddy leave the classroom. Do not plead with your child',generate_series(2500,2501),'2011-10-12');
-ALTER TABLE ONLY parquet_alter RENAME TO parquet_alter_table;
-INSERT INTO parquet_alter_table(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(generate_series(26,28),'F',2010,'f','b','Some students may need time to adjust to school.For most children, the adjustment is quick. Tears will usually disappear after Mommy and  Daddy leave the classroom. Do not plead with your child','2001-12-25 02:22:11','Some students may need time to adjust to school.For most children, the adjustment is quick. Tears will usually disappear after Mommy and  Daddy leave the classroom. Do not plead with your child',generate_series(2502,2503),'2011-10-12');
-select count(*) from parquet_alter_table;
- count 
--------
-    56
-(1 row)
-
-alter table parquet_alter_table rename partition df_p to df_p1;
-alter table parquet_alter_table truncate partition df_p1;
-alter table parquet_alter_table drop partition new_p;
-INSERT INTO parquet_alter_table(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(generate_series(1,28),'M',2010,'f','b','Some students may need time to adjust to school.For most children, the adjustment is quick. Tears will usually disappear after Mommy and  Daddy leave the classroom. Do not plead with your child','2001-12-25 02:22:11','Some students may need time to adjust to school.For most children, the adjustment is quick. Tears will usually disappear after Mommy and  Daddy leave the classroom. Do not plead with your child',2500,'2011-10-12');
-select count(*) from parquet_alter_table;
- count 
--------
-    68
-(1 row)
-
-CREATE TABLE parquet_alter_table_exch(id SERIAL,a1 int,a2 char(5),a3 numeric,a4 boolean DEFAULT false ,a5 char DEFAULT 'd',a6 text,a7 timestamp,a8 character varying(705),a9 bigint,a10 date) WITH (appendonly=true, orientation=parquet, compresstype=snappy)  distributed randomly;
-NOTICE:  CREATE TABLE will create implicit sequence "parquet_alter_table_exch_id_seq" for serial column "parquet_alter_table_exch.id"
-Insert into parquet_alter_table_exch(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(generate_series(29,30),'F',2011,'t','a','dfjjjjjj','2001-12-24 02:26:11','hghgh',333,'2011-10-11');
-Alter table parquet_alter_table exchange default partition with table parquet_alter_table_exch;
-NOTICE:  exchanged partition "df_p1" of relation "parquet_alter_table" with relation "parquet_alter_table_exch"
-Alter table parquet_alter_table split partition FOR (RANK(2)) at(10) into (partition splitc,partition splitd);
-NOTICE:  CREATE TABLE will create partition "parquet_alter_table_1_prt_splitc" for table "parquet_alter_table"
-NOTICE:  CREATE TABLE will create partition "parquet_alter_table_1_prt_splitd" for table "parquet_alter_table"
-INSERT INTO parquet_alter_table(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(generate_series(1,30),'F',2011,'f','b','If the scale of a value to be stored is greater than the declared scale of the column, the system will round the value to the specified number of fractional digits. Then, if the number of digits to the left of the decimal point ','2001-11-24 02:26:11','bbffruikjjjk89kjhjjdsdsjflsdkfjowikfo;lwejkufhekusgfuyewhfkdhsuyfgjsdbfjhkdshgciuvhdskfhiewyerwhkjehriur687687rt3ughjgd67567tcghjzvcnzfTYr7tugfTRE#$#^%*GGHJFTEW#RUYJBJHCFDGJBJGYrythgfT^&^tjhE655ugHD655uVtyr%^uygUYT&^R%^FJYFHGF',2802,'2011-11-12');
-select count(*) from parquet_alter_table;
- count 
--------
-    87
-(1 row)
-
-Create schema mytest;
-ALTER TABLE parquet_alter_table SET SCHEMA mytest;
-\d+ parquet_alter_table
-set search_path=mytest;
-\d+ parquet_alter_table
-                                         Parquet Table "mytest.parquet_alter_table"
- Column |            Type             |                         Modifiers                          | Storage  | Description 
---------+-----------------------------+------------------------------------------------------------+----------+-------------
- id     | integer                     | not null default nextval('parquet_alter_id_seq'::regclass) | plain    | 
- a1     | integer                     |                                                            | plain    | 
- a2     | character(5)                |                                                            | extended | 
- a3     | numeric                     |                                                            | main     | 
- a4     | boolean                     | default false                                              | plain    | 
- a5     | character(1)                | default 'd'::bpchar                                        | extended | 
- a6     | text                        |                                                            | extended | 
- a7     | timestamp without time zone |                                                            | plain    | 
- a8     | character varying(705)      |                                                            | extended | 
- a9     | bigint                      |                                                            | plain    | 
- a10    | date                        |                                                            | plain    | 
-Compression Type: gzip
-Compression Level: 1
-Page Size: 1048576
-RowGroup Size: 8388608
-Checksum: f
-Child tables: public.parquet_alter_table_1_prt_1,
-              public.parquet_alter_table_1_prt_df_p1,
-              public.parquet_alter_table_1_prt_splitc,
-              public.parquet_alter_table_1_prt_splitd
-Has OIDs: no
-Options: appendonly=true, orientation=parquet, compresstype=gzip, compresslevel=1, pagesize=1048576, rowgroupsize=8388608
-Distributed randomly
-Partition by: (a1)
-
-Alter table parquet_alter_table SET DISTRIBUTED BY (a9);
-INSERT INTO parquet_alter_table(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(31,'F',2011,'f','b','If the scale of a value to be stored is greater than the declared scale of the column, the system will round the value to the specified number of fractional digits. Then, if the number of digits to the left of the decimal point ','2001-11-24 02:26:11','bbffruikjjjk89kjhjjdsdsjflsdkfjowikfo;lwejkufhekusgfuyewhfkdhsuyfgjsdbfjhkdshgciuvhdskfhiewyerwhkjehriur687687rt3ughjgd67567tcghjzvcnzfTYr7tugfTRE#$#^%*GGHJFTEW#RUYJBJHCFDGJBJGYrythgfT^&^tjhE655ugHD655uVtyr%^uygUYT&^R%^FJYFHGF',2802,'2011-11-12');
-select count(*) from parquet_alter_table;
- count 
--------
-    88
-(1 row)
-
-Alter table parquet_alter_table SET WITH (REORGANIZE=true);
-INSERT INTO parquet_alter_table(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(32,'F',2011,'f','b','If the scale of a value to be stored is greater than the declared scale of the column, the system will round the value to the specified number of fractional digits. Then, if the number of digits to the left of the decimal point ','2001-11-24 02:26:11','bbffruikjjjk89kjhjjdsdsjflsdkfjowikfo;lwejkufhekusgfuyewhfkdhsuyfgjsdbfjhkdshgciuvhdskfhiewyerwhkjehriur687687rt3ughjgd67567tcghjzvcnzfTYr7tugfTRE#$#^%*GGHJFTEW#RUYJBJHCFDGJBJGYrythgfT^&^tjhE655ugHD655uVtyr%^uygUYT&^R%^FJYFHGF',2802,'2011-11-12');
-select count(*) from parquet_alter_table;
- count 
--------
-    89
-(1 row)
-
-ALTER TABLE parquet_alter_table ADD CONSTRAINT mychk CHECK (a1 <= 35);

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/62f2dfb3/src/test/regress/expected/parquet_multipletype.out
----------------------------------------------------------------------
diff --git a/src/test/regress/expected/parquet_multipletype.out b/src/test/regress/expected/parquet_multipletype.out
deleted file mode 100644
index 53653a9..0000000
--- a/src/test/regress/expected/parquet_multipletype.out
+++ /dev/null
@@ -1,14 +0,0 @@
-drop table if exists multipletype;
-NOTICE:  table "multipletype" does not exist, skipping
-create table multipletype (c0 int4, c1 polygon, c2 text, c3 time, c4 timetz, c5 macaddr, c6 timestamptz, c7 char(10), c8 int2, c9 bool, c10 cidr, c11 circle, c12 lseg, c13 interval, c14 bit, c15 money, c16 box, c17 bytea, c18 xml, c19 bit(5), c20 varchar(10), c21 inet, c22 int8, c23 varbit, c24 serial, c25 float4, c26 point, c27 date, c28 float8 ) WITH (APPENDONLY=true, ORIENTATION=parquet);
-NOTICE:  CREATE TABLE will create implicit sequence "multipletype_c24_seq" for serial column "multipletype.c24"
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'c0' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-insert into multipletype values (2147483647, null, null, '00:00:00', null, 'FF:89:71:45:AE:01', '2000-01-01 08:00:00+09', null, 32767, 'true', '192.168.1.255/32', '<(1,2),3>', '[(0,0),(6,6)]', '-178000000 years', '0', '-21474836.48', '((100,200),(200,400))', null, '<aa>bb</aa>', null, '123456789a', '2001:db8:85a3:8d3:1319:8a2e:370:7344/64', null, null, 1, 0, POINT(1,2), '4277-12-31 AD', 128);
-insert into multipletype values (0, '((100,123),(5,10),(7,2),(4,5))', 'hello world', null, '04:45:05.0012+08:40', null, null, 'bbccddeeff', 128, null, '2001:db8:85a3:8d3:1319:8a2e:370:7344/128', '<(1,2),3>', '[(0,0),(6,6)]', null, '1', '0', '((0,1),(2,3))', 'hello world', '<aa>bb</aa>', null, 'aaaa', '2001:db8:85a3:8d3:1319:8a2e:370:7344/64', 0, null, 2147483647, '-Infinity', POINT(1,2), '4277-12-31 AD', 'Infinity');
-select count(*) from multipletype;
- count 
--------
-     2
-(1 row)
-

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/62f2dfb3/src/test/regress/expected/parquet_pagerowgroup_size.out
----------------------------------------------------------------------
diff --git a/src/test/regress/expected/parquet_pagerowgroup_size.out b/src/test/regress/expected/parquet_pagerowgroup_size.out
deleted file mode 100644
index 624300c..0000000
--- a/src/test/regress/expected/parquet_pagerowgroup_size.out
+++ /dev/null
@@ -1,74 +0,0 @@
--- @description parquet insert vary pagesize/rowgroupsize 
--- @created 2013-08-09 20:33:16
--- @modified 2013-08-09 20:33:16
--- @tags HAWQ parquet
---start_ignore
-drop table if exists pTable1;
-NOTICE:  table "ptable1" does not exist, skipping
-drop table if exists pTable3_from;
-NOTICE:  table "ptable3_from" does not exist, skipping
-drop table if exists pTable3_to;
-NOTICE:  table "ptable3_to" does not exist, skipping
-drop table if exists pTable4_from;
-NOTICE:  table "ptable4_from" does not exist, skipping
-drop table if exists pTable4_to;
-NOTICE:  table "ptable4_to" does not exist, skipping
-drop table if exists pTable5_from;
-NOTICE:  table "ptable5_from" does not exist, skipping
-drop table if exists pTable5_to;
-NOTICE:  table "ptable5_to" does not exist, skipping
---end_ignore
---value/record size equal to pagesize/rowgroupsize
-create table pTable1 (a1 char(10485760), a2 char(10485760), a3 char(10485760), a4 char(10485760), a5 char(10485760), a6 char(10485760), a7 char(10485760), a8 char(10485760), a9 char(10485760), a10 char(10485760)) with(appendonly=true, orientation=parquet, pagesize=10485760, rowgroupsize=104857600);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'a1' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-insert into pTable1 values ( ('a'::char(10485760)), ('a'::char(10485760)), ('a'::char(10485760)), ('a'::char(10485760)), ('a'::char(10485760)), ('a'::char(10485760)), ('a'::char(10485760)), ('a'::char(10485760)), ('a'::char(10485760)), ('a'::char(10485760)) );
-ERROR:  value for column "a1" exceeds pagesize 10485760!  (seg0 mal10Mac:40000 pid=33557)
---single column, one data page contains several values, one rwo group contains several groups
-create table pTable3_from ( a1 text ) with(appendonly=true, orientation=parquet);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'a1' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-insert into pTable3_from values(repeat('parquet',100));
-insert into pTable3_from values(repeat('parquet',20));
-insert into pTable3_from values(repeat('parquet',30));
-create table pTable3_to ( a1 text ) with(appendonly=true, orientation=parquet, pagesize=1024, rowgroupsize=1025);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'a1' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-insert into pTable3_to select * from pTable3_from;
-select count(*) from pTable3_to;
- count 
--------
-     3
-(1 row)
-
---multiple columns, multiple rows combination
-create table pTable4_from ( a1 text , a2 text) with(appendonly=true, orientation=parquet);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'a1' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-insert into pTable4_from values(repeat('parquet',200), repeat('pq',200));
-insert into pTable4_from values(repeat('parquet',50), repeat('pq',200));
-create table pTable4_to ( a1 text, a2 text ) with(appendonly=true, orientation=parquet, pagesize=2048, rowgroupsize=4096);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'a1' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-insert into pTable4_to select * from pTable4_from;
-select count(*) from pTable4_to;
- count 
--------
-     2
-(1 row)
-
---large data insert, several column values in one page, several rows in one rowgroup
-create table pTable5_from (a1 char(1048576), a2 char(2048576), a3 char(3048576), a4 char(4048576), a5 char(5048576), a6 char(6048576), a7 char(7048576), a8 char(8048576), a9 char(9048576), a10 char(9)) with(appendonly=true, orientation=parquet, pagesize=10485760, rowgroupsize=90874386);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'a1' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-insert into pTable5_from values ( ('a'::char(1048576)), ('a'::char(2048576)), ('a'::char(3048576)), ('a'::char(4048576)), ('a'::char(5048576)), ('a'::char(6048576)), ('a'::char(7048576)), ('a'::char(8048576)), ('a'::char(9048576)), ('a'::char(9)) );
-create table pTable5_to (a1 char(1048576), a2 char(2048576), a3 char(3048576), a4 char(4048576), a5 char(5048576), a6 char(6048576), a7 char(7048576), a8 char(8048576), a9 char(9048576), a10 char(9)) with(appendonly=true, orientation=parquet, pagesize=10485760, rowgroupsize=17437200);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column named 'a1' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make sure column(s) chosen are the optimal data distribution key to minimize skew.
-insert into pTable5_to select * from pTable5_from;
-select count(a10) from pTable5_to;
- count 
--------
-     1
-(1 row)
-

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/62f2dfb3/src/test/regress/expected/parquet_subpartition.out
----------------------------------------------------------------------
diff --git a/src/test/regress/expected/parquet_subpartition.out b/src/test/regress/expected/parquet_subpartition.out
deleted file mode 100644
index bf6b97c..0000000
--- a/src/test/regress/expected/parquet_subpartition.out
+++ /dev/null
@@ -1,261 +0,0 @@
---
--- Drop table if exists
---
---start_ignore
-DROP TABLE if exists parquet_wt_subpartgzip7 cascade;
-NOTICE:  table "parquet_wt_subpartgzip7" does not exist, skipping
-DROP TABLE if exists parquet_wt_subpartgzip7_uncompr cascade;
-NOTICE:  table "parquet_wt_subpartgzip7_uncompr" does not exist, skipping
---end_ignore
---
--- Create table
---
-CREATE TABLE parquet_wt_subpartgzip7 
-	(id SERIAL,a1 int,a2 char(5),a3 numeric,a4 boolean DEFAULT false ,a5 char DEFAULT 'd',a6 text,a7 timestamp,a8 character varying(705),a9 bigint,a10 date)
- WITH (appendonly=true, orientation=parquet) distributed randomly  Partition by range(a1) Subpartition by list(a2) subpartition template ( default subpartition df_sp, subpartition sp1 values('M') , subpartition sp2 values('F')  
- WITH (appendonly=true, orientation=parquet,compresstype=gzip,compresslevel=7)) (start(1)  end(5000) every(1000) );
-NOTICE:  CREATE TABLE will create implicit sequence "parquet_wt_subpartgzip7_id_seq" for serial column "parquet_wt_subpartgzip7.id"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_1" for table "parquet_wt_subpartgzip7"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_2" for table "parquet_wt_subpartgzip7"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_3" for table "parquet_wt_subpartgzip7"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_4" for table "parquet_wt_subpartgzip7"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_5" for table "parquet_wt_subpartgzip7"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_1_2_prt_df_sp" for table "parquet_wt_subpartgzip7_1_prt_1"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_1_2_prt_sp1" for table "parquet_wt_subpartgzip7_1_prt_1"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_1_2_prt_sp2" for table "parquet_wt_subpartgzip7_1_prt_1"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_2_2_prt_df_sp" for table "parquet_wt_subpartgzip7_1_prt_2"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_2_2_prt_sp1" for table "parquet_wt_subpartgzip7_1_prt_2"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_2_2_prt_sp2" for table "parquet_wt_subpartgzip7_1_prt_2"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_3_2_prt_df_sp" for table "parquet_wt_subpartgzip7_1_prt_3"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_3_2_prt_sp1" for table "parquet_wt_subpartgzip7_1_prt_3"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_3_2_prt_sp2" for table "parquet_wt_subpartgzip7_1_prt_3"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_4_2_prt_df_sp" for table "parquet_wt_subpartgzip7_1_prt_4"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_4_2_prt_sp1" for table "parquet_wt_subpartgzip7_1_prt_4"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_4_2_prt_sp2" for table "parquet_wt_subpartgzip7_1_prt_4"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_5_2_prt_df_sp" for table "parquet_wt_subpartgzip7_1_prt_5"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_5_2_prt_sp1" for table "parquet_wt_subpartgzip7_1_prt_5"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_5_2_prt_sp2" for table "parquet_wt_subpartgzip7_1_prt_5"
---
--- Insert data to the table
---
- INSERT INTO parquet_wt_subpartgzip7(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(generate_series(1,20),'M',2011,'t','a','This is news of today: Deadlock between Republicans and Democrats over how best to reduce the U.S. deficit, and over what period, has blocked an agreement to allow the raising of the $14.3 trillion debt ceiling','2001-12-24 02:26:11','U.S. House of Representatives Speaker John Boehner, the top Republican in Congress who has put forward a deficit reduction plan to be voted on later on Thursday said he had no control over whether his bill would avert a credit downgrade.',generate_series(2490,2505),'2011-10-11'); 
- INSERT INTO parquet_wt_subpartgzip7(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(generate_series(500,510),'F',2010,'f','b','Some students may need time to adjust to school.For most children, the adjustment is quick. Tears will usually disappear after Mommy and  Daddy leave the classroom. Do not plead with your child','2001-12-25 02:22:11','Some students may need time to adjust to school.For most children, the adjustment is quick. Tears will usually disappear after Mommy and  Daddy leave the classroom. Do not plead with your child',generate_series(2500,2516),'2011-10-12'); 
---Create Uncompressed table of same schema definition
-CREATE TABLE parquet_wt_subpartgzip7_uncompr(id SERIAL,a1 int,a2 char(5),a3 numeric,a4 boolean DEFAULT false ,a5 char DEFAULT 'd',a6 text,a7 timestamp,a8 character varying(705),a9 bigint,a10 date) WITH (appendonly=true, orientation=parquet) distributed randomly Partition by range(a1) Subpartition by list(a2) subpartition template ( subpartition sp1 values('M') , subpartition sp2 values('F') ) (start(1)  end(5000) every(1000)) ;
-NOTICE:  CREATE TABLE will create implicit sequence "parquet_wt_subpartgzip7_uncompr_id_seq" for serial column "parquet_wt_subpartgzip7_uncompr.id"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_uncompr_1_prt_1" for table "parquet_wt_subpartgzip7_uncompr"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_uncompr_1_prt_2" for table "parquet_wt_subpartgzip7_uncompr"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_uncompr_1_prt_3" for table "parquet_wt_subpartgzip7_uncompr"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_uncompr_1_prt_4" for table "parquet_wt_subpartgzip7_uncompr"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_uncompr_1_prt_5" for table "parquet_wt_subpartgzip7_uncompr"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_uncompr_1_prt_1_2_prt_sp1" for table "parquet_wt_subpartgzip7_uncompr_1_prt_1"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_uncompr_1_prt_1_2_prt_sp2" for table "parquet_wt_subpartgzip7_uncompr_1_prt_1"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_uncompr_1_prt_2_2_prt_sp1" for table "parquet_wt_subpartgzip7_uncompr_1_prt_2"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_uncompr_1_prt_2_2_prt_sp2" for table "parquet_wt_subpartgzip7_uncompr_1_prt_2"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_uncompr_1_prt_3_2_prt_sp1" for table "parquet_wt_subpartgzip7_uncompr_1_prt_3"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_uncompr_1_prt_3_2_prt_sp2" for table "parquet_wt_subpartgzip7_uncompr_1_prt_3"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_uncompr_1_prt_4_2_prt_sp1" for table "parquet_wt_subpartgzip7_uncompr_1_prt_4"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_uncompr_1_prt_4_2_prt_sp2" for table "parquet_wt_subpartgzip7_uncompr_1_prt_4"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_uncompr_1_prt_5_2_prt_sp1" for table "parquet_wt_subpartgzip7_uncompr_1_prt_5"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_uncompr_1_prt_5_2_prt_sp2" for table "parquet_wt_subpartgzip7_uncompr_1_prt_5"
---
--- Insert to uncompressed table
---
- INSERT INTO parquet_wt_subpartgzip7_uncompr(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(generate_series(1,20),'M',2011,'t','a','This is news of today: Deadlock between Republicans and Democrats over how best to reduce the U.S. deficit, and over what period, has blocked an agreement to allow the raising of the $14.3 trillion debt ceiling','2001-12-24 02:26:11','U.S. House of Representatives Speaker John Boehner, the top Republican in Congress who has put forward a deficit reduction plan to be voted on later on Thursday said he had no control over whether his bill would avert a credit downgrade.',generate_series(2490,2505),'2011-10-11'); 
- INSERT INTO parquet_wt_subpartgzip7_uncompr(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(generate_series(500,510),'F',2010,'f','b','Some students may need time to adjust to school.For most children, the adjustment is quick. Tears will usually disappear after Mommy and  Daddy leave the classroom. Do not plead with your child','2001-12-25 02:22:11','Some students may need time to adjust to school.For most children, the adjustment is quick. Tears will usually disappear after Mommy and  Daddy leave the classroom. Do not plead with your child',generate_series(2500,2516),'2011-10-12'); 
---
--- ********Validation******* 
---
-\d+ parquet_wt_subpartgzip7_1_prt_1_2_prt_sp2
-                                   Parquet Table "public.parquet_wt_subpartgzip7_1_prt_1_2_prt_sp2"
- Column |            Type             |                              Modifiers                               | Storage  | Description 
---------+-----------------------------+----------------------------------------------------------------------+----------+-------------
- id     | integer                     | not null default nextval('parquet_wt_subpartgzip7_id_seq'::regclass) | plain    | 
- a1     | integer                     |                                                                      | plain    | 
- a2     | character(5)                |                                                                      | extended | 
- a3     | numeric                     |                                                                      | main     | 
- a4     | boolean                     | default false                                                        | plain    | 
- a5     | character(1)                | default 'd'::bpchar                                                  | extended | 
- a6     | text                        |                                                                      | extended | 
- a7     | timestamp without time zone |                                                                      | plain    | 
- a8     | character varying(705)      |                                                                      | extended | 
- a9     | bigint                      |                                                                      | plain    | 
- a10    | date                        |                                                                      | plain    | 
-Compression Type: gzip
-Compression Level: 7
-Page Size: 1048576
-RowGroup Size: 8388608
-Checksum: f
-Check constraints:
-    "parquet_wt_subpartgzip7_1_prt_1_2_prt_sp2_check" CHECK (a2 = 'F    '::bpchar)
-    "parquet_wt_subpartgzip7_1_prt_1_check" CHECK (a1 >= 1 AND a1 < 1001)
-Inherits: parquet_wt_subpartgzip7_1_prt_1
-Has OIDs: no
-Options: appendonly=true, orientation=parquet, compresstype=gzip, compresslevel=7, pagesize=1048576, rowgroupsize=8388608
-Distributed randomly
-
---
--- Compression ratio
---
---select 'compression_ratio', get_ao_compression_ratio('parquet_wt_subpartgzip7_1_prt_1_2_prt_sp2'); 
---Select from pg_attribute_encoding to see the table entry 
-select parencattnum, parencattoptions from pg_partition_encoding e, pg_partition p, pg_class c  where c.relname = 'parquet_wt_subpartgzip7' and c.oid = p.parrelid and p.oid = e.parencoid order by parencattnum limit 3; 
- parencattnum | parencattoptions 
---------------+------------------
-(0 rows)
-
---
--- Compare data with uncompressed table
---
---
--- Select number of rows from the uncompressed table 
---
-SELECT count(*) as count_uncompressed from  parquet_wt_subpartgzip7_uncompr ;
- count_uncompressed 
---------------------
-                267
-(1 row)
-
---
--- Select number of rows from the compressed table 
---
-SELECT count(*) as count_compressed from  parquet_wt_subpartgzip7;
- count_compressed 
-------------------
-              267
-(1 row)
-
---
--- Select number of rows using a FULL outer join on all the columns of the two tables 
--- Count should match with above result if the all the rows uncompressed correctly: 
---
-Select count(*) as count_join from parquet_wt_subpartgzip7 t1 full outer join parquet_wt_subpartgzip7_uncompr t2 on t1.id=t2.id and t1.a1=t2.a1 and t1.a2=t2.a2 and t1.a3=t2.a3 and t1.a4=t2.a4 and t1.a5=t2.a5 and t1.a6=t2.a6 and t1.a7=t2.a7 and t1.a8=t2.a8 and t1.a9=t2.a9 and t1.a10=t2.a10;
- count_join 
-------------
-        267
-(1 row)
-
---
--- Truncate the table 
---
-TRUNCATE table parquet_wt_subpartgzip7;
---
--- Insert data again 
---
-insert into parquet_wt_subpartgzip7 select * from parquet_wt_subpartgzip7_uncompr order by a1;
---Alter table Add Partition 
-alter table parquet_wt_subpartgzip7 add partition new_p start(5050) end (6051) WITH (appendonly=true, orientation=parquet, compresstype=gzip, compresslevel=1);
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_new_p" for table "parquet_wt_subpartgzip7"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_new_p_2_prt_sp1" for table "parquet_wt_subpartgzip7_1_prt_new_p"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_new_p_2_prt_sp2" for table "parquet_wt_subpartgzip7_1_prt_new_p"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_new_p_2_prt_df_sp" for table "parquet_wt_subpartgzip7_1_prt_new_p"
---Validation with psql utility 
-  \d+ parquet_wt_subpartgzip7_1_prt_new_p_2_prt_sp1
-                                 Parquet Table "public.parquet_wt_subpartgzip7_1_prt_new_p_2_prt_sp1"
- Column |            Type             |                              Modifiers                               | Storage  | Description 
---------+-----------------------------+----------------------------------------------------------------------+----------+-------------
- id     | integer                     | not null default nextval('parquet_wt_subpartgzip7_id_seq'::regclass) | plain    | 
- a1     | integer                     |                                                                      | plain    | 
- a2     | character(5)                |                                                                      | extended | 
- a3     | numeric                     |                                                                      | main     | 
- a4     | boolean                     | default false                                                        | plain    | 
- a5     | character(1)                | default 'd'::bpchar                                                  | extended | 
- a6     | text                        |                                                                      | extended | 
- a7     | timestamp without time zone |                                                                      | plain    | 
- a8     | character varying(705)      |                                                                      | extended | 
- a9     | bigint                      |                                                                      | plain    | 
- a10    | date                        |                                                                      | plain    | 
-Compression Type: gzip
-Compression Level: 1
-Page Size: 1048576
-RowGroup Size: 8388608
-Checksum: f
-Check constraints:
-    "parquet_wt_subpartgzip7_1_prt_new_p_2_prt_sp1_check" CHECK (a2 = 'M    '::bpchar)
-    "parquet_wt_subpartgzip7_1_prt_new_p_check" CHECK (a1 >= 5050 AND a1 < 6051)
-Inherits: parquet_wt_subpartgzip7_1_prt_new_p
-Has OIDs: no
-Options: appendonly=true, orientation=parquet, compresstype=gzip, compresslevel=1, pagesize=1048576, rowgroupsize=8388608
-Distributed randomly
-
-alter table parquet_wt_subpartgzip7 add default partition df_p ;
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_df_p" for table "parquet_wt_subpartgzip7"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_df_p_2_prt_sp1" for table "parquet_wt_subpartgzip7_1_prt_df_p"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_df_p_2_prt_sp2" for table "parquet_wt_subpartgzip7_1_prt_df_p"
-NOTICE:  CREATE TABLE will create partition "parquet_wt_subpartgzip7_1_prt_df_p_2_prt_df_sp" for table "parquet_wt_subpartgzip7_1_prt_df_p"
---Validation with psql utility 
-  \d+ parquet_wt_subpartgzip7_1_prt_df_p_2_prt_sp2
-                                 Parquet Table "public.parquet_wt_subpartgzip7_1_prt_df_p_2_prt_sp2"
- Column |            Type             |                              Modifiers                               | Storage  | Description 
---------+-----------------------------+----------------------------------------------------------------------+----------+-------------
- id     | integer                     | not null default nextval('parquet_wt_subpartgzip7_id_seq'::regclass) | plain    | 
- a1     | integer                     |                                                                      | plain    | 
- a2     | character(5)                |                                                                      | extended | 
- a3     | numeric                     |                                                                      | main     | 
- a4     | boolean                     | default false                                                        | plain    | 
- a5     | character(1)                | default 'd'::bpchar                                                  | extended | 
- a6     | text                        |                                                                      | extended | 
- a7     | timestamp without time zone |                                                                      | plain    | 
- a8     | character varying(705)      |                                                                      | extended | 
- a9     | bigint                      |                                                                      | plain    | 
- a10    | date                        |                                                                      | plain    | 
-Compression Type: gzip
-Compression Level: 7
-Page Size: 1048576
-RowGroup Size: 8388608
-Checksum: f
-Check constraints:
-    "parquet_wt_subpartgzip7_1_prt_df_p_2_prt_sp2_check" CHECK (a2 = 'F    '::bpchar)
-Inherits: parquet_wt_subpartgzip7_1_prt_df_p
-Has OIDs: no
-Options: appendonly=true, orientation=parquet, compresstype=gzip, compresslevel=7, pagesize=1048576, rowgroupsize=8388608
-Distributed randomly
-
---
--- Compression ratio
---
---select 'compression_ratio', get_ao_compression_ratio('parquet_wt_subpartgzip7_1_prt_new_p_2_prt_sp1'); 
---Alter table Exchange Partition 
---Create a table to use in exchange partition 
-Drop Table if exists parquet_wt_subpartgzip7_exch; 
-NOTICE:  table "parquet_wt_subpartgzip7_exch" does not exist, skipping
- CREATE TABLE parquet_wt_subpartgzip7_exch(id SERIAL,a1 int,a2 char(5),a3 numeric,a4 boolean DEFAULT false ,a5 char DEFAULT 'd',a6 text,a7 timestamp,a8 character varying(705),a9 bigint,a10 date) WITH (appendonly=true, orientation=parquet, compresstype=gzip)  distributed randomly;
-NOTICE:  CREATE TABLE will create implicit sequence "parquet_wt_subpartgzip7_exch_id_seq" for serial column "parquet_wt_subpartgzip7_exch.id"
- 
-Drop Table if exists parquet_wt_subpartgzip7_defexch; 
-NOTICE:  table "parquet_wt_subpartgzip7_defexch" does not exist, skipping
- CREATE TABLE parquet_wt_subpartgzip7_defexch(id SERIAL,a1 int,a2 char(5),a3 numeric,a4 boolean DEFAULT false ,a5 char DEFAULT 'd',a6 text,a7 timestamp,a8 character varying(705),a9 bigint,a10 date) WITH (appendonly=true, orientation=parquet, compresstype=gzip)  distributed randomly;
-NOTICE:  CREATE TABLE will create implicit sequence "parquet_wt_subpartgzip7_defexch_id_seq" for serial column "parquet_wt_subpartgzip7_defexch.id"
- 
-Insert into parquet_wt_subpartgzip7_exch(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) select a1,a2,a3,a4,a5,a6,a7,a8,a9,a10 from parquet_wt_subpartgzip7 where  a1=10 and a2!='C';
-Insert into parquet_wt_subpartgzip7_defexch(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) select a1,a2,a3,a4,a5,a6,a7,a8,a9,a10 from parquet_wt_subpartgzip7 where a1 =10 and a2!='C';
-Alter table parquet_wt_subpartgzip7 alter partition FOR (RANK(1)) exchange partition sp1 with table parquet_wt_subpartgzip7_exch;
-NOTICE:  exchanged partition "sp1" of partition for rank 1 of relation "parquet_wt_subpartgzip7" with relation "parquet_wt_subpartgzip7_exch"
-\d+ parquet_wt_subpartgzip7_1_prt_1_2_prt_sp1
-                                     Parquet Table "public.parquet_wt_subpartgzip7_1_prt_1_2_prt_sp1"
- Column |            Type             |                                 Modifiers                                 | Storage  | Description 
---------+-----------------------------+---------------------------------------------------------------------------+----------+-------------
- id     | integer                     | not null default nextval('parquet_wt_subpartgzip7_exch_id_seq'::regclass) | plain    | 
- a1     | integer                     |                                                                           | plain    | 
- a2     | character(5)                |                                                                           | extended | 
- a3     | numeric                     |                                                                           | main     | 
- a4     | boolean                     | default false                                                             | plain    | 
- a5     | character(1)                | default 'd'::bpchar                                                       | extended | 
- a6     | text                        |                                                                           | extended | 
- a7     | timestamp without time zone |                                                                           | plain    | 
- a8     | character varying(705)      |                                                                           | extended | 
- a9     | bigint                      |                                                                           | plain    | 
- a10    | date                        |                                                                           | plain    | 
-Compression Type: gzip
-Compression Level: 1
-Page Size: 1048576
-RowGroup Size: 8388608
-Checksum: f
-Check constraints:
-    "parquet_wt_subpartgzip7_1_prt_1_2_prt_sp1_check" CHECK (a2 = 'M    '::bpchar)
-    "parquet_wt_subpartgzip7_1_prt_1_check" CHECK (a1 >= 1 AND a1 < 1001)
-Inherits: parquet_wt_subpartgzip7_1_prt_1
-Has OIDs: no
-Options: appendonly=true, orientation=parquet, compresstype=gzip
-Distributed randomly
-

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/62f2dfb3/src/test/regress/known_good_schedule
----------------------------------------------------------------------
diff --git a/src/test/regress/known_good_schedule b/src/test/regress/known_good_schedule
index a0fd52e..eb105d8 100755
--- a/src/test/regress/known_good_schedule
+++ b/src/test/regress/known_good_schedule
@@ -210,11 +210,6 @@ ignore: orca_udfs
 #
 ignore: goh_tablespace
 ignore: hdfs_filesystem
-test: parquet_ddl
-test: parquet_multipletype
-test: parquet_pagerowgroup_size
-test: parquet_compression
-test: parquet_subpartition
 ignore: co_disabled
 # HCatalog tests
 test: caqlinmem

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/62f2dfb3/src/test/regress/sql/parquet_compression.sql
----------------------------------------------------------------------
diff --git a/src/test/regress/sql/parquet_compression.sql b/src/test/regress/sql/parquet_compression.sql
deleted file mode 100644
index 0544fce..0000000
--- a/src/test/regress/sql/parquet_compression.sql
+++ /dev/null
@@ -1,124 +0,0 @@
--- @description parquet compression test
--- @created 2013-08-09 19:57:38
--- @modified 2013-08-09 19:57:38
--- @tags HAWQ parquet
-
---start_ignore
-drop schema parquet cascade;
-create schema parquet;
-set search_path=parquet,"$user",public;
-set default_segment_num = 4;
-
-drop table parquet_gzip;
-drop table parquet_gzip_uncompr;
-drop table parquet_gzip_part;
-drop table parquet_gzip_part_unc;
-drop table parquet_snappy;
-drop table parquet_snappy_uncompr;
-drop table parquet_snappy_part;
-drop table parquet_snappy_part_unc;
-drop table parquet_gzip_2;
-
-alter resource queue pg_default with ( vseg_resource_quota='mem:4gb');
---end_ignore
-
---Datatypes covered: text,bytea,varchar,bit varying
-
--- parquet table ,compresstype = gzip
-
-create table parquet_gzip (i int,document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, compresstype=gzip, pagesize=150000000, rowgroupsize= 500000000);
-
-insert into parquet_gzip  values(1001,repeat('large data value for text data type',32732),repeat('large data value for text data type',32732),pg_catalog.decode(repeat('large data value for text data type',32732), 'escape'),1000::int::bit(3244447)::varbit);
-
--- Uncompressed table and join with compressed to see if select works
-
-create table parquet_gzip_uncompr(i int,document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, pagesize=150000000, rowgroupsize= 500000000);
-
-insert into parquet_gzip_uncompr values(1001,repeat('large data value for text data type',32732),repeat('large data value for text data type',32732),pg_catalog.decode(repeat('large data value for text data type',32732), 'escape'),1000::int::bit(3244447)::varbit);
-
-select count(*) from parquet_gzip c1 full outer join parquet_gzip_uncompr c2 on c1.document=c2.document and c1.vch1=c2.vch1 and c1.bta1=c2.bta1 and c1.bitv1=c2.bitv1;
-
---select get_ao_compression_ratio('parquet_gzip');
-
--- parquet partition table with hybrid partitions
-
-create table parquet_gzip_part (p1 int,document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, compresstype=gzip, pagesize=150000000, rowgroupsize= 500000000)
-      Partition by range(p1)
-      (Partition p1 start(1) end(4) with (appendonly=true),
-       Partition p2 start(4) end(7) with (appendonly=true),
-       Partition p3 start(7) end(11) with (appendonly=true,orientation=parquet,compresstype=gzip, pagesize=150000000, rowgroupsize= 500000000));
-
-insert into parquet_gzip_part values(generate_series(1,5), repeat('large data value for text data type',32732),repeat('large data value for text data type',32732),pg_catalog.decode(repeat('large data value for text data type',32732), 'escape'),1000::int::bit(3244447)::varbit);
-
--- Uncompressed table and join with compressed to see if select works
-
-create table parquet_gzip_part_unc (p1 int,document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, pagesize=150000000, rowgroupsize= 500000000)
-      Partition by range(p1)
-      (Partition p1 start(1) end(4) with (appendonly=true),
-       Partition p2 start(4) end(7) with (appendonly=true),
-       Partition p3 start(7) end(11) with (appendonly=true,orientation=parquet, pagesize=150000000, rowgroupsize= 500000000));
-
-insert into parquet_gzip_part_unc values(generate_series(1,5), repeat('large data value for text data type',32732),repeat('large data value for text data type',32732),pg_catalog.decode(repeat('large data value for text data type',32732), 'escape'),1000::int::bit(3244447)::varbit);
-
-select count(*) from parquet_gzip_part c1 full outer join parquet_gzip_part_unc c2 on c1.p1=c2.p1 and c1.document=c2.document and c1.vch1=c2.vch1 and c1.bta1=c2.bta1 and c1.bitv1=c2.bitv1;
-
---select get_ao_compression_ratio('parquet_gzip_part_1_prt_p3');
-
-
--- parquet table , compresstype = snappy
-
-create table parquet_snappy (p1 int, document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, compresstype=snappy, pagesize=150000000, rowgroupsize= 500000000);
-
-insert into parquet_snappy values(1001, repeat('large data value for text data type',32732),repeat('large data value for text data type',32732),pg_catalog.decode(repeat('large data value for text data type',32732), 'escape'),1000::int::bit(3244447)::varbit);
-
--- Uncompressed table and join with compressed to see if select works
-
-create table parquet_snappy_uncompr(p1 int, document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, pagesize=150000000, rowgroupsize= 500000000);
-
-insert into parquet_snappy_uncompr values(1001,repeat('large data value for text data type',32732),repeat('large data value for text data type',32732),pg_catalog.decode(repeat('large data value for text data type',32732), 'escape'),1000::int::bit(3244447)::varbit);
-
-select count(*) from parquet_snappy c1 full outer join parquet_snappy_uncompr c2 on c1.p1=c2.p1 and c1.document=c2.document and c1.vch1=c2.vch1 and c1.bta1=c2.bta1 and c1.bitv1=c2.bitv1;
-
---select get_ao_compression_ratio('parquet_snappy');
-
--- parquet partition table with hybrid partitions
-
-
-create table parquet_snappy_part (p1 int,document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, compresstype=snappy, pagesize=150000000, rowgroupsize= 500000000)
-      Partition by range(p1)
-      (Partition p1 start(1) end(4) with (appendonly=true),
-       Partition p2 start(4) end(7) with (appendonly=true),
-       Partition p3 start(7) end(11) with (appendonly=true,orientation=parquet,compresstype=snappy, pagesize=150000000, rowgroupsize= 500000000));
-
-insert into parquet_snappy_part values(generate_series(1,5), repeat('large data value for text data type',32732),repeat('large data value for text data type',32732),pg_catalog.decode(repeat('large data value for text data type',32732), 'escape'),1000::int::bit(3244447)::varbit);
-
--- Uncompressed table and join with compressed to see if select works
-
-create table parquet_snappy_part_unc (p1 int,document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, pagesize=150000000, rowgroupsize= 500000000)
-      Partition by range(p1)
-      (Partition p1 start(1) end(4) with (appendonly=true),
-       Partition p2 start(4) end(7) with (appendonly=true),
-       Partition p3 start(7) end(11) with (appendonly=true,orientation=parquet, pagesize=150000000, rowgroupsize= 500000000));
-
-insert into parquet_snappy_part_unc values(generate_series(1,5), repeat('large data value for text data type',32732),repeat('large data value for text data type',32732),pg_catalog.decode(repeat('large data value for text data type',32732), 'escape'),1000::int::bit(3244447)::varbit);
-
-select count(*) from parquet_snappy_part c1 full outer join parquet_snappy_part_unc c2 on c1.p1=c2.p1 and c1.document=c2.document and c1.vch1=c2.vch1 and c1.bta1=c2.bta1 and c1.bitv1=c2.bitv1;
-
---select get_ao_compression_ratio('parquet_snappy_part_1_prt_p3');
-create table parquet_gzip_2 (i int,document text not null, vch1 varchar,bta1 bytea, bitv1 varbit)
-      WITH (appendonly=true, orientation=parquet, compresstype=gzip, pagesize=150000000, rowgroupsize= 500000000);
-
-insert into parquet_gzip_2 values(12,array_to_string(ARRAY(SELECT chr((65 + round(random() * 25)) :: integer) FROM generate_series(1, 32732)), ''),array_to_string(ARRAY(SELECT chr((65 + round(random() * 25)) :: integer) FROM generate_series(1, 32732)), ''),pg_catalog.decode(array_to_string(ARRAY(SELECT chr((65 + round(random() * 25)) :: integer) FROM generate_series(1, 32732)), ''),'escape'),1010110::int::bit(3244447)::varbit);
-
-
-Select count(*) from parquet_gzip_2;
-
-alter resource queue pg_default with ( vseg_resource_quota='mem:256mb');

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/62f2dfb3/src/test/regress/sql/parquet_ddl.sql
----------------------------------------------------------------------
diff --git a/src/test/regress/sql/parquet_ddl.sql b/src/test/regress/sql/parquet_ddl.sql
deleted file mode 100644
index 2c64bc8..0000000
--- a/src/test/regress/sql/parquet_ddl.sql
+++ /dev/null
@@ -1,75 +0,0 @@
---start_ignore
-drop table if exists parquet_table_tobe_ctas;
-drop table if exists parquet_ctas_table1;
-DROP TABLE if exists parquet_alter cascade;
-DROP TABLE if exists parquet_alter_table cascade;
-Drop Table if exists parquet_alter_table_exch;
-DROP SCHEMA if exists mytest cascade;
---end_ignore
-
-CREATE TABLE parquet_table_tobe_ctas(text_col text, bigint_col bigint, char_vary_col character varying(30), numeric_col numeric, int_col int4, float_col float4, before_rename_col int4, change_datatype_col numeric, a_ts_without timestamp without time zone, b_ts_with timestamp with time zone, date_column date, col_set_default numeric) with (appendonly=true, orientation=parquet) DISTRIBUTED RANDOMLY;
-
-CREATE TABLE parquet_ctas_table1 with (appendonly=true, orientation=parquet) AS SELECT text_col,bigint_col,char_vary_col,numeric_col FROM parquet_table_tobe_ctas;
-
-
-CREATE TABLE parquet_alter
-    (id SERIAL,a1 int,a2 char(5),a3 numeric,a4 boolean DEFAULT false ,a5 char DEFAULT 'd',a6 text,a7 timestamp,a8 character varying(705),a9 bigint,a10 date)
-    WITH (appendonly=true, orientation=parquet,compresstype=gzip,compresslevel=1) distributed randomly Partition by range(a1) (start(1)  end(16) every(8)
-    WITH (appendonly=true, orientation=parquet,compresstype=snappy));
-
-
-INSERT INTO parquet_alter(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(generate_series(1,5),'M',2011,'t','a','This is news of today: Deadlock between Republicans and Democrats over how best to reduce the U.S. deficit, and over what period, has blocked an agreement to allow the raising of the $14.3 trillion debt ceiling','2001-12-24 02:26:11','U.S. House of Representatives Speaker John Boehner, the top Republican in Congress who has put forward a deficit reduction plan to be voted on later on Thursday said he had no control over whether his bill would avert a credit downgrade.',generate_series(2490,2495),'2011-10-11');
-
-alter table parquet_alter add partition new_p start(17) end (20) WITH (appendonly=true, orientation=parquet, compresstype=snappy);
-
-alter table parquet_alter add default partition df_p WITH (appendonly=true, orientation=parquet, compresstype=gzip, compresslevel=3);
-
-INSERT INTO parquet_alter(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(generate_series(6,25),'F',2010,'f','b','Some students may need time to adjust to school.For most children, the adjustment is quick. Tears will usually disappear after Mommy and  Daddy leave the classroom. Do not plead with your child','2001-12-25 02:22:11','Some students may need time to adjust to school.For most children, the adjustment is quick. Tears will usually disappear after Mommy and  Daddy leave the classroom. Do not plead with your child',generate_series(2500,2501),'2011-10-12');
-
-ALTER TABLE ONLY parquet_alter RENAME TO parquet_alter_table;
-
-INSERT INTO parquet_alter_table(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(generate_series(26,28),'F',2010,'f','b','Some students may need time to adjust to school.For most children, the adjustment is quick. Tears will usually disappear after Mommy and  Daddy leave the classroom. Do not plead with your child','2001-12-25 02:22:11','Some students may need time to adjust to school.For most children, the adjustment is quick. Tears will usually disappear after Mommy and  Daddy leave the classroom. Do not plead with your child',generate_series(2502,2503),'2011-10-12');
-
-select count(*) from parquet_alter_table;
-
-alter table parquet_alter_table rename partition df_p to df_p1;
-
-alter table parquet_alter_table truncate partition df_p1;
-
-alter table parquet_alter_table drop partition new_p;
-
-INSERT INTO parquet_alter_table(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(generate_series(1,28),'M',2010,'f','b','Some students may need time to adjust to school.For most children, the adjustment is quick. Tears will usually disappear after Mommy and  Daddy leave the classroom. Do not plead with your child','2001-12-25 02:22:11','Some students may need time to adjust to school.For most children, the adjustment is quick. Tears will usually disappear after Mommy and  Daddy leave the classroom. Do not plead with your child',2500,'2011-10-12');
-
-select count(*) from parquet_alter_table;
-
-CREATE TABLE parquet_alter_table_exch(id SERIAL,a1 int,a2 char(5),a3 numeric,a4 boolean DEFAULT false ,a5 char DEFAULT 'd',a6 text,a7 timestamp,a8 character varying(705),a9 bigint,a10 date) WITH (appendonly=true, orientation=parquet, compresstype=snappy)  distributed randomly;
-
-Insert into parquet_alter_table_exch(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(generate_series(29,30),'F',2011,'t','a','dfjjjjjj','2001-12-24 02:26:11','hghgh',333,'2011-10-11');
-
-Alter table parquet_alter_table exchange default partition with table parquet_alter_table_exch;
-
-Alter table parquet_alter_table split partition FOR (RANK(2)) at(10) into (partition splitc,partition splitd);
-
-INSERT INTO parquet_alter_table(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(generate_series(1,30),'F',2011,'f','b','If the scale of a value to be stored is greater than the declared scale of the column, the system will round the value to the specified number of fractional digits. Then, if the number of digits to the left of the decimal point ','2001-11-24 02:26:11','bbffruikjjjk89kjhjjdsdsjflsdkfjowikfo;lwejkufhekusgfuyewhfkdhsuyfgjsdbfjhkdshgciuvhdskfhiewyerwhkjehriur687687rt3ughjgd67567tcghjzvcnzfTYr7tugfTRE#$#^%*GGHJFTEW#RUYJBJHCFDGJBJGYrythgfT^&^tjhE655ugHD655uVtyr%^uygUYT&^R%^FJYFHGF',2802,'2011-11-12');
-
-select count(*) from parquet_alter_table;
-
-Create schema mytest;
-ALTER TABLE parquet_alter_table SET SCHEMA mytest;
-\d+ parquet_alter_table
-set search_path=mytest;
-\d+ parquet_alter_table
-
-Alter table parquet_alter_table SET DISTRIBUTED BY (a9);
-INSERT INTO parquet_alter_table(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(31,'F',2011,'f','b','If the scale of a value to be stored is greater than the declared scale of the column, the system will round the value to the specified number of fractional digits. Then, if the number of digits to the left of the decimal point ','2001-11-24 02:26:11','bbffruikjjjk89kjhjjdsdsjflsdkfjowikfo;lwejkufhekusgfuyewhfkdhsuyfgjsdbfjhkdshgciuvhdskfhiewyerwhkjehriur687687rt3ughjgd67567tcghjzvcnzfTYr7tugfTRE#$#^%*GGHJFTEW#RUYJBJHCFDGJBJGYrythgfT^&^tjhE655ugHD655uVtyr%^uygUYT&^R%^FJYFHGF',2802,'2011-11-12');
-
-select count(*) from parquet_alter_table;
-
-Alter table parquet_alter_table SET WITH (REORGANIZE=true);
-
-INSERT INTO parquet_alter_table(a1,a2,a3,a4,a5,a6,a7,a8,a9,a10) values(32,'F',2011,'f','b','If the scale of a value to be stored is greater than the declared scale of the column, the system will round the value to the specified number of fractional digits. Then, if the number of digits to the left of the decimal point ','2001-11-24 02:26:11','bbffruikjjjk89kjhjjdsdsjflsdkfjowikfo;lwejkufhekusgfuyewhfkdhsuyfgjsdbfjhkdshgciuvhdskfhiewyerwhkjehriur687687rt3ughjgd67567tcghjzvcnzfTYr7tugfTRE#$#^%*GGHJFTEW#RUYJBJHCFDGJBJGYrythgfT^&^tjhE655ugHD655uVtyr%^uygUYT&^R%^FJYFHGF',2802,'2011-11-12');
-
-select count(*) from parquet_alter_table;
-
-ALTER TABLE parquet_alter_table ADD CONSTRAINT mychk CHECK (a1 <= 35);
-

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/62f2dfb3/src/test/regress/sql/parquet_multipletype.sql
----------------------------------------------------------------------
diff --git a/src/test/regress/sql/parquet_multipletype.sql b/src/test/regress/sql/parquet_multipletype.sql
deleted file mode 100644
index 885f794..0000000
--- a/src/test/regress/sql/parquet_multipletype.sql
+++ /dev/null
@@ -1,5 +0,0 @@
-drop table if exists multipletype;
-create table multipletype (c0 int4, c1 polygon, c2 text, c3 time, c4 timetz, c5 macaddr, c6 timestamptz, c7 char(10), c8 int2, c9 bool, c10 cidr, c11 circle, c12 lseg, c13 interval, c14 bit, c15 money, c16 box, c17 bytea, c18 xml, c19 bit(5), c20 varchar(10), c21 inet, c22 int8, c23 varbit, c24 serial, c25 float4, c26 point, c27 date, c28 float8 ) WITH (APPENDONLY=true, ORIENTATION=parquet);
-insert into multipletype values (2147483647, null, null, '00:00:00', null, 'FF:89:71:45:AE:01', '2000-01-01 08:00:00+09', null, 32767, 'true', '192.168.1.255/32', '<(1,2),3>', '[(0,0),(6,6)]', '-178000000 years', '0', '-21474836.48', '((100,200),(200,400))', null, '<aa>bb</aa>', null, '123456789a', '2001:db8:85a3:8d3:1319:8a2e:370:7344/64', null, null, 1, 0, POINT(1,2), '4277-12-31 AD', 128);
-insert into multipletype values (0, '((100,123),(5,10),(7,2),(4,5))', 'hello world', null, '04:45:05.0012+08:40', null, null, 'bbccddeeff', 128, null, '2001:db8:85a3:8d3:1319:8a2e:370:7344/128', '<(1,2),3>', '[(0,0),(6,6)]', null, '1', '0', '((0,1),(2,3))', 'hello world', '<aa>bb</aa>', null, 'aaaa', '2001:db8:85a3:8d3:1319:8a2e:370:7344/64', 0, null, 2147483647, '-Infinity', POINT(1,2), '4277-12-31 AD', 'Infinity');
-select count(*) from multipletype;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/62f2dfb3/src/test/regress/sql/parquet_pagerowgroup_size.sql
----------------------------------------------------------------------
diff --git a/src/test/regress/sql/parquet_pagerowgroup_size.sql b/src/test/regress/sql/parquet_pagerowgroup_size.sql
deleted file mode 100644
index f99710a..0000000
--- a/src/test/regress/sql/parquet_pagerowgroup_size.sql
+++ /dev/null
@@ -1,45 +0,0 @@
--- @description parquet insert vary pagesize/rowgroupsize 
--- @created 2013-08-09 20:33:16
--- @modified 2013-08-09 20:33:16
--- @tags HAWQ parquet
-
---start_ignore
-drop table if exists pTable1;
-drop table if exists pTable3_from;
-drop table if exists pTable3_to;
-drop table if exists pTable4_from;
-drop table if exists pTable4_to;
-drop table if exists pTable5_from;
-drop table if exists pTable5_to;
---end_ignore
-
---value/record size equal to pagesize/rowgroupsize
-create table pTable1 (a1 char(10485760), a2 char(10485760), a3 char(10485760), a4 char(10485760), a5 char(10485760), a6 char(10485760), a7 char(10485760), a8 char(10485760), a9 char(10485760), a10 char(10485760)) with(appendonly=true, orientation=parquet, pagesize=10485760, rowgroupsize=104857600);
-
-insert into pTable1 values ( ('a'::char(10485760)), ('a'::char(10485760)), ('a'::char(10485760)), ('a'::char(10485760)), ('a'::char(10485760)), ('a'::char(10485760)), ('a'::char(10485760)), ('a'::char(10485760)), ('a'::char(10485760)), ('a'::char(10485760)) );
-
---single column, one data page contains several values, one rwo group contains several groups
-create table pTable3_from ( a1 text ) with(appendonly=true, orientation=parquet);
-insert into pTable3_from values(repeat('parquet',100));
-insert into pTable3_from values(repeat('parquet',20));
-insert into pTable3_from values(repeat('parquet',30));
-create table pTable3_to ( a1 text ) with(appendonly=true, orientation=parquet, pagesize=1024, rowgroupsize=1025);
-insert into pTable3_to select * from pTable3_from;
-select count(*) from pTable3_to;
-
---multiple columns, multiple rows combination
-create table pTable4_from ( a1 text , a2 text) with(appendonly=true, orientation=parquet);
-insert into pTable4_from values(repeat('parquet',200), repeat('pq',200));
-insert into pTable4_from values(repeat('parquet',50), repeat('pq',200));
-
-create table pTable4_to ( a1 text, a2 text ) with(appendonly=true, orientation=parquet, pagesize=2048, rowgroupsize=4096);
-insert into pTable4_to select * from pTable4_from;
-select count(*) from pTable4_to;
-
---large data insert, several column values in one page, several rows in one rowgroup
-create table pTable5_from (a1 char(1048576), a2 char(2048576), a3 char(3048576), a4 char(4048576), a5 char(5048576), a6 char(6048576), a7 char(7048576), a8 char(8048576), a9 char(9048576), a10 char(9)) with(appendonly=true, orientation=parquet, pagesize=10485760, rowgroupsize=90874386);
-insert into pTable5_from values ( ('a'::char(1048576)), ('a'::char(2048576)), ('a'::char(3048576)), ('a'::char(4048576)), ('a'::char(5048576)), ('a'::char(6048576)), ('a'::char(7048576)), ('a'::char(8048576)), ('a'::char(9048576)), ('a'::char(9)) );
-
-create table pTable5_to (a1 char(1048576), a2 char(2048576), a3 char(3048576), a4 char(4048576), a5 char(5048576), a6 char(6048576), a7 char(7048576), a8 char(8048576), a9 char(9048576), a10 char(9)) with(appendonly=true, orientation=parquet, pagesize=10485760, rowgroupsize=17437200);
-insert into pTable5_to select * from pTable5_from;
-select count(a10) from pTable5_to;
\ No newline at end of file


Mime
View raw message