carbondata-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From jack...@apache.org
Subject [05/38] incubator-carbondata git commit: reuse test case for integration module
Date Sat, 07 Jan 2017 16:36:39 GMT
http://git-wip-us.apache.org/repos/asf/incubator-carbondata/blob/af2f204e/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/AllDataTypesTestCaseFilter.scala
----------------------------------------------------------------------
diff --git a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/AllDataTypesTestCaseFilter.scala b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/AllDataTypesTestCaseFilter.scala
deleted file mode 100644
index 344b0c4..0000000
--- a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/AllDataTypesTestCaseFilter.scala
+++ /dev/null
@@ -1,66 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *    http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied.  See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-
-package org.apache.carbondata.spark.testsuite.filterexpr
-
-import org.apache.spark.sql.Row
-import org.apache.spark.sql.common.util.CarbonHiveContext._
-import org.apache.spark.sql.common.util.QueryTest
-import org.scalatest.BeforeAndAfterAll
-
-/**
- * Test Class for filter expression query on multiple datatypes
- * @author N00902756
- *
- */
-
-class AllDataTypesTestCaseFilter extends QueryTest with BeforeAndAfterAll {
-
-  override def beforeAll {
-    sql("CREATE TABLE alldatatypestableFilter (empno int, empname String, designation String, doj Timestamp, workgroupcategory int, workgroupcategoryname String, deptno int, deptname String, projectcode int, projectjoindate Timestamp, projectenddate Timestamp,attendance int,utilization int,salary int) STORED BY 'org.apache.carbondata.format'")
-    sql("LOAD DATA local inpath './src/test/resources/data.csv' INTO TABLE alldatatypestableFilter OPTIONS('DELIMITER'= ',', 'QUOTECHAR'= '\"')");
-    
-    sql("CREATE TABLE alldatatypestableFilter_hive (empno int, empname String, designation String, doj Timestamp, workgroupcategory int, workgroupcategoryname String, deptno int, deptname String, projectcode int, projectjoindate Timestamp, projectenddate Timestamp,attendance int,utilization int,salary int)row format delimited fields terminated by ','")
-    sql("LOAD DATA local inpath './src/test/resources/datawithoutheader.csv' INTO TABLE alldatatypestableFilter_hive");
-
-  }
-
-  test("select empno,empname,utilization,count(salary),sum(empno) from alldatatypestableFilter where empname in ('arvind','ayushi') group by empno,empname,utilization") {
-    checkAnswer(
-      sql("select empno,empname,utilization,count(salary),sum(empno) from alldatatypestableFilter where empname in ('arvind','ayushi') group by empno,empname,utilization"),
-      sql("select empno,empname,utilization,count(salary),sum(empno) from alldatatypestableFilter_hive where empname in ('arvind','ayushi') group by empno,empname,utilization"))
-  }
-  
-  test("select empno,empname from alldatatypestableFilter where regexp_replace(workgroupcategoryname, 'er', 'ment') NOT IN ('development')") {
-    checkAnswer(
-      sql("select empno,empname from alldatatypestableFilter where regexp_replace(workgroupcategoryname, 'er', 'ment') NOT IN ('development')"),
-      sql("select empno,empname from alldatatypestableFilter_hive where regexp_replace(workgroupcategoryname, 'er', 'ment') NOT IN ('development')"))
-  }
-  
-  test("select empno,empname from alldatatypescubeFilter where regexp_replace(workgroupcategoryname, 'er', 'ment') != 'development'") {
-    checkAnswer(
-      sql("select empno,empname from alldatatypestableFilter where regexp_replace(workgroupcategoryname, 'er', 'ment') != 'development'"),
-      sql("select empno,empname from alldatatypestableFilter_hive where regexp_replace(workgroupcategoryname, 'er', 'ment') != 'development'"))
-  }
-  
-  override def afterAll {
-    sql("drop table alldatatypestableFilter")
-    sql("drop table alldatatypestableFilter_hive")
-  }
-}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-carbondata/blob/af2f204e/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/CountStarTestCase.scala
----------------------------------------------------------------------
diff --git a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/CountStarTestCase.scala b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/CountStarTestCase.scala
deleted file mode 100644
index cf3f5f6..0000000
--- a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/CountStarTestCase.scala
+++ /dev/null
@@ -1,73 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *    http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied.  See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-
-package org.apache.carbondata.spark.testsuite.filterexpr
-
-import java.io.File
-
-import org.apache.spark.sql.Row
-import org.apache.spark.sql.common.util.CarbonHiveContext._
-import org.apache.spark.sql.common.util.QueryTest
-import org.apache.carbondata.core.constants.CarbonCommonConstants
-import org.apache.carbondata.core.util.CarbonProperties
-import org.scalatest.BeforeAndAfterAll
-
-/**
-  * Test Class for filter expression query on String datatypes
-  *
-  * @author N00902756
-  *
-  */
-class CountStarTestCase extends QueryTest with BeforeAndAfterAll {
-
-  override def beforeAll {
-    sql("drop table if exists filtertestTables")
-    sql("drop table if exists filterTimestampDataType")
-
-    sql("CREATE TABLE filterTimestampDataType (ID int, date Timestamp, country String, " +
-      "name String, phonetype String, serialname String, salary int) " +
-      "STORED BY 'org.apache.carbondata.format'"
-    )
-    CarbonProperties.getInstance()
-      .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "yyyy/MM/dd")
-    val currentDirectory = new File(this.getClass.getResource("/").getPath + "/../../")
-      .getCanonicalPath
-    var csvFilePath = currentDirectory + "/src/test/resources/datanullmeasurecol.csv";
-      sql(
-      s"LOAD DATA LOCAL INPATH '" + csvFilePath + "' INTO TABLE " +
-        s"filterTimestampDataType " +
-        s"OPTIONS('DELIMITER'= ',', " +
-        s"'FILEHEADER'= '')"
-    )
-  }
-
-  test("select count ") {
-    checkAnswer(
-      sql("select count(*) from filterTimestampDataType where country='china'"),
-      Seq(Row(2))
-    )
-  }
-
-  override def afterAll {
-    sql("drop table if exists filtertestTables")
-    sql("drop table if exists filterTimestampDataType")
-    CarbonProperties.getInstance()
-      .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "dd-MM-yyyy")
-  }
-}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-carbondata/blob/af2f204e/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/FilterProcessorTestCase.scala
----------------------------------------------------------------------
diff --git a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/FilterProcessorTestCase.scala b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/FilterProcessorTestCase.scala
deleted file mode 100644
index 7d45e88..0000000
--- a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/FilterProcessorTestCase.scala
+++ /dev/null
@@ -1,289 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *    http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied.  See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-
-package org.apache.carbondata.spark.testsuite.filterexpr
-
-import java.sql.Timestamp
-
-import org.apache.spark.sql.Row
-import org.apache.spark.sql.common.util.CarbonHiveContext._
-import org.apache.spark.sql.common.util.QueryTest
-import org.apache.carbondata.core.constants.CarbonCommonConstants
-import org.apache.carbondata.core.util.CarbonProperties
-import org.scalatest.BeforeAndAfterAll
-
-/**
-  * Test Class for filter expression query on String datatypes
-  *
-  */
-class FilterProcessorTestCase extends QueryTest with BeforeAndAfterAll {
-
-  override def beforeAll {
-    sql("drop table if exists filtertestTables")
-    sql("drop table if exists filtertestTablesWithDecimal")
-    sql("drop table if exists filtertestTablesWithNull")
-    sql("drop table if exists filterTimestampDataType")
-    sql("drop table if exists noloadtable")
-
-    sql("CREATE TABLE filtertestTables (ID int, date Timestamp, country String, " +
-      "name String, phonetype String, serialname String, salary int) " +
-        "STORED BY 'org.apache.carbondata.format'"
-    )
-    sql("CREATE TABLE noloadtable (ID int, date Timestamp, country String, " +
-      "name String, phonetype String, serialname String, salary int) " +
-      "STORED BY 'org.apache.carbondata.format'"
-    )
-     CarbonProperties.getInstance()
-        .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "MM-dd-yyyy HH:mm:ss")
-
-     sql("CREATE TABLE filterTimestampDataType (ID int, date Timestamp, country String, " +
-      "name String, phonetype String, serialname String, salary int) " +
-        "STORED BY 'org.apache.carbondata.format'"
-    )
-       CarbonProperties.getInstance()
-        .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "MM-dd-yyyy HH:mm:ss")
-    sql(
-      s"LOAD DATA LOCAL INPATH './src/test/resources/data2_DiffTimeFormat.csv' INTO TABLE " +
-        s"filterTimestampDataType " +
-        s"OPTIONS('DELIMITER'= ',', " +
-        s"'FILEHEADER'= '')"
-    )
-    CarbonProperties.getInstance()
-      .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "yyyy/MM/dd")
-    sql(
-      s"LOAD DATA local inpath './src/test/resources/dataDiff.csv' INTO TABLE filtertestTables " +
-        s"OPTIONS('DELIMITER'= ',', " +
-        s"'FILEHEADER'= '')"
-    )
-    sql(
-      "CREATE TABLE filtertestTablesWithDecimal (ID decimal, date Timestamp, country " +
-        "String, " +
-        "name String, phonetype String, serialname String, salary int) " +
-      "STORED BY 'org.apache.carbondata.format'"
-    )
-    sql(
-      s"LOAD DATA LOCAL INPATH './src/test/resources/dataDiff.csv' INTO TABLE " +
-        s"filtertestTablesWithDecimal " +
-        s"OPTIONS('DELIMITER'= ',', " +
-        s"'FILEHEADER'= '')"
-    )
-    sql("DROP TABLE IF EXISTS filtertestTablesWithNull")
-    sql(
-      "CREATE TABLE filtertestTablesWithNull (ID int, date Timestamp, country " +
-        "String, " +
-        "name String, phonetype String, serialname String,salary int) " +
-      "STORED BY 'org.apache.carbondata.format'"
-    )
-    sql("DROP TABLE IF EXISTS filtertestTablesWithNullJoin")
-    sql(
-      "CREATE TABLE filtertestTablesWithNullJoin (ID int, date Timestamp, country " +
-        "String, " +
-        "name String, phonetype String, serialname String,salary int) " +
-      "STORED BY 'org.apache.carbondata.format'"
-    )
-    CarbonProperties.getInstance()
-      .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "yyyy-MM-dd HH:mm:ss")
-    sql(
-      s"LOAD DATA LOCAL INPATH './src/test/resources/data2.csv' INTO TABLE " +
-        s"filtertestTablesWithNull " +
-        s"OPTIONS('DELIMITER'= ',', " +
-        s"'FILEHEADER'= '')"
-    )
-        sql(
-      s"LOAD DATA LOCAL INPATH './src/test/resources/data2.csv' INTO TABLE " +
-        s"filtertestTablesWithNullJoin " +
-        s"OPTIONS('DELIMITER'= ',', " +
-        s"'FILEHEADER'= '')"
-    )
-
-    sql("DROP TABLE IF EXISTS big_int_basicc")
-    sql("DROP TABLE IF EXISTS big_int_basicc_1")
-    sql("DROP TABLE IF EXISTS big_int_basicc_Hive")
-    sql("DROP TABLE IF EXISTS big_int_basicc_Hive_1")
-    sql("CREATE TABLE big_int_basicc (imei string,age int,task bigint,name string,country string,city string,sale int,num double,level decimal(10,3),quest bigint,productdate timestamp,enddate timestamp,PointId double,score decimal(10,3))STORED BY 'org.apache.carbondata.format'")
-    sql("CREATE TABLE big_int_basicc_1 (imei string,age int,task bigint,name string,country string,city string,sale int,num double,level decimal(10,3),quest bigint,productdate timestamp,enddate timestamp,PointId double,score decimal(10,3))STORED BY 'org.apache.carbondata.format'")
-    sql("CREATE TABLE big_int_basicc_Hive (imei string,age int,task bigint,name string,country string,city string,sale int,num double,level decimal(10,3),quest bigint,productdate date,enddate date,PointId double,score decimal(10,3))row format delimited fields terminated by ',' " +
-        "tblproperties(\"skip.header.line.count\"=\"1\") ")
-    sql("CREATE TABLE big_int_basicc_Hive_1 (imei string,age int,task bigint,name string,country string,city string,sale int,num double,level decimal(10,3),quest bigint,productdate date,enddate date,PointId double,score decimal(10,3))row format delimited fields terminated by ',' " +
-        "tblproperties(\"skip.header.line.count\"=\"1\") ")
-    CarbonProperties.getInstance().addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "yyyy-MM-dd HH:mm:ss")
-    sql("LOAD DATA INPATH './src/test/resources/big_int_Decimal.csv'  INTO TABLE big_int_basicc options ('DELIMITER'=',', 'QUOTECHAR'='\"', 'COMPLEX_DELIMITER_LEVEL_1'='$','COMPLEX_DELIMITER_LEVEL_2'=':', 'FILEHEADER'= '')")
-    sql("LOAD DATA INPATH './src/test/resources/big_int_Decimal.csv'  INTO TABLE big_int_basicc_1 options ('DELIMITER'=',', 'QUOTECHAR'='\"', 'COMPLEX_DELIMITER_LEVEL_1'='$','COMPLEX_DELIMITER_LEVEL_2'=':', 'FILEHEADER'= '')")
-    sql("load data local inpath './src/test/resources/big_int_Decimal.csv' into table big_int_basicc_Hive")
-    sql("load data local inpath './src/test/resources/big_int_Decimal.csv' into table big_int_basicc_Hive_1")
-  }
-
-  test("Is not null filter") {
-    checkAnswer(
-      sql("select id from filtertestTablesWithNull " + "where id is not null"),
-      Seq(Row(4), Row(6))
-    )
-  }
-  
-    test("join filter") {
-    checkAnswer(
-      sql("select b.name from filtertestTablesWithNull a join filtertestTablesWithNullJoin b  " + "on a.name=b.name"),
-      Seq(Row("aaa4"), Row("aaa5"),Row("aaa6"))
-    )
-  }
-  
-    test("Between  filter") {
-    checkAnswer(
-      sql("select date from filtertestTablesWithNull " + " where date between '2014-01-20 00:00:00' and '2014-01-28 00:00:00'"),
-      Seq(Row(Timestamp.valueOf("2014-01-21 00:00:00")), Row(Timestamp.valueOf("2014-01-22 00:00:00")))
-    )
-  }
-    test("Multi column with invalid member filter") {
-    checkAnswer(
-      sql("select id from filtertestTablesWithNull " + "where id = salary"),
-      Seq()
-    )
-  }
-
-  test("Greater Than Filter") {
-    checkAnswer(
-      sql("select id from filtertestTables " + "where id >999"),
-      Seq(Row(1000))
-    )
-  }
-  test("Greater Than Filter with decimal") {
-    checkAnswer(
-      sql("select id from filtertestTablesWithDecimal " + "where id >999"),
-      Seq(Row(1000))
-    )
-  }
-
-  test("Greater Than equal to Filter") {
-    checkAnswer(
-      sql("select id from filtertestTables " + "where id >=999"),
-      Seq(Row(999), Row(1000))
-    )
-  }
-  
-    test("Greater Than equal to Filter with limit") {
-    checkAnswer(
-      sql("select id from filtertestTables " + "where id >=999 order by id desc limit 1"),
-      Seq(Row(1000))
-    )
-  }
-
-      test("Greater Than equal to Filter with aggregation limit") {
-    checkAnswer(
-      sql("select count(id),country from filtertestTables " + "where id >=999 group by country limit 1"),
-      Seq(Row(2,"china"))
-    )
-  }
-  test("Greater Than equal to Filter with decimal") {
-    checkAnswer(
-      sql("select id from filtertestTables " + "where id >=999"),
-      Seq(Row(999), Row(1000))
-    )
-  }
-  test("Include Filter") {
-    checkAnswer(
-      sql("select id from filtertestTables " + "where id =999"),
-      Seq(Row(999))
-    )
-  }
-  test("In Filter") {
-    checkAnswer(
-      sql(
-        "select Country from filtertestTables where Country in ('china','france') group by Country"
-      ),
-      Seq(Row("china"), Row("france"))
-    )
-  }
-
-  test("Logical condition") {
-    checkAnswer(
-      sql("select id,country from filtertestTables " + "where country='china' and name='aaa1'"),
-      Seq(Row(1, "china"))
-    )
-  }
-  
-  test("filter query over table having no data") {
-    checkAnswer(
-      sql("select * from noloadtable " + "where country='china' and name='aaa1'"),
-      Seq()
-    )
-  }
-
-
-    
-     test("Time stamp filter with diff time format for load greater") {
-    checkAnswer(
-      sql("select date  from filterTimestampDataType where date > '2014-07-10 00:00:00'"),
-      Seq(Row(Timestamp.valueOf("2014-07-20 00:00:00.0")),
-        Row(Timestamp.valueOf("2014-07-25 00:00:00.0"))
-      )
-    )
-  }
-    test("Time stamp filter with diff time format for load less") {
-    checkAnswer(
-      sql("select date  from filterTimestampDataType where date < '2014-07-20 00:00:00'"),
-      Seq(Row(Timestamp.valueOf("2014-07-10 00:00:00.0"))
-      )
-    )
-  }
-   test("Time stamp filter with diff time format for load less than equal") {
-    checkAnswer(
-      sql("select date  from filterTimestampDataType where date <= '2014-07-20 00:00:00'"),
-      Seq(Row(Timestamp.valueOf("2014-07-10 00:00:00.0")),Row(Timestamp.valueOf("2014-07-20 00:00:00.0"))
-      )
-    )
-  }
-      test("Time stamp filter with diff time format for load greater than equal") {
-    checkAnswer(
-      sql("select date  from filterTimestampDataType where date >= '2014-07-20 00:00:00'"),
-      Seq(Row(Timestamp.valueOf("2014-07-20 00:00:00.0")),Row(Timestamp.valueOf("2014-07-25 00:00:00.0"))
-      )
-    )
-  }
-    test("join query with bigdecimal filter") {
-
-    checkAnswer(
-      sql("select b.level from big_int_basicc_Hive a join big_int_basicc_Hive_1 b on a.level=b.level order by level"),
-      sql("select b.level from big_int_basicc a join big_int_basicc_1 b on a.level=b.level order by level")
-    )
-  }
-    
-        test("join query with bigint filter") {
-
-    checkAnswer(
-      sql("select b.task from big_int_basicc_Hive a join big_int_basicc_Hive_1 b on a.task=b.task"),
-      sql("select b.task from big_int_basicc a join big_int_basicc_1 b on a.task=b.task")
-    )
-  }
-
-  override def afterAll {
-    sql("drop table if exists filtertestTables")
-    sql("drop table if exists filtertestTablesWithDecimal")
-    sql("drop table if exists filtertestTablesWithNull")
-    sql("drop table if exists filterTimestampDataType")
-    sql("drop table if exists noloadtable")
-    sql("DROP TABLE IF EXISTS big_int_basicc")
-    sql("DROP TABLE IF EXISTS big_int_basicc_1")
-    sql("DROP TABLE IF EXISTS big_int_basicc_Hive")
-    sql("DROP TABLE IF EXISTS big_int_basicc_Hive_1")
-    sql("DROP TABLE IF EXISTS filtertestTablesWithNull")
-    sql("DROP TABLE IF EXISTS filtertestTablesWithNullJoin")
-    CarbonProperties.getInstance()
-      .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "dd-MM-yyyy")
-  }
-}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-carbondata/blob/af2f204e/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/GrtLtFilterProcessorTestCase.scala
----------------------------------------------------------------------
diff --git a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/GrtLtFilterProcessorTestCase.scala b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/GrtLtFilterProcessorTestCase.scala
index b33b65f..eddde30 100644
--- a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/GrtLtFilterProcessorTestCase.scala
+++ b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/GrtLtFilterProcessorTestCase.scala
@@ -19,14 +19,12 @@
 
 package org.apache.carbondata.spark.testsuite.filterexpr
 
-import java.io.File
-
 import org.apache.spark.sql.Row
-import org.apache.spark.sql.common.util.CarbonHiveContext._
 import org.apache.spark.sql.common.util.QueryTest
+import org.scalatest.BeforeAndAfterAll
+
 import org.apache.carbondata.core.constants.CarbonCommonConstants
 import org.apache.carbondata.core.util.CarbonProperties
-import org.scalatest.BeforeAndAfterAll
 
 /**
   * Test Class for filter expression query on String datatypes
@@ -35,158 +33,33 @@ import org.scalatest.BeforeAndAfterAll
 class GrtLtFilterProcessorTestCase extends QueryTest with BeforeAndAfterAll {
 
   override def beforeAll {
-    sql("drop table if exists a12")
     sql("drop table if exists a12_allnull")
-    sql("drop table if exists a12_no_null")
-     sql("drop table if exists Test_Boundary1")
 
     sql(
-      "create table a12(empid String,ename String,sal double,deptno int,mgr string,gender string," +
-        "dob timestamp,comm decimal(4,2),desc string) stored by 'org.apache.carbondata.format'"
-    )
-    sql(
       "create table a12_allnull(empid String,ename String,sal double,deptno int,mgr string,gender" +
         " string," +
         "dob timestamp,comm decimal(4,2),desc string) stored by 'org.apache.carbondata.format'"
     )
-    sql(
-      "create table a12_no_null(empid String,ename String,sal double,deptno int,mgr string,gender" +
-        " string," +
-        "dob timestamp,comm decimal(4,2),desc string) stored by 'org.apache.carbondata.format'"
-    )
-    sql("create table Test_Boundary1 (c1_int int,c2_Bigint Bigint,c3_Decimal Decimal(38,38),c4_double double,c5_string string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 'org.apache.carbondata.format'")
     CarbonProperties.getInstance()
       .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "yyyy-MM-dd HH:mm:ss")
-    val basePath = new File(this.getClass.getResource("/").getPath + "/../../")
-      .getCanonicalPath
-    var testData = basePath + "/src/test/resources/filter/emp2.csv"
-    sql(
-      s"""LOAD DATA LOCAL INPATH '$testData' into table a12 OPTIONS('DELIMITER'=',',
-         'QUOTECHAR'='"','FILEHEADER'='empid,ename,sal,deptno,mgr,gender,dob,comm,desc')"""
-        .stripMargin
-    )
-    testData = basePath + "/src/test/resources/filter/emp2allnull.csv"
+    val testData = s"$resourcesPath/filter/emp2allnull.csv"
 
     sql(
       s"""LOAD DATA LOCAL INPATH '$testData' into table a12_allnull OPTIONS('DELIMITER'=',',
          'QUOTECHAR'='"','FILEHEADER'='empid,ename,sal,deptno,mgr,gender,dob,comm,desc')"""
         .stripMargin
     )
-    testData = basePath + "/src/test/resources/filter/emp2nonull.csv"
-
-    sql(
-      s"""LOAD DATA LOCAL INPATH '$testData' into table a12_no_null OPTIONS('DELIMITER'=',',
-         'QUOTECHAR'='"')"""
-        .stripMargin
-    )
-    
-    sql(
-      s"LOAD DATA INPATH './src/test/resources/Test_Data1_Logrithmic.csv' INTO table Test_Boundary1 OPTIONS('DELIMITER'=',','QUOTECHAR'='','FILEHEADER'='')")
-  }
-  //mixed value test
-  test("Less Than Filter") {
-    checkAnswer(
-      sql("select count(empid) from a12 where dob < '2014-07-01 12:07:28'"),
-      Seq(Row(0))
-    )
-  }
-
-  test("Les Than equal Filter") {
-    checkAnswer(
-      sql("select count (empid) from a12 where dob <= '2014-07-01 12:07:28'"),
-      Seq(Row(2))
-    )
-  }
-
-  test("Greater Than Filter") {
-    checkAnswer(
-      sql("select count (empid) from a12 where dob > '2014-07-01 12:07:28'"),
-      Seq(Row(3))
-    )
-  }
-  test("0.0 and -0.0 equality check for double data type applying log function") {
-    checkAnswer(
-      sql("select log(c4_double,1) from Test_Boundary1 where log(c4_double,1)= -0.0"),
-      Seq(Row(0.0),Row(0.0))
-    )
-  }
-
-  test("Greater Than equal to Filter") {
-    sql("select count (empid) from a12 where dob >= '2014-07-01 12:07:28'").show()
-    checkAnswer(
-      sql("select count (empid) from a12 where dob >= '2014-07-01 12:07:28'"),
-      Seq(Row(5))
-    )
-  }
-  //all null test cases
-
-  test("Less Than Filter all null") {
-    checkAnswer(
-      sql("select count(empid) from a12_allnull where dob < '2014-07-01 12:07:28'"),
-      Seq(Row(0))
-    )
-  }
-
-  test("Les Than equal Filter all null") {
-    checkAnswer(
-      sql("select count (empid) from a12_allnull where dob <= '2014-07-01 12:07:28'"),
-      Seq(Row(0))
-    )
   }
 
-  test("Greater Than Filter all null") {
-    checkAnswer(
-      sql("select count (empid) from a12_allnull where dob > '2014-07-01 12:07:28'"),
-      Seq(Row(0))
-    )
-  }
-
-  test("Greater Than equal to Filter all null") {
-    checkAnswer(
-      sql("select count (empid) from a12_allnull where dob >= '2014-07-01 12:07:28'"),
-      Seq(Row(0))
-    )
-  }
-      test("In condition With improper format query regarding Null filter") {
+  test("In condition With improper format query regarding Null filter") {
     checkAnswer(
       sql("select empid from a12_allnull " + "where empid not in ('china',NULL)"),
       Seq()
     )
-      }
-
-  //no null test cases
-
-  test("Less Than Filter no null") {
-    checkAnswer(
-      sql("select count(empid) from a12_no_null where dob < '2014-07-01 12:07:28'"),
-      Seq(Row(0))
-    )
-  }
-
-  test("Les Than equal Filter no null") {
-    sql("select empid from a12_no_null where dob <= '2014-07-01 12:07:28'").show()
-    checkAnswer(
-      sql("select count(empid) from a12_no_null where dob <= '2014-07-01 12:07:28'"),
-      Seq(Row(4))
-    )
-  }
-
-  test("Greater Than Filter no null") {
-    checkAnswer(
-      sql("select count (empid) from a12_no_null where dob > '2014-07-01 12:07:28'"),
-      Seq(Row(3))
-    )
-  }
-
-  test("Greater Than equal to Filter no null") {
-    checkAnswer(
-      sql("select count (empid) from a12_no_null where dob >= '2014-07-01 12:07:28'"),
-      Seq(Row(7))
-    )
   }
 
   override def afterAll {
-    sql("drop table a12")
+    sql("drop table if exists a12_allnull")
     CarbonProperties.getInstance()
       .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "dd-MM-yyyy")
   }

http://git-wip-us.apache.org/repos/asf/incubator-carbondata/blob/af2f204e/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/IntegerDataTypeTestCase.scala
----------------------------------------------------------------------
diff --git a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/IntegerDataTypeTestCase.scala b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/IntegerDataTypeTestCase.scala
deleted file mode 100644
index db82e9c..0000000
--- a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/IntegerDataTypeTestCase.scala
+++ /dev/null
@@ -1,48 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *    http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied.  See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-
-package org.apache.carbondata.spark.testsuite.filterexpr
-
-import org.apache.spark.sql.Row
-import org.apache.spark.sql.common.util.CarbonHiveContext._
-import org.apache.spark.sql.common.util.QueryTest
-import org.scalatest.BeforeAndAfterAll
-
-/**
- * Test Class for filter expression query on Integer datatypes
- * @author N00902756
- *
- */
-class IntegerDataTypeTestCase extends QueryTest with BeforeAndAfterAll {
-
-  override def beforeAll {
-    sql("CREATE TABLE integertypetableFilter (empno int, workgroupcategory string, deptno int, projectcode int,attendance int) STORED BY 'org.apache.carbondata.format'")
-    sql("LOAD DATA local inpath './src/test/resources/data.csv' INTO TABLE integertypetableFilter OPTIONS('DELIMITER'= ',', 'QUOTECHAR'= '\"')")
-  }
-
-  test("select empno from integertypetableFilter") {
-    checkAnswer(
-      sql("select empno from integertypetableFilter"),
-      Seq(Row(11), Row(12), Row(13), Row(14), Row(15), Row(16), Row(17), Row(18), Row(19), Row(20)))
-  }
-
-  override def afterAll {
-    sql("drop table integertypetableFilter")
-  }
-}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-carbondata/blob/af2f204e/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/NullMeasureValueTestCaseFilter.scala
----------------------------------------------------------------------
diff --git a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/NullMeasureValueTestCaseFilter.scala b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/NullMeasureValueTestCaseFilter.scala
deleted file mode 100644
index 381308f..0000000
--- a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/NullMeasureValueTestCaseFilter.scala
+++ /dev/null
@@ -1,58 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *    http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied.  See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-package org.apache.carbondata.spark.testsuite.filterexpr
-
-import org.apache.spark.sql.Row
-import org.apache.spark.sql.common.util.CarbonHiveContext._
-import org.apache.spark.sql.common.util.QueryTest
-import org.apache.carbondata.core.constants.CarbonCommonConstants
-import org.apache.carbondata.core.util.CarbonProperties
-import org.scalatest.BeforeAndAfterAll
-
-class NullMeasureValueTestCaseFilter extends QueryTest with BeforeAndAfterAll {
-
-  override def beforeAll {
-    sql("drop table if exists t3")
-    sql(
-      "CREATE TABLE t3 (ID bigInt, date Timestamp, country String, name String, " +
-        "phonetype String, serialname String, salary Int) STORED BY 'org.apache.carbondata.format'"
-    )
-    CarbonProperties.getInstance()
-      .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "yyyy/MM/dd")
-    sql("LOAD DATA LOCAL INPATH './src/test/resources/datawithnullmeasure.csv' into table t3");
-  }
-
-  test("select ID from t3 where salary is not null") {
-    checkAnswer(
-      sql("select ID from t3 where salary is not null"),
-      Seq(Row(1),Row(4)))
-  }
-
-  test("select ID from t3 where salary is null") {
-    checkAnswer(
-      sql("select ID from t3 where salary is null"),
-      Seq(Row(2),Row(3)))
-  }
-
-  override def afterAll {
-    sql("drop table t3")
-    CarbonProperties.getInstance()
-      .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "dd-MM-yyyy")
-  }
-}

http://git-wip-us.apache.org/repos/asf/incubator-carbondata/blob/af2f204e/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/TestAndEqualFilterEmptyOperandValue.scala
----------------------------------------------------------------------
diff --git a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/TestAndEqualFilterEmptyOperandValue.scala b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/TestAndEqualFilterEmptyOperandValue.scala
deleted file mode 100644
index 2da2b7d..0000000
--- a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/TestAndEqualFilterEmptyOperandValue.scala
+++ /dev/null
@@ -1,89 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *    http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied.  See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-
-package org.apache.carbondata.spark.testsuite.filterexpr
-
-import java.io.File
-
-import org.apache.spark.sql.common.util.CarbonHiveContext._
-import org.apache.spark.sql.common.util.QueryTest
-import org.scalatest.BeforeAndAfterAll
-
-import org.apache.carbondata.core.constants.CarbonCommonConstants
-import org.apache.carbondata.core.util.CarbonProperties
-
-/**
- * Test cases for testing columns having \N or \null values for non numeric columns
- */
-class TestAndEqualFilterEmptyOperandValue extends QueryTest with BeforeAndAfterAll {
-
-  override def beforeAll {
-    sql("drop table if exists carbonTable")
-    sql("drop table if exists hiveTable")
-    CarbonProperties.getInstance()
-      .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT,
-        CarbonCommonConstants.CARBON_TIMESTAMP_DEFAULT_FORMAT
-      )
-    val currentDirectory = new File(this.getClass.getResource("/").getPath + "/../../")
-      .getCanonicalPath
-    val csvFilePath = currentDirectory + "/src/test/resources/filter/datagrtlrt.csv"
-    sql(
-      "CREATE TABLE IF NOT EXISTS carbonTable(date Timestamp, country String, salary Int) STORED " +
-      "BY " +
-      "'carbondata'"
-    )
-    sql(
-      "create table if not exists hiveTable(date Timestamp, country String, salary Int)row format" +
-      " delimited fields " +
-      "terminated by ','"
-    )
-    sql(
-      "LOAD DATA LOCAL INPATH '" + csvFilePath + "' into table carbonTable OPTIONS " +
-      "('FILEHEADER'='date,country,salary')"
-    )
-    sql(
-      "LOAD DATA local inpath '" + csvFilePath + "' INTO table hiveTable"
-    )
-  }
-
-  test("select * from carbonTable where country='' and salary='')") {
-    sql("select * from carbonTable where country='' and salary=''").show()
-    sql("select * from hiveTable where country='' and salary=''").show()
-    checkAnswer(
-      sql("select * from carbonTable where country='' and salary=''"),
-      sql("select * from hiveTable where country='' and salary=''")
-    )
-  }
-
-  test("select * from carbonTable where date='' and salary='')") {
-    sql("select * from carbonTable where date='' and salary=''").show()
-    sql("select * from hiveTable where date='' and salary=''").show()
-    checkAnswer(
-      sql("select * from carbonTable where date='' and salary=''"),
-      sql("select * from hiveTable where date='' and salary=''")
-    )
-  }
-
-  override def afterAll {
-    sql("drop table if exists carbonTable")
-    sql("drop table if exists hiveTable")
-    CarbonProperties.getInstance()
-      .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "dd-MM-yyyy")
-  }
-}

http://git-wip-us.apache.org/repos/asf/incubator-carbondata/blob/af2f204e/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/TestGrtLessFilter.scala
----------------------------------------------------------------------
diff --git a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/TestGrtLessFilter.scala b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/TestGrtLessFilter.scala
deleted file mode 100644
index ea28252..0000000
--- a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/filterexpr/TestGrtLessFilter.scala
+++ /dev/null
@@ -1,96 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *    http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied.  See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-
-package org.apache.carbondata.spark.testsuite.filterexpr
-
-import java.io.File
-
-import org.apache.carbondata.core.constants.CarbonCommonConstants
-import org.apache.carbondata.core.util.CarbonProperties
-import org.apache.spark.sql.common.util.CarbonHiveContext._
-import org.apache.spark.sql.common.util.QueryTest
-import org.scalatest.BeforeAndAfterAll
-
-/**
- * Test cases for testing columns having \N or \null values for non numeric columns
- */
-class TestGrtLessFilter extends QueryTest with BeforeAndAfterAll {
-
-  override def beforeAll {
-    sql("drop table if exists carbonTable")
-    sql("drop table if exists hiveTable")
-    CarbonProperties.getInstance()
-      .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT,
-        CarbonCommonConstants.CARBON_TIMESTAMP_DEFAULT_FORMAT
-      )
-    val currentDirectory = new File(this.getClass.getResource("/").getPath + "/../../")
-      .getCanonicalPath
-    val csvFilePath = currentDirectory + "/src/test/resources/filter/datagrtlrt.csv"
-    sql(
-      "CREATE TABLE IF NOT EXISTS carbonTable(date Timestamp, country String, salary Int) STORED " +
-      "BY " +
-      "'carbondata'"
-    )
-    sql(
-      "create table if not exists hiveTable(date Timestamp, country String, salary Int)row format" +
-      " delimited fields " +
-      "terminated by ','"
-    )
-    sql(
-      "LOAD DATA LOCAL INPATH '" + csvFilePath + "' into table carbonTable OPTIONS " +
-      "('FILEHEADER'='date,country,salary')"
-    )
-    sql(
-      "LOAD DATA local inpath '" + csvFilePath + "' INTO table hiveTable"
-    )
-  }
-
-
-  test("select * from carbonTable where date > cast('2017-7-25 12:07:29' as timestamp)") {
-    checkAnswer(
-      sql("select * from carbonTable where date > cast('2017-7-25 12:07:29' as timestamp)"),
-      sql("select * from hiveTable where date > cast('2017-7-25 12:07:29' as timestamp)")
-    )
-  }
-  test("select * from carbonTable where date < cast('2017-7-25 12:07:29' as timestamp)") {
-    checkAnswer(
-      sql("select * from carbonTable where date < cast('2017-7-25 12:07:29' as timestamp)"),
-      sql("select * from hiveTable where date < cast('2017-7-25 12:07:29' as timestamp)")
-    )
-  }
-  test("select * from carbonTable where date > cast('2018-7-24 12:07:28' as timestamp)") {
-    checkAnswer(
-      sql("select * from carbonTable where date > cast('2018-7-24 12:07:28' as timestamp)"),
-      sql("select * from hiveTable where date > cast('2018-7-24 12:07:28' as timestamp)")
-    )
-  }
-  test("select * from carbonTable where date < cast('2018-7-24 12:07:28' as timestamp)") {
-    checkAnswer(
-      sql("select * from carbonTable where date < cast('2018-7-24 12:07:28' as timestamp)"),
-      sql("select * from hiveTable where date < cast('2018-7-24 12:07:28' as timestamp)")
-    )
-  }
-
-  override def afterAll {
-    sql("drop table if exists carbonTable")
-    sql("drop table if exists hiveTable")
-    CarbonProperties.getInstance()
-      .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "dd-MM-yyyy")
-  }
-}

http://git-wip-us.apache.org/repos/asf/incubator-carbondata/blob/af2f204e/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/hadooprelation/HadoopFSRelationTestCase.scala
----------------------------------------------------------------------
diff --git a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/hadooprelation/HadoopFSRelationTestCase.scala b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/hadooprelation/HadoopFSRelationTestCase.scala
index 0838a2b..95a94b5 100644
--- a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/hadooprelation/HadoopFSRelationTestCase.scala
+++ b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/hadooprelation/HadoopFSRelationTestCase.scala
@@ -19,8 +19,7 @@
 
 package org.apache.carbondata.spark.testsuite.hadooprelation
 
-import org.apache.spark.sql.{DataFrame, Row}
-import org.apache.spark.sql.common.util.CarbonHiveContext._
+import org.apache.spark.sql.DataFrame
 import org.apache.spark.sql.common.util.QueryTest
 import org.scalatest.BeforeAndAfterAll
 
@@ -31,6 +30,8 @@ import org.scalatest.BeforeAndAfterAll
 class HadoopFSRelationTestCase extends QueryTest with BeforeAndAfterAll {
 
   override def beforeAll {
+    sql("drop table if exists hadoopfsrelation")
+    sql("drop table if exists hadoopfsrelation_hive")
     sql(
       "CREATE TABLE hadoopfsrelation (empno int, empname String, designation " +
       "String, doj Timestamp, workgroupcategory int, workgroupcategoryname String, deptno " +
@@ -38,23 +39,23 @@ class HadoopFSRelationTestCase extends QueryTest with BeforeAndAfterAll {
       "Timestamp,attendance int,utilization int,salary int)" +
       "STORED BY 'org.apache.carbondata.format'")
     sql(
-      "LOAD DATA local inpath './src/test/resources/data.csv' INTO TABLE hadoopfsrelation " +
+      s"LOAD DATA local inpath '$resourcesPath/data.csv' INTO TABLE hadoopfsrelation " +
       "OPTIONS('DELIMITER'= ',', 'QUOTECHAR'= '\"')");
     
     sql("CREATE TABLE hadoopfsrelation_hive (empno int, empname String, designation String, doj Timestamp, workgroupcategory int, workgroupcategoryname String, deptno int, deptname String, projectcode int, projectjoindate Timestamp, projectenddate Timestamp,attendance int,utilization int,salary int)row format delimited fields terminated by ','")
     
     sql(
-      "LOAD DATA local inpath './src/test/resources/datawithoutheader.csv' INTO TABLE hadoopfsrelation_hive ");
+      s"LOAD DATA local inpath '$resourcesPath/datawithoutheader.csv' INTO TABLE hadoopfsrelation_hive ");
   }
 
   test("hadoopfsrelation select all test") {
-    val rdd = read.format("org.apache.spark.sql.CarbonSource")
+    val rdd = sqlContext.read.format("org.apache.spark.sql.CarbonSource")
       .option("tableName", "hadoopfsrelation").load()
     assert(rdd.collect().length > 0)
   }
 
   test("hadoopfsrelation filters test") {
-    val rdd: DataFrame = read.format("org.apache.spark.sql.CarbonSource")
+    val rdd: DataFrame = sqlContext.read.format("org.apache.spark.sql.CarbonSource")
       .option("tableName", "hadoopfsrelation").load()
       .select("empno", "empname", "utilization").where("empname in ('arvind','ayushi')")
     checkAnswer(

http://git-wip-us.apache.org/repos/asf/incubator-carbondata/blob/af2f204e/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/iud/DeleteCarbonTableTestCase.scala
----------------------------------------------------------------------
diff --git a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/iud/DeleteCarbonTableTestCase.scala b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/iud/DeleteCarbonTableTestCase.scala
index 5f76477..fd12262 100644
--- a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/iud/DeleteCarbonTableTestCase.scala
+++ b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/iud/DeleteCarbonTableTestCase.scala
@@ -19,7 +19,6 @@
 package org.apache.carbondata.spark.testsuite.iud
 
 import org.apache.spark.sql.Row
-import org.apache.spark.sql.common.util.CarbonHiveContext._
 import org.apache.spark.sql.common.util.QueryTest
 import org.scalatest.BeforeAndAfterAll
 
@@ -31,12 +30,12 @@ class DeleteCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
     sql("create database  iud_db")
 
     sql("""create table iud_db.source2 (c11 string,c22 int,c33 string,c55 string, c66 int) STORED BY 'org.apache.carbondata.format'""")
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/source2.csv' INTO table iud_db.source2""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/source2.csv' INTO table iud_db.source2""")
     sql("use iud_db")
   }
   test("delete data from carbon table with alias [where clause ]") {
     sql("""create table iud_db.dest (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud_db.dest""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud_db.dest""")
     sql("""delete from iud_db.dest d where d.c1 = 'a'""").show
     checkAnswer(
       sql("""select c2 from iud_db.dest"""),
@@ -46,7 +45,7 @@ class DeleteCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
   test("delete data from  carbon table[where clause ]") {
     sql("""drop table if exists iud_db.dest""")
     sql("""create table iud_db.dest (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud_db.dest""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud_db.dest""")
     sql("""delete from dest where c2 = 2""").show
     checkAnswer(
       sql("""select c1 from dest"""),
@@ -56,7 +55,7 @@ class DeleteCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
   test("delete data from  carbon table[where IN  ]") {
     sql("""drop table if exists iud_db.dest""")
     sql("""create table iud_db.dest (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud_db.dest""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud_db.dest""")
     sql("""delete from dest where c1 IN ('d', 'e')""").show
     checkAnswer(
       sql("""select c1 from dest"""),
@@ -67,7 +66,7 @@ class DeleteCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
   test("delete data from  carbon table[with alias No where clause]") {
     sql("""drop table if exists iud_db.dest""")
     sql("""create table iud_db.dest (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud_db.dest""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud_db.dest""")
     sql("""delete from iud_db.dest a""").show
     checkAnswer(
       sql("""select c1 from iud_db.dest"""),
@@ -77,7 +76,7 @@ class DeleteCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
   test("delete data from  carbon table[No alias No where clause]") {
     sql("""drop table if exists iud_db.dest""")
     sql("""create table iud_db.dest (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud_db.dest""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud_db.dest""")
     sql("""delete from dest""").show()
     checkAnswer(
       sql("""select c1 from dest"""),
@@ -107,7 +106,7 @@ class DeleteCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
   test("delete data from  carbon table[where numeric condition  ]") {
     sql("""drop table if exists iud_db.dest""")
     sql("""create table iud_db.dest (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud_db.dest""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud_db.dest""")
     sql("""delete from  iud_db.dest where c2 >= 4""").show()
     checkAnswer(
       sql("""select count(*) from iud_db.dest"""),

http://git-wip-us.apache.org/repos/asf/incubator-carbondata/blob/af2f204e/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/iud/IUDCompactionTestCases.scala
----------------------------------------------------------------------
diff --git a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/iud/IUDCompactionTestCases.scala b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/iud/IUDCompactionTestCases.scala
index 57555dd..a9d3902 100644
--- a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/iud/IUDCompactionTestCases.scala
+++ b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/iud/IUDCompactionTestCases.scala
@@ -20,7 +20,6 @@
 package org.apache.carbondata.spark.testsuite.iud
 
 import org.apache.spark.sql.Row
-import org.apache.spark.sql.common.util.CarbonHiveContext._
 import org.apache.spark.sql.common.util.QueryTest
 import org.scalatest.BeforeAndAfterAll
 
@@ -38,20 +37,20 @@ class HorizontalCompactionTestCase extends QueryTest with BeforeAndAfterAll {
       """create table iud4.dest (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""")
 
       .show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/comp1.csv' INTO table iud4.dest""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/comp1.csv' INTO table iud4.dest""")
     sql(
       """create table iud4.source2 (c11 string,c22 int,c33 string,c55 string, c66 int) STORED BY 'org.apache.carbondata.format'""")
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/source3.csv' INTO table iud4.source2""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/source3.csv' INTO table iud4.source2""")
     sql("""create table iud4.other (c1 string,c2 int) STORED BY 'org.apache.carbondata.format'""")
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/other.csv' INTO table iud4.other""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/other.csv' INTO table iud4.other""")
     sql(
       """create table iud4.hdest (c1 string,c2 int,c3 string,c5 string) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' LINES TERMINATED BY '\n' STORED AS TEXTFILE""")
       .show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/comp1.csv' INTO table iud4.hdest""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/comp1.csv' INTO table iud4.hdest""")
     sql(
       """CREATE TABLE iud4.update_01(imei string,age int,task bigint,num double,level decimal(10,3),name string)STORED BY 'org.apache.carbondata.format' """)
     sql(
-      """LOAD DATA LOCAL INPATH './src/test/resources/IUD/update01.csv' INTO TABLE iud4.update_01 OPTIONS('BAD_RECORDS_LOGGER_ENABLE' = 'FALSE', 'BAD_RECORDS_ACTION' = 'FORCE') """)
+      s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/update01.csv' INTO TABLE iud4.update_01 OPTIONS('BAD_RECORDS_LOGGER_ENABLE' = 'FALSE', 'BAD_RECORDS_ACTION' = 'FORCE') """)
     CarbonProperties.getInstance()
       .addProperty(CarbonCommonConstants.isHorizontalCompactionEnabled, "true")
   }
@@ -65,14 +64,14 @@ class HorizontalCompactionTestCase extends QueryTest with BeforeAndAfterAll {
     sql(
       """create table dest2 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""")
       .show()
-    sql("""load data local inpath './src/test/resources/IUD/comp1.csv' INTO table dest2""").show()
-    sql("""load data local inpath './src/test/resources/IUD/comp2.csv' INTO table dest2""").show()
-    sql("""load data local inpath './src/test/resources/IUD/comp3.csv' INTO table dest2""").show()
-    sql("""load data local inpath './src/test/resources/IUD/comp4.csv' INTO table dest2""").show()
+    sql(s"""load data local inpath '$resourcesPath/IUD/comp1.csv' INTO table dest2""").show()
+    sql(s"""load data local inpath '$resourcesPath/IUD/comp2.csv' INTO table dest2""").show()
+    sql(s"""load data local inpath '$resourcesPath/IUD/comp3.csv' INTO table dest2""").show()
+    sql(s"""load data local inpath '$resourcesPath/IUD/comp4.csv' INTO table dest2""").show()
     sql(
       """create table source2 (c11 string,c22 int,c33 string,c55 string, c66 int) STORED BY 'org.apache.carbondata.format'""")
       .show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/source3.csv' INTO table source2""").show()
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/source3.csv' INTO table source2""").show()
     sql(
       """update dest2 d set (d.c3, d.c5 ) = (select s.c33,s.c55 from source2 s where d.c1 = s.c11 and s.c22 < 3 or (s.c22 > 10 and s.c22 < 13) or (s.c22 > 20 and s.c22 < 23) or (s.c22 > 30 and s.c22 < 33))""")
       .show()

http://git-wip-us.apache.org/repos/asf/incubator-carbondata/blob/af2f204e/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/iud/UpdateCarbonTableTestCase.scala
----------------------------------------------------------------------
diff --git a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/iud/UpdateCarbonTableTestCase.scala b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/iud/UpdateCarbonTableTestCase.scala
index 3a408f3..b4c6940 100644
--- a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/iud/UpdateCarbonTableTestCase.scala
+++ b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/iud/UpdateCarbonTableTestCase.scala
@@ -18,29 +18,29 @@
  */
 package org.apache.carbondata.spark.testsuite.iud
 
-import org.apache.carbondata.core.constants.CarbonCommonConstants
-import org.apache.carbondata.core.util.CarbonProperties
 import org.apache.spark.sql.Row
-import org.apache.spark.sql.common.util.CarbonHiveContext._
 import org.apache.spark.sql.common.util.QueryTest
 import org.scalatest.BeforeAndAfterAll
 
+import org.apache.carbondata.core.constants.CarbonCommonConstants
+import org.apache.carbondata.core.util.CarbonProperties
+
 class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
   override def beforeAll {
 
     sql("drop database if exists iud cascade")
-    sql("create database  iud")
+    sql("create database iud")
     sql("use iud")
     sql("""create table iud.dest (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest""")
     sql("""create table iud.source2 (c11 string,c22 int,c33 string,c55 string, c66 int) STORED BY 'org.apache.carbondata.format'""")
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/source2.csv' INTO table iud.source2""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/source2.csv' INTO table iud.source2""")
     sql("""create table iud.other (c1 string,c2 int) STORED BY 'org.apache.carbondata.format'""")
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/other.csv' INTO table iud.other""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/other.csv' INTO table iud.other""")
     sql("""create table iud.hdest (c1 string,c2 int,c3 string,c5 string) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' LINES TERMINATED BY '\n' STORED AS TEXTFILE""").show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.hdest""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.hdest""")
     sql("""CREATE TABLE iud.update_01(imei string,age int,task bigint,num double,level decimal(10,3),name string)STORED BY 'org.apache.carbondata.format' """)
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/update01.csv' INTO TABLE iud.update_01 OPTIONS('BAD_RECORDS_LOGGER_ENABLE' = 'FALSE', 'BAD_RECORDS_ACTION' = 'FORCE') """)
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/update01.csv' INTO TABLE iud.update_01 OPTIONS('BAD_RECORDS_LOGGER_ENABLE' = 'FALSE', 'BAD_RECORDS_ACTION' = 'FORCE') """)
     CarbonProperties.getInstance()
       .addProperty(CarbonCommonConstants.isHorizontalCompactionEnabled , "true")
   }
@@ -49,7 +49,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
   test("test update operation with 0 rows updation.") {
     sql("""drop table iud.zerorows""").show
     sql("""create table iud.zerorows (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.zerorows""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.zerorows""")
     sql("""update zerorows d  set (d.c2) = (d.c2 + 1) where d.c1 = 'a'""").show()
     sql("""update zerorows d  set (d.c2) = (d.c2 + 1) where d.c1 = 'xxx'""").show()
      checkAnswer(
@@ -65,7 +65,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
   test("update carbon table[select from source table with where and exist]") {
       sql("""drop table iud.dest11""").show
       sql("""create table iud.dest11 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-      sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest11""")
+      sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest11""")
       sql("""update iud.dest11 d set (d.c3, d.c5 ) = (select s.c33,s.c55 from iud.source2 s where d.c1 = s.c11) where 1 = 1""").show()
       checkAnswer(
         sql("""select c3,c5 from iud.dest11"""),
@@ -77,7 +77,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
    test("update carbon table[using destination table columns with where and exist]") {
     sql("""drop table iud.dest22""").show
     sql("""create table iud.dest22 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest22""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest22""")
     checkAnswer(
       sql("""select c2 from iud.dest22 where c1='a'"""),
       Seq(Row(1))
@@ -93,7 +93,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
    test("update carbon table without alias in set columns") {
       sql("""drop table iud.dest33""").show
       sql("""create table iud.dest33 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-      sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest33""")
+      sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest33""")
       sql("""update iud.dest33 d set (c3,c5 ) = (select s.c33 ,s.c55  from iud.source2 s where d.c1 = s.c11) where d.c1 = 'a'""").show()
       checkAnswer(
         sql("""select c3,c5 from iud.dest33 where c1='a'"""),
@@ -105,7 +105,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
    test("update carbon table without alias in set three columns") {
      sql("""drop table iud.dest44""").show
      sql("""create table iud.dest44 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-     sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest44""")
+     sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest44""")
      sql("""update iud.dest44 d set (c1,c3,c5 ) = (select s.c11, s.c33 ,s.c55  from iud.source2 s where d.c1 = s.c11) where d.c1 = 'a'""").show()
      checkAnswer(
        sql("""select c1,c3,c5 from iud.dest44 where c1='a'"""),
@@ -117,7 +117,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
    test("update carbon table[single column select from source with where and exist]") {
       sql("""drop table iud.dest55""").show
       sql("""create table iud.dest55 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-      sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest55""")
+      sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest55""")
      sql("""update iud.dest55 d set (c3)  = (select s.c33 from iud.source2 s where d.c1 = s.c11) where 1 = 1""").show()
       checkAnswer(
         sql("""select c1,c3 from iud.dest55 """),
@@ -129,7 +129,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
   test("update carbon table[single column SELECT from source with where and exist]") {
     sql("""drop table iud.dest55""").show
     sql("""create table iud.dest55 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest55""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest55""")
     sql("""update iud.dest55 d set (c3)  = (SELECT s.c33 from iud.source2 s where d.c1 = s.c11) where 1 = 1""").show()
     checkAnswer(
       sql("""select c1,c3 from iud.dest55 """),
@@ -141,7 +141,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
    test("update carbon table[using destination table columns without where clause]") {
      sql("""drop table iud.dest66""").show
      sql("""create table iud.dest66 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-     sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest66""")
+     sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest66""")
      sql("""update iud.dest66 d set (c2, c5 ) = (c2 + 1, concat(c5 , "z"))""").show()
      checkAnswer(
        sql("""select c2,c5 from iud.dest66 """),
@@ -153,7 +153,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
    test("update carbon table[using destination table columns with where clause]") {
        sql("""drop table iud.dest77""").show
        sql("""create table iud.dest77 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-       sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest77""")
+       sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest77""")
        sql("""update iud.dest77 d set (c2, c5 ) = (c2 + 1, concat(c5 , "z")) where d.c3 = 'dd'""").show()
        checkAnswer(
          sql("""select c2,c5 from iud.dest77 where c3 = 'dd'"""),
@@ -165,7 +165,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
    test("update carbon table[using destination table( no alias) columns without where clause]") {
      sql("""drop table iud.dest88""").show
      sql("""create table iud.dest88 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-     sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest88""")
+     sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest88""")
      sql("""update iud.dest88  set (c2, c5 ) = (c2 + 1, concat(c5 , "y" ))""").show()
      checkAnswer(
        sql("""select c2,c5 from iud.dest88 """),
@@ -177,7 +177,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
    test("update carbon table[using destination table columns with hard coded value ]") {
      sql("""drop table iud.dest99""").show
      sql("""create table iud.dest99 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-     sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest99""")
+     sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest99""")
      sql("""update iud.dest99 d set (c2, c5 ) = (c2 + 1, "xyx")""").show()
      checkAnswer(
        sql("""select c2,c5 from iud.dest99 """),
@@ -189,7 +189,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
    test("update carbon tableusing destination table columns with hard coded value and where condition]") {
      sql("""drop table iud.dest110""").show
      sql("""create table iud.dest110 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-     sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest110""")
+     sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest110""")
      sql("""update iud.dest110 d set (c2, c5 ) = (c2 + 1, "xyx") where d.c1 = 'e'""").show()
      checkAnswer(
        sql("""select c2,c5 from iud.dest110 where c1 = 'e' """),
@@ -201,7 +201,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
    test("update carbon table[using source  table columns with where and exist and no destination table condition]") {
      sql("""drop table iud.dest120""").show
      sql("""create table iud.dest120 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-     sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest120""")
+     sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest120""")
      sql("""update iud.dest120 d  set (c3, c5 ) = (select s.c33 ,s.c55  from iud.source2 s where d.c1 = s.c11)""").show()
      checkAnswer(
        sql("""select c3,c5 from iud.dest120 """),
@@ -213,7 +213,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
    test("update carbon table[using destination table where and exist]") {
      sql("""drop table iud.dest130""").show
      sql("""create table iud.dest130 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-     sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest130""")
+     sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest130""")
      sql("""update iud.dest130 dd  set (c2, c5 ) = (c2 + 1, "xyx")  where dd.c1 = 'a'""").show()
      checkAnswer(
        sql("""select c2,c5 from iud.dest130 where c1 = 'a' """),
@@ -225,7 +225,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
    test("update carbon table[using destination table (concat) where and exist]") {
      sql("""drop table iud.dest140""").show
      sql("""create table iud.dest140 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-     sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest140""")
+     sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest140""")
      sql("""update iud.dest140 d set (c2, c5 ) = (c2 + 1, concat(c5 , "z"))  where d.c1 = 'a'""").show()
      checkAnswer(
        sql("""select c2,c5 from iud.dest140 where c1 = 'a'"""),
@@ -237,7 +237,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
    test("update carbon table[using destination table (concat) with  where") {
      sql("""drop table iud.dest150""").show
      sql("""create table iud.dest150 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-     sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest150""")
+     sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest150""")
      sql("""update iud.dest150 d set (c5) = (concat(c5 , "z"))  where d.c1 = 'b'""").show()
      checkAnswer(
        sql("""select c5 from iud.dest150 where c1 = 'b' """),
@@ -294,7 +294,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
  test("""update carbon [special characters  in value- test parsing logic ]""") {
     sql("""drop table iud.dest160""").show
     sql("""create table iud.dest160 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest160""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest160""")
     sql("""update iud.dest160 set(c1) = ("ab\')$*)(&^)")""").show()
     sql("""update iud.dest160 set(c1) =  ('abd$asjdh$adasj$l;sdf$*)$*)(&^')""").show()
     sql("""update iud.dest160 set(c1) =("\\")""").show()
@@ -311,7 +311,7 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
   test("""update carbon [sub query, between and existing in outer condition.(Customer query ) ]""") {
     sql("""drop table iud.dest170""").show
     sql("""create table iud.dest170 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest170""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest170""")
     sql("""update iud.dest170 d set (c3)=(select s.c33 from iud.source2 s where d.c1 = s.c11 and d.c2 = s.c22) where  d.c2 between 1 and 3""").show()
     checkAnswer(
       sql("""select c3 from  iud.dest170 as d where d.c2 between 1 and 3"""),
@@ -323,11 +323,11 @@ class UpdateCarbonTableTestCase extends QueryTest with BeforeAndAfterAll {
   test("""update carbon [self join select query ]""") {
     sql("""drop table iud.dest171""").show
     sql("""create table iud.dest171 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest171""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest171""")
     sql("""update iud.dest171 d set (c3)=(select concat(s.c3 , "z") from iud.dest171 s where d.c2 = s.c2)""").show
     sql("""drop table iud.dest172""").show
     sql("""create table iud.dest172 (c1 string,c2 int,c3 string,c5 string) STORED BY 'org.apache.carbondata.format'""").show()
-    sql("""LOAD DATA LOCAL INPATH './src/test/resources/IUD/dest.csv' INTO table iud.dest172""")
+    sql(s"""LOAD DATA LOCAL INPATH '$resourcesPath/IUD/dest.csv' INTO table iud.dest172""")
     sql("""update iud.dest172 d set (c3)=( concat(c3 , "z"))""").show
     checkAnswer(
       sql("""select c3 from  iud.dest171"""),

http://git-wip-us.apache.org/repos/asf/incubator-carbondata/blob/af2f204e/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/joinquery/AllDataTypesTestCaseJoin.scala
----------------------------------------------------------------------
diff --git a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/joinquery/AllDataTypesTestCaseJoin.scala b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/joinquery/AllDataTypesTestCaseJoin.scala
deleted file mode 100644
index 83fdd59..0000000
--- a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/joinquery/AllDataTypesTestCaseJoin.scala
+++ /dev/null
@@ -1,68 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *    http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied.  See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-
-package org.apache.carbondata.spark.testsuite.joinquery
-
-import org.apache.spark.sql.Row
-import org.apache.spark.sql.common.util.CarbonHiveContext._
-import org.apache.spark.sql.common.util.QueryTest
-import org.scalatest.BeforeAndAfterAll
-
-/**
- * Test Class for join query on multiple datatypes
- * @author N00902756
- *
- */
-
-class AllDataTypesTestCaseJoin extends QueryTest with BeforeAndAfterAll {
-
-  override def beforeAll {
-    sql("CREATE TABLE alldatatypestableJoin (empno int, empname String, designation String, doj Timestamp, workgroupcategory int, workgroupcategoryname String, deptno int, deptname String, projectcode int, projectjoindate Timestamp, projectenddate Timestamp,attendance int,utilization int,salary int) STORED BY 'org.apache.carbondata.format'")
-    sql("LOAD DATA local inpath './src/test/resources/data.csv' INTO TABLE alldatatypestableJoin OPTIONS('DELIMITER'= ',', 'QUOTECHAR'= '\"')");
-
-    sql("CREATE TABLE alldatatypestableJoin_hive (empno int, empname String, designation String, doj Timestamp, workgroupcategory int, workgroupcategoryname String, deptno int, deptname String, projectcode int, projectjoindate Timestamp, projectenddate Timestamp,attendance int,utilization int,salary int)row format delimited fields terminated by ','")
-    sql("LOAD DATA local inpath './src/test/resources/datawithoutheader.csv' INTO TABLE alldatatypestableJoin_hive");
-
-  }
-
-  test("select empno,empname,utilization,count(salary),sum(empno) from alldatatypestableJoin where empname in ('arvind','ayushi') group by empno,empname,utilization") {
-    checkAnswer(
-      sql("select empno,empname,utilization,count(salary),sum(empno) from alldatatypestableJoin where empname in ('arvind','ayushi') group by empno,empname,utilization"),
-      sql("select empno,empname,utilization,count(salary),sum(empno) from alldatatypestableJoin_hive where empname in ('arvind','ayushi') group by empno,empname,utilization"))
-  }
-
-  test("select e.empid from employee e inner join manager m on e.mgrid=m.empid") {
-    sql("drop table if exists employee")
-    sql("create table employee(name string, empid string, mgrid string, mobileno bigint) stored by 'carbondata'")
-    sql("load data inpath './src/test/resources/join/emp.csv' into table employee options('fileheader'='name,empid,mgrid,mobileno')")
-    
-    sql("drop table if exists manager")
-    sql("create table manager(name string, empid string, mgrid string, mobileno bigint) stored by 'carbondata'")
-    sql("load data inpath './src/test/resources/join/mgr.csv' into table manager options('fileheader'='name,empid,mgrid,mobileno')")
-    checkAnswer(
-    sql("select e.empid from employee e inner join manager m on e.mgrid=m.empid"),
-    Seq(Row("t23717"))
-    )
-   
-  }
-  override def afterAll {
-    sql("drop table alldatatypestableJoin")
-    sql("drop table alldatatypestableJoin_hive")
-  }
-}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-carbondata/blob/af2f204e/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/joinquery/IntegerDataTypeTestCase.scala
----------------------------------------------------------------------
diff --git a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/joinquery/IntegerDataTypeTestCase.scala b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/joinquery/IntegerDataTypeTestCase.scala
deleted file mode 100644
index e963a4e..0000000
--- a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/joinquery/IntegerDataTypeTestCase.scala
+++ /dev/null
@@ -1,48 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *    http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied.  See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-
-package org.apache.carbondata.spark.testsuite.joinquery
-
-import org.apache.spark.sql.Row
-import org.apache.spark.sql.common.util.CarbonHiveContext._
-import org.apache.spark.sql.common.util.QueryTest
-import org.scalatest.BeforeAndAfterAll
-
-/**
- * Test Class for join query on Integer datatypes
- * @author N00902756
- *
- */
-class IntegerDataTypeTestCase extends QueryTest with BeforeAndAfterAll {
-
-  override def beforeAll {
-    sql("CREATE TABLE integertypetablejoin (empno int, workgroupcategory string, deptno int, projectcode int,attendance int) STORED BY 'org.apache.carbondata.format'")
-    sql("LOAD DATA local inpath './src/test/resources/data.csv' INTO TABLE integertypetablejoin OPTIONS('DELIMITER'= ',', 'QUOTECHAR'= '\"')")
-  }
-
-  test("select empno from integertypetablejoin") {
-    checkAnswer(
-      sql("select empno from integertypetablejoin"),
-      Seq(Row(11), Row(12), Row(13), Row(14), Row(15), Row(16), Row(17), Row(18), Row(19), Row(20)))
-  }
-
-  override def afterAll {
-    sql("drop table integertypetablejoin")
-  }
-}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-carbondata/blob/af2f204e/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/measurenullvalue/NullMeasureValueTestCaseAggregate.scala
----------------------------------------------------------------------
diff --git a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/measurenullvalue/NullMeasureValueTestCaseAggregate.scala b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/measurenullvalue/NullMeasureValueTestCaseAggregate.scala
deleted file mode 100644
index c593a43..0000000
--- a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/measurenullvalue/NullMeasureValueTestCaseAggregate.scala
+++ /dev/null
@@ -1,85 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *    http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied.  See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-package org.apache.carbondata.spark.testsuite.measurenullvalue
-
-import org.apache.spark.sql.Row
-import org.apache.spark.sql.common.util.CarbonHiveContext._
-import org.apache.spark.sql.common.util.QueryTest
-import org.scalatest.BeforeAndAfterAll
-
-import org.apache.carbondata.core.constants.CarbonCommonConstants
-import org.apache.carbondata.core.util.CarbonProperties
-
-class NullMeasureValueTestCaseAggregate extends QueryTest with BeforeAndAfterAll {
-
-  override def beforeAll {
-    sql("drop table IF EXISTS t3")
-    sql(
-      "CREATE TABLE IF NOT EXISTS t3 (ID Int, date Timestamp, country String, name String, " +
-        "phonetype String, serialname String, salary Int) STORED BY 'org.apache.carbondata.format'"
-    )
-    CarbonProperties.getInstance()
-      .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "yyyy/MM/dd")
-    sql("LOAD DATA LOCAL INPATH './src/test/resources/nullmeasurevalue.csv' into table t3");
-  }
-
-  test("select count(salary) from t3") {
-    checkAnswer(
-      sql("select count(salary) from t3"),
-      Seq(Row(0)))
-  }
-  test("select count(ditinct salary) from t3") {
-    checkAnswer(
-      sql("select count(distinct salary) from t3"),
-      Seq(Row(0)))
-  }
-  
-  test("select sum(salary) from t3") {
-    checkAnswer(
-      sql("select sum(salary) from t3"),
-      Seq(Row(null)))
-  }
-  test("select avg(salary) from t3") {
-    checkAnswer(
-      sql("select avg(salary) from t3"),
-      Seq(Row(null)))
-  }
-  
-   test("select max(salary) from t3") {
-    checkAnswer(
-      sql("select max(salary) from t3"),
-      Seq(Row(null)))
-   }
-   test("select min(salary) from t3") {
-    checkAnswer(
-      sql("select min(salary) from t3"),
-      Seq(Row(null)))
-   }
-   test("select sum(distinct salary) from t3") {
-    checkAnswer(
-      sql("select sum(distinct salary) from t3"),
-      Seq(Row(null)))
-   }
-   
-  override def afterAll {
-    sql("drop table t3")
-    CarbonProperties.getInstance()
-      .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "dd-MM-yyyy")
-  }
-}

http://git-wip-us.apache.org/repos/asf/incubator-carbondata/blob/af2f204e/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/nullvalueserialization/TestNullValueSerialization.scala
----------------------------------------------------------------------
diff --git a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/nullvalueserialization/TestNullValueSerialization.scala b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/nullvalueserialization/TestNullValueSerialization.scala
deleted file mode 100644
index 2d84851..0000000
--- a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/nullvalueserialization/TestNullValueSerialization.scala
+++ /dev/null
@@ -1,112 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *    http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied.  See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-
-package org.apache.carbondata.spark.testsuite.nullvalueserialization
-
-import java.io.File
-
-import org.apache.spark.sql.common.util.CarbonHiveContext._
-import org.apache.spark.sql.common.util.QueryTest
-import org.apache.carbondata.core.constants.CarbonCommonConstants
-import org.apache.carbondata.core.util.CarbonProperties
-import org.scalatest.BeforeAndAfterAll
-
-/**
-  * Test cases for testing columns having \N or \null values for non numeric columns
-  */
-class TestNullValueSerialization extends QueryTest with BeforeAndAfterAll {
-
-  override def beforeAll {
-    sql("drop table if exists carbonTable")
-    sql("drop table if exists hiveTable")
-    CarbonProperties.getInstance()
-      .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT,
-        CarbonCommonConstants.CARBON_TIMESTAMP_DEFAULT_FORMAT
-      )
-    val currentDirectory = new File(this.getClass.getResource("/").getPath + "/../../")
-      .getCanonicalPath
-    val csvFilePath = currentDirectory + "/src/test/resources/nullvalueserialization.csv"
-    sql(
-      "CREATE TABLE IF NOT EXISTS carbonTable (ID String, date Timestamp, country String, name " +
-        "String, phonetype String, serialname String, salary Decimal(17,2))STORED BY 'org.apache" +
-        ".carbondata.format'"
-    )
-    sql(
-      "create table if not exists hiveTable(ID String, date Timestamp, country String, name " +
-        "String, " +
-        "phonetype String, serialname String, salary Decimal(17,2))row format delimited fields " +
-        "terminated by ','"
-    )
-    sql(
-      "LOAD DATA LOCAL INPATH '" + csvFilePath + "' into table carbonTable OPTIONS " +
-        "('FILEHEADER'='ID,date," +
-        "country,name,phonetype,serialname,salary')"
-    )
-    sql(
-      "LOAD DATA local inpath '" + csvFilePath + "' INTO table hiveTable"
-    )
-  }
-
-
-  test("test detail query on column having null values") {
-    System.out.println("Carbon Table")
-    sql("select * from carbonTable").show()
-    System.out.println("Hive Table")
-    sql("select * from hiveTable").show()
-    checkAnswer(
-      sql("select * from carbonTable"),
-      sql("select * from hiveTable")
-    )
-  }
-
-    test("test filter query on column is null") {
-      checkAnswer(
-        sql("select * from carbonTable where salary is null"),
-        sql("select * from hiveTable where salary is null")
-      )
-    }
-
-    test("test filter query on column is not null") {
-      checkAnswer(
-        sql("select * from carbonTable where salary is not null"),
-        sql("select * from hiveTable where salary is not null")
-      )
-    }
-
-    test("test filter query on columnValue=null") {
-      checkAnswer(
-        sql("select * from carbonTable where salary=null"),
-        sql("select * from hiveTable where salary=null")
-      )
-    }
-
-    test("test filter query where date is null") {
-      checkAnswer(
-        sql("select * from carbonTable where date is null"),
-        sql("select * from hiveTable where date is null")
-      )
-    }
-
-  override def afterAll {
-    sql("drop table if exists carbonTable")
-    sql("drop table if exists hiveTable")
-    CarbonProperties.getInstance()
-      .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "dd-MM-yyyy")
-  }
-}



Mime
View raw message