carbondata-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "sounak chakraborty (JIRA)" <j...@apache.org>
Subject [jira] [Issue Comment Deleted] (CARBONDATA-657) We are not able to create table with shared dictionary columns in spark 2.1
Date Fri, 20 Jan 2017 13:34:26 GMT

     [ https://issues.apache.org/jira/browse/CARBONDATA-657?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]

sounak chakraborty updated CARBONDATA-657:
------------------------------------------
    Comment: was deleted

(was: Please reverify. In latest code the output is proper. 

1 row selected (0.204 seconds)
0: jdbc:hive2://localhost:10000> drop table uniqdata_date;
+---------+--+| Result  |
+---------+--+
+---------+--+
No rows selected (0.215 seconds)
0: jdbc:hive2://localhost:10000> CREATE TABLE uniqdata_date (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION
string, DOB date, DOJ date, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10),
DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1
int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= "256 MB");
+---------+--+
| Result  |
+---------+--+
+---------+--+
No rows selected (0.115 seconds)
0: jdbc:hive2://localhost:10000>   LOAD DATA INPATH 'hdfs://localhost:54310/sc/one_row.csv'
into table uniqdata_date OPTIONS('FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
+---------+--+
| Result  |
+---------+--+
+---------+--+
No rows selected (0.913 seconds)
0: jdbc:hive2://localhost:10000> select count(CUST_NAME) from uniqdata_date;
+------+--+| _c0  |
+------+--+
| 1    |
+------+--+
1 row selected (0.237 seconds)
0: jdbc:hive2://localhost:10000> show segments for table uniqdata_date;
+--------------------+----------+--------------------------+--------------------------+--+|
SegmentSequenceId  |  Status  |     Load Start Time      |      Load End Time       |
+--------------------+----------+--------------------------+--------------------------+--+
| 0                  | Success  | 2017-01-20 19:00:48.798  | 2017-01-20 19:00:49.473  |
+--------------------+----------+--------------------------+--------------------------+--+
1 row selected (0.018 seconds)
0: jdbc:hive2://localhost:10000>   LOAD DATA INPATH 'hdfs://localhost:54310/sc/one_row.csv'
into table uniqdata_date OPTIONS('FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');+---------+--+
| Result  |
+---------+--+
+---------+--+
No rows selected (0.363 seconds)
0: jdbc:hive2://localhost:10000> select count(CUST_NAME) from uniqdata_date;
+------+--+| _c0  |
+------+--+
| 2    |
+------+--+
1 row selected (0.219 seconds)
0: jdbc:hive2://localhost:10000> show segments for table uniqdata_date;
+--------------------+----------+--------------------------+--------------------------+--+
| SegmentSequenceId  |  Status  |     Load Start Time      |      Load End Time       |
+--------------------+----------+--------------------------+--------------------------+--+
| 1                  | Success  | 2017-01-20 19:01:01.961  | 2017-01-20 19:01:02.167  |
| 0                  | Success  | 2017-01-20 19:00:48.798  | 2017-01-20 19:00:49.473  |
+--------------------+----------+--------------------------+--------------------------+--+
2 rows selected (0.017 seconds)
0: jdbc:hive2://localhost:10000>   LOAD DATA INPATH 'hdfs://localhost:54310/sc/one_row.csv'
into table uniqdata_date OPTIONS('FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
+---------+--+
| Result  |
+---------+--+
+---------+--+
No rows selected (0.396 seconds)
0: jdbc:hive2://localhost:10000> select count(CUST_NAME) from uniqdata_date;
+------+--+
| _c0  |
+------+--+
| 3    |
+------+--+
1 row selected (0.22 seconds)
0: jdbc:hive2://localhost:10000> show segments for table uniqdata_date;
+--------------------+----------+--------------------------+--------------------------+--+
| SegmentSequenceId  |  Status  |     Load Start Time      |      Load End Time       |
+--------------------+----------+--------------------------+--------------------------+--+
| 2                  | Success  | 2017-01-20 19:01:20.479  | 2017-01-20 19:01:20.734  |
| 1                  | Success  | 2017-01-20 19:01:01.961  | 2017-01-20 19:01:02.167  |
| 0                  | Success  | 2017-01-20 19:00:48.798  | 2017-01-20 19:00:49.473  |
+--------------------+----------+--------------------------+--------------------------+--+
3 rows selected (0.014 seconds)
0: jdbc:hive2://localhost:10000>   LOAD DATA INPATH 'hdfs://localhost:54310/sc/one_row.csv'
into table uniqdata_date OPTIONS('FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
+---------+--+
| Result  |
+---------+--+
+---------+--+
No rows selected (0.357 seconds)
0: jdbc:hive2://localhost:10000> select count(CUST_NAME) from uniqdata_date;
+------+--+
| _c0  |
+------+--+
| 4    |
+------+--+
1 row selected (0.217 seconds)
0: jdbc:hive2://localhost:10000> show segments for table uniqdata_date;
+--------------------+----------+--------------------------+--------------------------+--+
| SegmentSequenceId  |  Status  |     Load Start Time      |      Load End Time       |
+--------------------+----------+--------------------------+--------------------------+--+
| 3                  | Success  | 2017-01-20 19:01:30.826  | 2017-01-20 19:01:31.04   |
| 2                  | Success  | 2017-01-20 19:01:20.479  | 2017-01-20 19:01:20.734  |
| 1                  | Success  | 2017-01-20 19:01:01.961  | 2017-01-20 19:01:02.167  |
| 0                  | Success  | 2017-01-20 19:00:48.798  | 2017-01-20 19:00:49.473  |
+--------------------+----------+--------------------------+--------------------------+--+
4 rows selected (0.022 seconds)
0: jdbc:hive2://localhost:10000>   LOAD DATA INPATH 'hdfs://localhost:54310/sc/one_row.csv'
into table uniqdata_date OPTIONS('FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
+---------+--+
| Result  |
+---------+--+
+---------+--+
No rows selected (0.552 seconds)
0: jdbc:hive2://localhost:10000> select count(CUST_NAME) from uniqdata_date;
+------+--+
| _c0  |
+------+--+
| 5    |
+------+--+
1 row selected (0.233 seconds)
0: jdbc:hive2://localhost:10000> show segments for table uniqdata_date;
+--------------------+----------+--------------------------+--------------------------+--+
| SegmentSequenceId  |  Status  |     Load Start Time      |      Load End Time       |
+--------------------+----------+--------------------------+--------------------------+--+
| 4                  | Success  | 2017-01-20 19:01:48.578  | 2017-01-20 19:01:48.899  |
| 3                  | Success  | 2017-01-20 19:01:30.826  | 2017-01-20 19:01:31.04   |
| 2                  | Success  | 2017-01-20 19:01:20.479  | 2017-01-20 19:01:20.734  |
| 1                  | Success  | 2017-01-20 19:01:01.961  | 2017-01-20 19:01:02.167  |
| 0                  | Success  | 2017-01-20 19:00:48.798  | 2017-01-20 19:00:49.473  |
+--------------------+----------+--------------------------+--------------------------+--+
5 rows selected (0.021 seconds)
0: jdbc:hive2://localhost:10000>   LOAD DATA INPATH 'hdfs://localhost:54310/sc/one_row.csv'
into table uniqdata_date OPTIONS('FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
+---------+--+
| Result  |
+---------+--+
+---------+--+
No rows selected (0.895 seconds)
0: jdbc:hive2://localhost:10000> select count(CUST_NAME) from uniqdata_date;
+------+--+
| _c0  |
+------+--+
| 6    |
+------+--+
1 row selected (0.203 seconds)
0: jdbc:hive2://localhost:10000> show segments for table uniqdata_date;
+--------------------+------------+--------------------------+--------------------------+--+
| SegmentSequenceId  |   Status   |     Load Start Time      |      Load End Time       |
+--------------------+------------+--------------------------+--------------------------+--+
| 5                  | Success    | 2017-01-20 19:01:59.212  | 2017-01-20 19:01:59.493  |
| 4                  | Success    | 2017-01-20 19:01:48.578  | 2017-01-20 19:01:48.899  |
| 3                  | Compacted  | 2017-01-20 19:01:30.826  | 2017-01-20 19:01:31.04   |
| 2                  | Compacted  | 2017-01-20 19:01:20.479  | 2017-01-20 19:01:20.734  |
| 1                  | Compacted  | 2017-01-20 19:01:01.961  | 2017-01-20 19:01:02.167  |
| 0.1                | Success    | 2017-01-20 19:01:59.545  | 2017-01-20 19:01:59.852  |
| 0                  | Compacted  | 2017-01-20 19:00:48.798  | 2017-01-20 19:00:49.473  |
+--------------------+------------+--------------------------+--------------------------+--+
7 rows selected (0.018 seconds)
0: jdbc:hive2://localhost:10000> select count(CUST_NAME) from uniqdata_date;
+------+--+
| _c0  |
+------+--+
| 6    |
+------+--+
1 row selected (0.219 seconds)
)

> We are not able to create table with shared dictionary columns in spark 2.1
> ---------------------------------------------------------------------------
>
>                 Key: CARBONDATA-657
>                 URL: https://issues.apache.org/jira/browse/CARBONDATA-657
>             Project: CarbonData
>          Issue Type: Bug
>          Components: sql
>    Affects Versions: 1.0.0-incubating
>         Environment: Spark-2.1
>            Reporter: Payal
>            Priority: Minor
>
> We are not able to create table with shared dictionary columns not working with spark-2.1
but  it is working fine with spark 1.6 
> 0: jdbc:hive2://hadoop-master:10000> CREATE TABLE uniq_shared_dictionary (CUST_ID
int,CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1
bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1
double, Double_COLUMN2 double,INTEGER_COLUMN1 int) STORED BY 'org.apache.carbondata.format'
TBLPROPERTIES('DICTIONARY_INCLUDE'='CUST_ID,Double_COLUMN2,DECIMAL_COLUMN2','columnproperties.CUST_ID.shared_column'='shared.CUST_ID','columnproperties.decimal_column2.shared_column'='shared.decimal_column2');
> Error: org.apache.carbondata.spark.exception.MalformedCarbonCommandException: Invalid
table properties columnproperties.cust_id.shared_column (state=,code=0)
> LOGS
> ERROR 18-01 13:31:18,147 - Error executing query, currentState RUNNING, 
> org.apache.carbondata.spark.exception.MalformedCarbonCommandException: Invalid table
properties columnproperties.cust_id.shared_column
> at org.apache.carbondata.spark.util.CommonUtil$$anonfun$validateTblProperties$1.apply(CommonUtil.scala:141)
> at org.apache.carbondata.spark.util.CommonUtil$$anonfun$validateTblProperties$1.apply(CommonUtil.scala:137)
> at scala.collection.Iterator$class.foreach(Iterator.scala:893)
> at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
> at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
> at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
> at org.apache.carbondata.spark.util.CommonUtil$.validateTblProperties(CommonUtil.scala:137)
> at org.apache.spark.sql.parser.CarbonSqlAstBuilder.visitCreateTable(CarbonSparkSqlParser.scala:135)
> at org.apache.spark.sql.parser.CarbonSqlAstBuilder.visitCreateTable(CarbonSparkSqlParser.scala:60)
> at org.apache.spark.sql.catalyst.parser.SqlBaseParser$CreateTableContext.accept(SqlBaseParser.java:503)
> at org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visit(AbstractParseTreeVisitor.java:42)
> at org.apache.spark.sql.catalyst.parser.AstBuilder$$anonfun$visitSingleStatement$1.apply(AstBuilder.scala:66)
> at org.apache.spark.sql.catalyst.parser.AstBuilder$$anonfun$visitSingleStatement$1.apply(AstBuilder.scala:66)
> at org.apache.spark.sql.catalyst.parser.ParserUtils$.withOrigin(ParserUtils.scala:93)
> at org.apache.spark.sql.catalyst.parser.AstBuilder.visitSingleStatement(AstBuilder.scala:65)
> at org.apache.spark.sql.catalyst.parser.AbstractSqlParser$$anonfun$parsePlan$1.apply(ParseDriver.scala:54)
> at org.apache.spark.sql.catalyst.parser.AbstractSqlParser$$anonfun$parsePlan$1.apply(ParseDriver.scala:53)
> at org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parse(ParseDriver.scala:82)
> at org.apache.spark.sql.parser.CarbonSparkSqlParser.parse(CarbonSparkSqlParser.scala:45)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Mime
View raw message