carbondata-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Vandana Yadav (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (CARBONDATA-2007) Unexpected behavior while inserting data into a partitioned table
Date Thu, 18 Jan 2018 11:36:00 GMT

    [ https://issues.apache.org/jira/browse/CARBONDATA-2007?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16330401#comment-16330401
] 

Vandana Yadav commented on CARBONDATA-2007:
-------------------------------------------

its working fine as there was one column was missing in create table query.

> Unexpected behavior while inserting data into a partitioned table
> -----------------------------------------------------------------
>
>                 Key: CARBONDATA-2007
>                 URL: https://issues.apache.org/jira/browse/CARBONDATA-2007
>             Project: CarbonData
>          Issue Type: Bug
>          Components: data-load
>    Affects Versions: 1.3.0
>         Environment: spark 2.1
>            Reporter: Vandana Yadav
>            Assignee: anubhav tarar
>            Priority: Major
>         Attachments: 2000_UniqData.csv
>
>
> Unexpected behavior while inserting data into a partitioned table
> Steps to reproduce
> 1) Create hive table:
> a) CREATE TABLE uniqdata_hive (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string,
DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1
decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,
INTEGER_COLUMN1 int)ROW FORMAT DELIMITED FIELDS TERMINATED BY ',';
> b) Load data:
> LOAD DATA LOCAL INPATH '/home/knoldus/Desktop/csv/TestData/Data/uniqdata/2000_UniqData.csv'
into table UNIQDATA_HIVE;
> 2) create Partoitioned table:
> a) CREATE TABLE uniqdata_bigint (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string,
DOB timestamp, DOJ timestamp,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2
decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double, INTEGER_COLUMN1 int) partitioned
by (BIGINT_COLUMN1 bigint) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"=
"256 MB");
> b) CREATE TABLE uniqdata_double (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string,
DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint, DECIMAL_COLUMN2
decimal(36,10), Double_COLUMN2 double, INTEGER_COLUMN1 int) partitioned by (Double_COLUMN1
double) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= "256 MB");
> 3) Load data into the partitioned table from hive table:
> a) insert into uniqdata_bigint partition(BIGINT_COLUMN1)select * from uniqdata_hive limit
5;
> output:
> +---------+--+
> | Result  |
> +---------+--+
> +---------+--+
> b) select * from uniqdata_bigint order by cust_id;
> output:
> +----------+------------------+----------------------------+------------------------+-------+-----------------+------------------+------------------+-----------------+-----------------+------------------+-----------------+--+
> | cust_id  |    cust_name     |    active_emui_version     |          dob           |
 doj  | bigint_column2  | decimal_column1  | decimal_column2  | double_column1  | double_column2
 | integer_column1  | bigint_column1  |
> +----------+------------------+----------------------------+------------------------+-------+-----------------+------------------+------------------+-----------------+-----------------+------------------+-----------------+--+
> | NULL     |                  |                            | NULL                   |
NULL  | NULL            | NULL             | NULL             | NULL            | NULL   
        | NULL             | NULL            |
> | NULL     |                  |                            | 1970-01-01 11:00:03.0  |
NULL  | NULL            | NULL             | NULL             | NULL            | NULL   
        | NULL             | NULL            |
> | NULL     |                  | ACTIVE_EMUI_VERSION_00000  | NULL                   |
NULL  | NULL            | NULL             | NULL             | NULL            | NULL   
        | NULL             | NULL            |
> | NULL     | CUST_NAME_00000  |                            | NULL                   |
NULL  | NULL            | NULL             | NULL             | NULL            | NULL   
        | NULL             | NULL            |
> | 8999     |                  |                            | NULL                   |
NULL  | NULL            | NULL             | NULL             | NULL            | NULL   
        | NULL             | NULL            |
> +----------+------------------+----------------------------+------------------------+-------+-----------------+------------------+------------------+-----------------+-----------------+------------------+-----------------+--+
> 5 rows selected (0.769 seconds)
> c)insert into uniqdata_double partition(Double_COLUMN1)select * from uniqdata_hive limit
5;
> Expected Output: data should be inserted successfully into the table.
> Actual Output:
> Error: org.apache.spark.sql.AnalysisException: Cannot insert into table `newcarbon`.`uniqdata_double`
because the number of columns are different: need 11 columns, but query has 12 columns.; (state=,code=0)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Mime
View raw message