Return-Path: X-Original-To: apmail-hbase-dev-archive@www.apache.org Delivered-To: apmail-hbase-dev-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 0ABC11041C for ; Mon, 7 Apr 2014 10:20:01 +0000 (UTC) Received: (qmail 2526 invoked by uid 500); 7 Apr 2014 10:19:53 -0000 Delivered-To: apmail-hbase-dev-archive@hbase.apache.org Received: (qmail 2227 invoked by uid 500); 7 Apr 2014 10:19:53 -0000 Mailing-List: contact dev-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hbase.apache.org Delivered-To: mailing list dev@hbase.apache.org Received: (qmail 2211 invoked by uid 99); 7 Apr 2014 10:19:52 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 07 Apr 2014 10:19:52 +0000 X-ASF-Spam-Status: No, hits=-0.7 required=5.0 tests=RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of ashish.singhi@huawei.com designates 119.145.14.64 as permitted sender) Received: from [119.145.14.64] (HELO szxga01-in.huawei.com) (119.145.14.64) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 07 Apr 2014 10:19:46 +0000 Received: from 172.24.2.119 (EHLO szxeml210-edg.china.huawei.com) ([172.24.2.119]) by szxrg01-dlp.huawei.com (MOS 4.3.7-GA FastPath queued) with ESMTP id BTT38034; Mon, 07 Apr 2014 18:19:22 +0800 (CST) Received: from SZXEML406-HUB.china.huawei.com (10.82.67.93) by szxeml210-edg.china.huawei.com (172.24.2.183) with Microsoft SMTP Server (TLS) id 14.3.158.1; Mon, 7 Apr 2014 18:19:21 +0800 Received: from SZXEML513-MBS.china.huawei.com ([169.254.8.124]) by szxeml406-hub.china.huawei.com ([10.82.67.93]) with mapi id 14.03.0158.001; Mon, 7 Apr 2014 18:19:17 +0800 From: ashish singhi To: "dev@hbase.apache.org" CC: "user@hbase.apache.org" Subject: RE: One question regarding bulk load Thread-Topic: One question regarding bulk load Thread-Index: Ac9SOvz5ceUFHicISUeTAuWuHW9+YwADJjjAAACX83A= Date: Mon, 7 Apr 2014 10:19:17 +0000 Message-ID: <8E858C3A39F0D046B420FBA6F75448C02FA973DC@szxeml513-mbs.china.huawei.com> References: <8E858C3A39F0D046B420FBA6F75448C02FA97380@szxeml513-mbs.china.huawei.com> <3FEC773ED933774285EE811E6F24DF7A673FF425@szxeml521-mbx.china.huawei.com> In-Reply-To: <3FEC773ED933774285EE811E6F24DF7A673FF425@szxeml521-mbx.china.huawei.com> Accept-Language: en-US, zh-CN Content-Language: en-US X-MS-Has-Attach: X-MS-TNEF-Correlator: x-originating-ip: [10.18.168.119] Content-Type: text/plain; charset="iso-8859-1" Content-Transfer-Encoding: quoted-printable MIME-Version: 1.0 X-CFilter-Loop: Reflected X-Virus-Checked: Checked by ClamAV on apache.org Yes. Thanks Kashif for pointing it out. There was an empty line at the end = of the file. Regards Ashish -----Original Message----- From: Kashif Jawed Siddiqui [mailto:kashifjs@huawei.com]=20 Sent: 07 April 2014 15:28 To: user@hbase.apache.org Cc: dev@hbase.apache.org Subject: RE: One question regarding bulk load Hi, Please check if your file contains empty lines(maybe in the beginning or t= he end). Since -Dimporttsv.skip.bad.lines=3Dfalse is set, any empty lines will caus= e this error. Regards KASHIF -----Original Message----- From: ashish singhi [mailto:ashish.singhi@huawei.com]=20 Sent: 07 April 2014 13:56 To: user@hbase.apache.org Cc: dev@hbase.apache.org Subject: One question regarding bulk load Hi all. I have one question regarding bulk load. How to load data with table empty column values in few rows using bulk load= tool ? I tried the following simple example in HBase 0.94.11 and Hadoop-2, with ta= ble having three columns and second column value is empty in few rows using= bulk load tool. =D8 Data in file is in below format row0,value1,value0 row1,,value1 row2,value3,value2 row3,,value3 row4,value5,value4 row5,,value5 row6,value7,value6 row7,,value7 row8,value9,value8 =D8 When I execute the command hadoop jar /hbase-0.94.11-security.jar importtsv -Dimporttsv.sk= ip.bad.lines=3Dfalse -Dimporttsv.separator=3D, -Dimporttsv.columns=3DHBASE_= ROW_KEY,cf1:c1,cf1:c2 -Dimporttsv.bulk.output=3D /bulkdata/comma_separated = _3columns comma_separated_3columns /comma_separated_ 3columns.txt I get the below Exception. 2014-04-07 11:15:01,870 INFO [main] mapreduce.Job (Job.java:printTaskEvent= s(1424)) - Task Id : attempt_1396526639698_0028_m_000000_2, Status : FAILED Error: java.io.IOException: org.apache.hadoop.hbase.mapreduce.ImportTsv$Tsv= Parser$BadTsvLineException: No delimiter at org.apache.hadoop.hbase.mapreduce.TsvImporterTextMapper.map(TsvI= mporterTextMapper.java:135) at org.apache.hadoop.hbase.mapreduce.TsvImporterTextMapper.map(TsvI= mporterTextMapper.java:33) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:763) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:339) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:167) Regards, Ashish Singhi