Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id F19FC200BF8 for ; Fri, 13 Jan 2017 11:51:28 +0100 (CET) Received: by cust-asf.ponee.io (Postfix) id F0122160B3F; Fri, 13 Jan 2017 10:51:28 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 1DAB0160B32 for ; Fri, 13 Jan 2017 11:51:27 +0100 (CET) Received: (qmail 25670 invoked by uid 500); 13 Jan 2017 10:51:26 -0000 Mailing-List: contact user-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hbase.apache.org Delivered-To: mailing list user@hbase.apache.org Received: (qmail 25538 invoked by uid 99); 13 Jan 2017 10:51:25 -0000 Received: from pnap-us-west-generic-nat.apache.org (HELO spamd3-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 13 Jan 2017 10:51:25 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd3-us-west.apache.org (ASF Mail Server at spamd3-us-west.apache.org) with ESMTP id 48DD218051E; Fri, 13 Jan 2017 10:51:25 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd3-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: -3.3 X-Spam-Level: X-Spam-Status: No, score=-3.3 tagged_above=-999 required=6.31 tests=[HTML_MESSAGE=2, RCVD_IN_DNSWL_MED=-2.3, RP_MATCHES_RCVD=-2.999, SPF_PASS=-0.001] autolearn=disabled Received: from mx1-lw-eu.apache.org ([10.40.0.8]) by localhost (spamd3-us-west.apache.org [10.40.0.10]) (amavisd-new, port 10024) with ESMTP id 46kH4ys-5Gij; Fri, 13 Jan 2017 10:51:17 +0000 (UTC) Received: from szxga02-in.huawei.com (szxga02-in.huawei.com [119.145.14.65]) by mx1-lw-eu.apache.org (ASF Mail Server at mx1-lw-eu.apache.org) with ESMTPS id F0DBD5FC7B; Fri, 13 Jan 2017 10:51:13 +0000 (UTC) Received: from 172.24.1.47 (EHLO SZXEMA418-HUB.china.huawei.com) ([172.24.1.47]) by szxrg02-dlp.huawei.com (MOS 4.3.7-GA FastPath queued) with ESMTP id DTI14099; Fri, 13 Jan 2017 18:47:52 +0800 (CST) Received: from BLREML407-HUB.china.huawei.com (10.20.4.45) by SZXEMA418-HUB.china.huawei.com (10.82.72.36) with Microsoft SMTP Server (TLS) id 14.3.235.1; Fri, 13 Jan 2017 18:47:51 +0800 Received: from BLREML501-MBB.china.huawei.com ([10.20.5.200]) by BLREML407-HUB.china.huawei.com ([10.20.4.45]) with mapi id 14.03.0301.000; Fri, 13 Jan 2017 16:17:40 +0530 From: Pankaj kr To: "dev@hbase.apache.org" , "user@hbase.apache.org" Subject: Region comapction failed Thread-Topic: Region comapction failed Thread-Index: AdJtinVkQOyKJFlaSUGkANGhzcN1Pw== Date: Fri, 13 Jan 2017 10:47:40 +0000 Message-ID: <74ECFFA8DC3B6847888649793C770FE0A2E46CAE@blreml501-mbb> Accept-Language: en-US, zh-CN Content-Language: en-US X-MS-Has-Attach: X-MS-TNEF-Correlator: x-originating-ip: [10.18.248.96] Content-Type: multipart/alternative; boundary="_000_74ECFFA8DC3B6847888649793C770FE0A2E46CAEblreml501mbb_" MIME-Version: 1.0 X-CFilter-Loop: Reflected X-Mirapoint-Virus-RAPID-Raw: score=unknown(0), refid=str=0001.0A090202.5878B058.00C3,ss=1,re=0.000,recu=0.000,reip=0.000,cl=1,cld=1,fgs=0, ip=0.0.0.0, so=2013-06-18 04:22:30, dmn=2013-03-21 17:37:32 X-Mirapoint-Loop-Id: bdcf075c1d88ac50378abd75b1c6acaa archived-at: Fri, 13 Jan 2017 10:51:29 -0000 --_000_74ECFFA8DC3B6847888649793C770FE0A2E46CAEblreml501mbb_ Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: quoted-printable Hi, We met a weird issue in our production environment. Region compaction is always failing with following errors, 1. 2017-01-10 02:19:10,427 | ERROR | regionserver/RS-HOST/RS-IP:PORT-longCompa= ctions-1483858654825 | Compaction failed Request =3D regionName=3DXXXX., st= oreName=3DXYZ, fileCount=3D6, fileSize=3D100.7 M (3.2 M, 20.8 M, 15.1 M, 20= .9 M, 21.0 M, 19.7 M), priority=3D-5, time=3D1747414906352088 | org.apache.= hadoop.hbase.regionserver.CompactSplitThread$CompactionRunner.doCompaction(= CompactSplitThread.java:562) java.io.IOException: ScanWildcardColumnTracker.checkColumn ran into a colum= n actually smaller than the previous column: XXXXXXX at org.apache.hadoop.hbase.regionserver.ScanWildcardColumnT= racker.checkVersions(ScanWildcardColumnTracker.java:114) at org.apache.hadoop.hbase.regionserver.ScanQueryMatcher.ma= tch(ScanQueryMatcher.java:457) at org.apache.hadoop.hbase.regionserver.StoreScanner.next(S= toreScanner.java:551) at org.apache.hadoop.hbase.regionserver.compactions.Compact= or.performCompaction(Compactor.java:328) at org.apache.hadoop.hbase.regionserver.compactions.Default= Compactor.compact(DefaultCompactor.java:104) at org.apache.hadoop.hbase.regionserver.DefaultStoreEngine$= DefaultCompactionContext.compact(DefaultStoreEngine.java:133) at org.apache.hadoop.hbase.regionserver.HStore.compact(HSto= re.java:1243) at org.apache.hadoop.hbase.regionserver.HRegion.compact(HRe= gion.java:1895) at org.apache.hadoop.hbase.regionserver.CompactSplitThread$= CompactionRunner.doCompaction(CompactSplitThread.java:546) at org.apache.hadoop.hbase.regionserver.CompactSplitThread$= CompactionRunner.run(CompactSplitThread.java:583) at java.util.concurrent.ThreadPoolExecutor.runWorker(Thread= PoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecuto 2. 2017-01-10 02:33:53,009 | ERROR | regionserver/RS-HOST/RS-IP:PORT-longCompa= ctions-1483686810953 | Compaction failed Request =3D regionName=3DYYYYYY, s= toreName=3DABC, fileCount=3D6, fileSize=3D125.3 M (20.9 M, 20.9 M, 20.9 M, = 20.9 M, 20.9 M, 20.9 M), priority=3D-68, time=3D1748294500157323 | org.apac= he.hadoop.hbase.regionserver.CompactSplitThread$CompactionRunner.doCompacti= on(CompactSplitThread.java:562) java.io.IOException: Non-increasing Bloom keys: XXXXXXXXXXXXXXXXXXXXXX afte= r XXXXXXXXXXXX at org.apache.hadoop.hbase.regionserver.StoreFile$Writer.ap= pendGeneralBloomfilter(StoreFile.java:911) at org.apache.hadoop.hbase.regionserver.StoreFile$Writer.ap= pend(StoreFile.java:947) at org.apache.hadoop.hbase.regionserver.compactions.Compact= or.performCompaction(Compactor.java:337) at org.apache.hadoop.hbase.regionserver.compactions.Default= Compactor.compact(DefaultCompactor.java:104) at org.apache.hadoop.hbase.regionserver.DefaultStoreEngine$= DefaultCompactionContext.compact(DefaultStoreEngine.java:133) at org.apache.hadoop.hbase.regionserver.HStore.compact(HSto= re.java:1243) at org.apache.hadoop.hbase.regionserver.HRegion.compact(HRe= gion.java:1895) at org.apache.hadoop.hbase.regionserver.CompactSplitThread$= CompactionRunner.doCompaction(CompactSplitThread.java:546) at org.apache.hadoop.hbase.regionserver.CompactSplitThread$= CompactionRunner.run(CompactSplitThread.java:583) at java.util.concurrent.ThreadPoolExecutor.runWorker(Thread= PoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa= dPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) HBase version : 1.0.2 We have verified all the HFiles in the store using HFilePrettyPrinter with = "k" (checkrow), all report is normal. Full scan is also successful. We don't have the access to the actual data and may be customer wont agree = to share that . Have anyone faced this issue, any pointers will be much appreciated. Thanks & Regards, Pankaj --_000_74ECFFA8DC3B6847888649793C770FE0A2E46CAEblreml501mbb_--