Return-Path: X-Original-To: apmail-hbase-user-archive@www.apache.org Delivered-To: apmail-hbase-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 9F69410475 for ; Mon, 2 Dec 2013 14:38:45 +0000 (UTC) Received: (qmail 2429 invoked by uid 500); 2 Dec 2013 14:38:21 -0000 Delivered-To: apmail-hbase-user-archive@hbase.apache.org Received: (qmail 2385 invoked by uid 500); 2 Dec 2013 14:38:19 -0000 Mailing-List: contact user-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hbase.apache.org Delivered-To: mailing list user@hbase.apache.org Received: (qmail 2373 invoked by uid 99); 2 Dec 2013 14:38:18 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 02 Dec 2013 14:38:18 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: local policy includes SPF record at spf.trusted-forwarder.org) Received: from [209.85.128.177] (HELO mail-ve0-f177.google.com) (209.85.128.177) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 02 Dec 2013 14:38:10 +0000 Received: by mail-ve0-f177.google.com with SMTP id db12so8701072veb.22 for ; Mon, 02 Dec 2013 06:37:50 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:in-reply-to:references:from:date :message-id:subject:to:content-type; bh=5iMi3+n2uqlHdIJCiPbyGlSG2gxKXOp1ehSFjwwyCFo=; b=MY58QXb/S5GQU1aatmTQQmxl4H8Fq+BLcfp4LqB9CnOCLeTd8Mxao/w4+16XezlJZL RX8TrVTKoKicIHWCkbkrQ3ABO5JBQqOFxiIWcff6jHvSPQKAiu8YvP5SiMvCQNMFwh8S kviVYBW0f1pk1ZI+mts4pLC0F3NMLJ58PilIEq7KxdKahBD+PRRC3RTvj3PuQrzCPECC 6103D7AvHY6UZk3Xcxexinjh+L7gQP/q38GZSNxsQf16zq0ph0iDhEKcoa4ctLCCTdcb SlSlKagRieonFEwACTVlxSnGMBifcgHvHgTi/MEK0PGVBO6rnw918/zMOIFR15vJMzgr VWmQ== X-Gm-Message-State: ALoCoQlZvbwtTTUzfL2vitcp2087Ltk6B54hSs3iB8PQz9Za5jJHqSBSEMZ1tzc28VR+ZIx3jIXy X-Received: by 10.221.37.9 with SMTP id tc9mr121350vcb.39.1385995070035; Mon, 02 Dec 2013 06:37:50 -0800 (PST) MIME-Version: 1.0 Received: by 10.52.163.68 with HTTP; Mon, 2 Dec 2013 06:37:29 -0800 (PST) In-Reply-To: References: From: Jean-Marc Spaggiari Date: Mon, 2 Dec 2013 09:37:29 -0500 Message-ID: Subject: Re: What is HBase compaction-queue-size at all? To: user Content-Type: multipart/alternative; boundary=001a11337ed4cf6f1604ec8e1fed X-Virus-Checked: Checked by ClamAV on apache.org --001a11337ed4cf6f1604ec8e1fed Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: quoted-printable - Is it the *number of Store* of regionserver need to be major compacted ? or numbers of which is* being* compacted currently ? This is the number that are currently in the pipe. Doesn.'t mean they are compacting right now, but they are queued for compaction. and not necessary major compaction. Major is only if all the regions need to compact. "I was discovering that at some time it got *regionserver compaction-queue-size =3D 4*.(I check it from Ambari). That's theoretically impossible since I have only *one Store *to write(sequential key) at any time, incurring only one major compaction is more reasonable." Why is this "impossible"? A store file is a dump of HBase memory blocks written into the disk. Even if you write to a single region, single table, with keys all close-by (even if it's all the same exact key). When the block in memory reach a threshold, it's then written into the disk. When more than x blocks (3 is the default) are there in disk, compaction is launched. - Just more confusing is : Isn't multi-thread enabled at earlier version that will allocate each compaction job to a thread , by this reason why there exists compaction queue waiting for processing ? Yes, compaction is done on a separate thread, but there is one single queue. You don't want to take 100% of you RS resources to do compactions... Depending if you are doing mostly writes and almost no reads, you might want to tweek some parameters. And also, you might want to look into bulk loading... Last, maybe you should review you key and distribution. And last again ;) What is your table definition? Multiplying the columns famillies can also sometime lend to this kind of issues... JM 2013/12/2 =E6=9E=97=E7=85=92=E6=B8=85 > Any one knows what compaction queue size is meant? > > By doc's definition: > > *9.2.5.* hbase.regionserver.compactionQueueSize Size of the compaction > queue. This is the number of stores in the region that have been targeted > for compaction. > > > - Is it the *number of Store* of regionserver need to be major compact= ed > ? or numbers of which is* being* compacted currently ? > > I have a job writing data in a hotspot style using sequential key(non > distributed) with 1 family so that 1 Store each region. > > I was discovering that at some time it got *regionserver > compaction-queue-size =3D 4*.(I check it from Ambari). That's theoretical= ly > impossible since I have only *one Store *to write(sequential key) at any > time, incurring only one major compaction is more reasonable. > > > - Then I dig into the logs ,found there is no thing about hints of > queue size > 0: Every major compaction just say *"This selection was = in > queue for 0sec", *I don't really understand what's it to means? is it > saying hbase has nothing in compaction queue? > > 013-11-26 12:28:00,778 INFO > [regionserver60020-smallCompactions-1385440028938] regionserver.HStore: > Completed major compaction of 3 file(s) in f1 of myTable.key.md5.... into > md5....(size=3D607.8 M), total size for store is 645.8 M.*This selection = was > in queue for 0sec*, and took 39sec to execute. > > > - Just more confusing is : Isn't multi-thread enabled at earlier versi= on > that will allocate each compaction job to a thread , by this reason w= hy > there exists compaction queue waiting for processing ? > --001a11337ed4cf6f1604ec8e1fed--