Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 924C8200C46 for ; Wed, 29 Mar 2017 11:02:46 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id 90D6D160B95; Wed, 29 Mar 2017 09:02:46 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id D9C46160B8A for ; Wed, 29 Mar 2017 11:02:45 +0200 (CEST) Received: (qmail 14825 invoked by uid 500); 29 Mar 2017 09:02:45 -0000 Mailing-List: contact issues-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Delivered-To: mailing list issues@hbase.apache.org Received: (qmail 14813 invoked by uid 99); 29 Mar 2017 09:02:45 -0000 Received: from pnap-us-west-generic-nat.apache.org (HELO spamd4-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 29 Mar 2017 09:02:45 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd4-us-west.apache.org (ASF Mail Server at spamd4-us-west.apache.org) with ESMTP id A5E37C03A8 for ; Wed, 29 Mar 2017 09:02:44 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd4-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: -99.202 X-Spam-Level: X-Spam-Status: No, score=-99.202 tagged_above=-999 required=6.31 tests=[KAM_ASCII_DIVIDERS=0.8, RP_MATCHES_RCVD=-0.001, SPF_PASS=-0.001, USER_IN_WHITELIST=-100] autolearn=disabled Received: from mx1-lw-eu.apache.org ([10.40.0.8]) by localhost (spamd4-us-west.apache.org [10.40.0.11]) (amavisd-new, port 10024) with ESMTP id o7_iAW3gjrVl for ; Wed, 29 Mar 2017 09:02:43 +0000 (UTC) Received: from mailrelay1-us-west.apache.org (mailrelay1-us-west.apache.org [209.188.14.139]) by mx1-lw-eu.apache.org (ASF Mail Server at mx1-lw-eu.apache.org) with ESMTP id 4E5B85FE48 for ; Wed, 29 Mar 2017 09:02:43 +0000 (UTC) Received: from jira-lw-us.apache.org (unknown [207.244.88.139]) by mailrelay1-us-west.apache.org (ASF Mail Server at mailrelay1-us-west.apache.org) with ESMTP id 8911DE0674 for ; Wed, 29 Mar 2017 09:02:42 +0000 (UTC) Received: from jira-lw-us.apache.org (localhost [127.0.0.1]) by jira-lw-us.apache.org (ASF Mail Server at jira-lw-us.apache.org) with ESMTP id BAE1724172 for ; Wed, 29 Mar 2017 09:02:41 +0000 (UTC) Date: Wed, 29 Mar 2017 09:02:41 +0000 (UTC) From: "Yu Li (JIRA)" To: issues@hbase.apache.org Message-ID: In-Reply-To: References: Subject: [jira] [Commented] (HBASE-17215) Separate small/large file delete threads in HFileCleaner to accelerate hfile cleanup speed MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit X-JIRA-FingerPrint: 30527f35849b9dde25b450d4833f0394 archived-at: Wed, 29 Mar 2017 09:02:46 -0000 [ https://issues.apache.org/jira/browse/HBASE-17215?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15946755#comment-15946755 ] Yu Li commented on HBASE-17215: ------------------------------- Thanks for the reference [~anoop.hbase], will take a look when making the patch. The issue here exactly targets at accelerating cleaning the archived files, and changes will be mainly in HFileCleaner. > Separate small/large file delete threads in HFileCleaner to accelerate hfile cleanup speed > ------------------------------------------------------------------------------------------ > > Key: HBASE-17215 > URL: https://issues.apache.org/jira/browse/HBASE-17215 > Project: HBase > Issue Type: Improvement > Reporter: Yu Li > Assignee: Yu Li > > When using PCIe-SSD the flush speed will be really quick, and although we have per CF flush, we still have the {{hbase.regionserver.optionalcacheflushinterval}} setting and some other mechanism to avoid data kept in memory for too long to flush small hfiles. In our online environment we found the single thread cleaner kept cleaning earlier flushed small files while large files got no chance, which caused disk full then many other problems. > Deleting hfiles in parallel with too many threads will also increase the workload of namenode, so here we propose to separate large/small hfile cleaner threads just like we do for compaction, and it turned out to work well in our cluster. -- This message was sent by Atlassian JIRA (v6.3.15#6346)