Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id C3387200C47 for ; Thu, 30 Mar 2017 19:08:48 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id C1BC8160B8B; Thu, 30 Mar 2017 17:08:48 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 11FB9160B7E for ; Thu, 30 Mar 2017 19:08:47 +0200 (CEST) Received: (qmail 48956 invoked by uid 500); 30 Mar 2017 17:08:44 -0000 Mailing-List: contact issues-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Delivered-To: mailing list issues@hbase.apache.org Received: (qmail 48942 invoked by uid 99); 30 Mar 2017 17:08:44 -0000 Received: from pnap-us-west-generic-nat.apache.org (HELO spamd3-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 30 Mar 2017 17:08:44 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd3-us-west.apache.org (ASF Mail Server at spamd3-us-west.apache.org) with ESMTP id 58749188A0F for ; Thu, 30 Mar 2017 17:08:44 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd3-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: -99.202 X-Spam-Level: X-Spam-Status: No, score=-99.202 tagged_above=-999 required=6.31 tests=[KAM_ASCII_DIVIDERS=0.8, RP_MATCHES_RCVD=-0.001, SPF_PASS=-0.001, USER_IN_WHITELIST=-100] autolearn=disabled Received: from mx1-lw-eu.apache.org ([10.40.0.8]) by localhost (spamd3-us-west.apache.org [10.40.0.10]) (amavisd-new, port 10024) with ESMTP id y-nV7CU70qnS for ; Thu, 30 Mar 2017 17:08:43 +0000 (UTC) Received: from mailrelay1-us-west.apache.org (mailrelay1-us-west.apache.org [209.188.14.139]) by mx1-lw-eu.apache.org (ASF Mail Server at mx1-lw-eu.apache.org) with ESMTP id F3A395FCBA for ; Thu, 30 Mar 2017 17:08:42 +0000 (UTC) Received: from jira-lw-us.apache.org (unknown [207.244.88.139]) by mailrelay1-us-west.apache.org (ASF Mail Server at mailrelay1-us-west.apache.org) with ESMTP id 4E406E0AFC for ; Thu, 30 Mar 2017 17:08:42 +0000 (UTC) Received: from jira-lw-us.apache.org (localhost [127.0.0.1]) by jira-lw-us.apache.org (ASF Mail Server at jira-lw-us.apache.org) with ESMTP id B2F9D21DDF for ; Thu, 30 Mar 2017 17:08:41 +0000 (UTC) Date: Thu, 30 Mar 2017 17:08:41 +0000 (UTC) From: "huaxiang sun (JIRA)" To: issues@hbase.apache.org Message-ID: In-Reply-To: References: Subject: [jira] [Commented] (HBASE-17215) Separate small/large file delete threads in HFileCleaner to accelerate archived hfile cleanup speed MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit X-JIRA-FingerPrint: 30527f35849b9dde25b450d4833f0394 archived-at: Thu, 30 Mar 2017 17:08:48 -0000 [ https://issues.apache.org/jira/browse/HBASE-17215?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15949419#comment-15949419 ] huaxiang sun commented on HBASE-17215: -------------------------------------- Thanks [~carp84], it is definitely fine to go with a follow-up to address the shared queue. {quote} I'm very much interested in this "some reason" (smile). Ever dig into it and mind share if so? Thanks. huaxiang sun {quote} We have some rough idea about it. First, hdfs operation is slow, it takes 120 ms ~ 140 ms to delete a file from hdfs. I think this Jira will help to address this issue (speed up + delete large files to free up the diskspace). Another issue is related with SnapshotHFileCleaner, refresh snapshot file cache is called very frequently, still looking into code to see if something can be done there. > Separate small/large file delete threads in HFileCleaner to accelerate archived hfile cleanup speed > --------------------------------------------------------------------------------------------------- > > Key: HBASE-17215 > URL: https://issues.apache.org/jira/browse/HBASE-17215 > Project: HBase > Issue Type: Improvement > Reporter: Yu Li > Assignee: Yu Li > Attachments: HBASE-17215.patch, HBASE-17215.v2.patch > > > When using PCIe-SSD the flush speed will be really quick, and although we have per CF flush, we still have the {{hbase.regionserver.optionalcacheflushinterval}} setting and some other mechanism to avoid data kept in memory for too long to flush small hfiles. In our online environment we found the single thread cleaner kept cleaning earlier flushed small files while large files got no chance, which caused disk full then many other problems. > Deleting hfiles in parallel with too many threads will also increase the workload of namenode, so here we propose to separate large/small hfile cleaner threads just like we do for compaction, and it turned out to work well in our cluster. -- This message was sent by Atlassian JIRA (v6.3.15#6346)