Return-Path: X-Original-To: apmail-hbase-user-archive@www.apache.org Delivered-To: apmail-hbase-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id C815AE3B3 for ; Fri, 11 Jan 2013 17:37:47 +0000 (UTC) Received: (qmail 69384 invoked by uid 500); 11 Jan 2013 17:37:45 -0000 Delivered-To: apmail-hbase-user-archive@hbase.apache.org Received: (qmail 69329 invoked by uid 500); 11 Jan 2013 17:37:45 -0000 Mailing-List: contact user-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hbase.apache.org Delivered-To: mailing list user@hbase.apache.org Received: (qmail 69321 invoked by uid 99); 11 Jan 2013 17:37:45 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 11 Jan 2013 17:37:45 +0000 X-ASF-Spam-Status: No, hits=-0.7 required=5.0 tests=RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of bryanck@gmail.com designates 209.85.160.52 as permitted sender) Received: from [209.85.160.52] (HELO mail-pb0-f52.google.com) (209.85.160.52) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 11 Jan 2013 17:37:40 +0000 Received: by mail-pb0-f52.google.com with SMTP id ro2so1065801pbb.25 for ; Fri, 11 Jan 2013 09:37:20 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=x-received:content-type:mime-version:subject:from:in-reply-to:date :content-transfer-encoding:message-id:references:to:x-mailer; bh=6cdDY4x4eYPsZFSI/NISYaj1UVGtMvBuEOFssDbWAaE=; b=I9Ur+GbLb06h3DkSrNr2OmdcBpA8w1IaZLBbSLAuXll6axavAjqQNSlVsnw9J3pLP6 lD45OUQcgYg7gDEp1PW13xzYWXXpFDQC3Y6SV1a5mghnxY6wGhnMpC6nOB7cq7nnY8mP 3DmEjnAobXQz2KgD3ZkKHdW83yulChADEonsfx4a3Evc3iPfBCW7lS5+wRG/oM5dXKS7 0Kxohdla2y5/2kM6/RMyCVKzihlguEk/R1z2zDev+8kuVULnEk2Qeg7F30hkyPi+wxMQ C+D0j2MBouAc3v5Ioyj2SYJBYEwkWCI/2wPC9FauRIv+BR++tBHn2wjQTKd0EHW0iqoy Pvnw== X-Received: by 10.66.76.104 with SMTP id j8mr1714911paw.72.1357925840190; Fri, 11 Jan 2013 09:37:20 -0800 (PST) Received: from [10.41.41.236] (71-6-107-162.static-ip.telepacific.net. [71.6.107.162]) by mx.google.com with ESMTPS id kl3sm3126544pbc.15.2013.01.11.09.37.18 (version=TLSv1 cipher=ECDHE-RSA-RC4-SHA bits=128/128); Fri, 11 Jan 2013 09:37:19 -0800 (PST) Content-Type: text/plain; charset=us-ascii Mime-Version: 1.0 (Mac OS X Mail 6.2 \(1499\)) Subject: Re: Maximizing throughput From: Bryan Keller In-Reply-To: <-7724766211969777899@unknownmsgid> Date: Fri, 11 Jan 2013 09:37:17 -0800 Content-Transfer-Encoding: quoted-printable Message-Id: <3053A6C1-CBB4-43C9-87AC-38E3499AF0D2@gmail.com> References: <37A338FB-327D-4FA6-B616-61A6E181CA0D@gmail.com> <-7724766211969777899@unknownmsgid> To: user@hbase.apache.org X-Mailer: Apple Mail (2.1499) X-Virus-Checked: Checked by ClamAV on apache.org Thanks for the responses. I'm running HBase 0.92.1 (Cloudera CDH4). The program is very simple, it inserts batches of rows into a table via = multiple threads. I've tried running it with different parameters = (column count, threads, batch size, etc.), but throughput didn't = improve. I've pasted the code here: http://pastebin.com/gPXfdkPy I have auto flush on (default) as I am inserting rows in batch so don't = need to use the internal HTable write buffer. I've posted my config as well: http://pastebin.com/LVG9h6Z4 The regionservers have 12 cores (24 with HT), 128 GB RAM, 6 SCSI drives = Max throughput is 90-100mb/sec on a drive. I've also tested this on an = EC2 High I/O instance type with 2 SSDs, 64GB RAM, and 8 cores (16 with = HT). Both the EC2 and my colo cluster have the same issue of seemingly = underutilizing resources. I measure disk usage using iostat and measured the theoretical max using = hdparm dd. I use iftop to monitor network bandwidth usage, and used = iperf to test theoretical max. CPU usage I use top and iostat. The maximum write performance I'm getting is usually around 20mb/sec on = a drive (this is my colo cluster) on each of the 2 data nodes. That's = about 20% of the max, and it is only sporadic, not a consistent 20mb/sec = per drive. Network usage also seems to top out around 20% (200mbit/sec) = to each node. CPU usage on each node is around 10%. The problem is more = pronounced on EC2 which has much higher theoretical limits for storage = and network I/O. Copying a 133gb file to HDFS looks like it gives similar performance as = HBase (sporadic disk usage topping out at 20%, low CPU, 30-40% network = I/O) so it seems this is more of an HDFS issue than an HBase issue.