Return-Path: X-Original-To: apmail-hbase-user-archive@www.apache.org Delivered-To: apmail-hbase-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 804FA10548 for ; Mon, 7 Oct 2013 05:29:55 +0000 (UTC) Received: (qmail 16678 invoked by uid 500); 7 Oct 2013 05:29:42 -0000 Delivered-To: apmail-hbase-user-archive@hbase.apache.org Received: (qmail 16590 invoked by uid 500); 7 Oct 2013 05:29:40 -0000 Mailing-List: contact user-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hbase.apache.org Delivered-To: mailing list user@hbase.apache.org Received: (qmail 16573 invoked by uid 99); 7 Oct 2013 05:29:37 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 07 Oct 2013 05:29:37 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS,T_REMOTE_IMAGE X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of bharathv@cloudera.com designates 74.125.82.53 as permitted sender) Received: from [74.125.82.53] (HELO mail-wg0-f53.google.com) (74.125.82.53) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 07 Oct 2013 05:29:33 +0000 Received: by mail-wg0-f53.google.com with SMTP id x12so6734740wgg.32 for ; Sun, 06 Oct 2013 22:29:12 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:in-reply-to:references:from:date :message-id:subject:to:content-type; bh=tjigTUOzfqg/QzHKXWSBuBJt978lFqaiVkld9G617UI=; b=O6xdyENRaTfZ9QE5q4OVZaIU9wMmIWvQ74968qM/u+3wtNRg6xlPG1FdqVXsyLDeGC 1dl0n03OAm4kpvs8XLIGO7WgSxqIMAuvLBT0dkO/LS07rVJmsPEqbPidrYbpr6RaM8oi zYMrIwCw6974RM6NSsj4YEv4flvmvIXoopABO0wQzVsldgHfz9P8fDbUAZiHHa0Nhz0Y q5Hvt7UaX9lV8FKbYxoWKIyBIBxyaQHo3omIRoJ7B/+YXrK/o8YUMsNCn8opEf8eXPaH dQP+yiFm5Lkxvlzg+hIzHJjPeu4HGM0SOZ0fOt2uLlGE88q3auh1HQUbir/o2VnXSCvM ENBg== X-Gm-Message-State: ALoCoQnZENg+SjNKj6pgb1Dj31s8RsikKQtSZVjhs0GMlGyVKJduOhwMlpQjgz9yG6nXZbRI520T X-Received: by 10.180.86.230 with SMTP id s6mr3126501wiz.64.1381123752720; Sun, 06 Oct 2013 22:29:12 -0700 (PDT) MIME-Version: 1.0 Received: by 10.227.144.9 with HTTP; Sun, 6 Oct 2013 22:28:52 -0700 (PDT) In-Reply-To: <1381123300.88874.YahooMailNeo@web140602.mail.bf1.yahoo.com> References: <1381123300.88874.YahooMailNeo@web140602.mail.bf1.yahoo.com> From: Bharath Vissapragada Date: Mon, 7 Oct 2013 10:58:52 +0530 Message-ID: Subject: Re: HBase Random Read latency > 100ms To: user@hbase.apache.org, lars hofhansl Content-Type: multipart/alternative; boundary=f46d04428624ac11a304e81fee9d X-Virus-Checked: Checked by ClamAV on apache.org --f46d04428624ac11a304e81fee9d Content-Type: text/plain; charset=ISO-8859-1 Adding to what Lars said, you can enable bloom filters on column families for read performance. On Mon, Oct 7, 2013 at 10:51 AM, lars hofhansl wrote: > Have you enabled short circuit reading? See here: > http://hbase.apache.org/book/perf.hdfs.html > > How's your data locality (shown on the RegionServer UI page). > > > How much memory are you giving your RegionServers? > If you reads are truly random and the data set does not fit into the > aggregate cache, you'll be dominated by the disk and network. > Each read would need to bring in a 64k (default) HFile block. If short > circuit reading is not enabled you'll get two or three context switches. > > So I would try: > 1. Enable short circuit reading > 2. Increase the block cache size per RegionServer > 3. Decrease the HFile block size > 4. Make sure your data is local (if it is not, issue a major compaction). > > > -- Lars > > > > ________________________________ > From: Ramu M S > To: user@hbase.apache.org > Sent: Sunday, October 6, 2013 10:01 PM > Subject: HBase Random Read latency > 100ms > > > Hi All, > > My HBase cluster has 8 Region Servers (CDH 4.4.0, HBase 0.94.6). > > Each Region Server is with the following configuration, > 16 Core CPU, 192 GB RAM, 800 GB SATA (7200 RPM) Disk > (Unfortunately configured with RAID 1, can't change this as the Machines > are leased temporarily for a month). > > I am running YCSB benchmark tests on HBase and currently inserting around > 1.8 Billion records. > (1 Key + 7 Fields of 100 Bytes = 724 Bytes per record) > > Currently I am getting a write throughput of around 100K OPS, but random > reads are very very slow, all gets have more than 100ms or more latency. > > I have changed the following default configuration, > 1. HFile Size: 16GB > 2. HDFS Block Size: 512 MB > > Total Data size is around 1.8 TB (Excluding the replicas). > My Table is split into 128 Regions (No pre-splitting used, started with 1 > and grew to 128 over the insertion time) > > Taking some inputs from earlier discussions I have done the following > changes to disable Nagle (In both Client and Server hbase-site.xml, > hdfs-site.xml) > > > hbase.ipc.client.tcpnodelay > true > > > > ipc.server.tcpnodelay > true > > > Ganglia stats shows large CPU IO wait (>30% during reads). > > I agree that disk configuration is not ideal for Hadoop cluster, but as > told earlier it can't change for now. > I feel the latency is way beyond any reported results so far. > > Any pointers on what can be wrong? > > Thanks, > Ramu > -- Bharath Vissapragada --f46d04428624ac11a304e81fee9d--