Return-Path: X-Original-To: apmail-hbase-user-archive@www.apache.org Delivered-To: apmail-hbase-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id E709F179CA for ; Wed, 3 Jun 2015 00:18:17 +0000 (UTC) Received: (qmail 7171 invoked by uid 500); 2 Jun 2015 23:16:15 -0000 Delivered-To: apmail-hbase-user-archive@hbase.apache.org Received: (qmail 7005 invoked by uid 500); 2 Jun 2015 23:16:15 -0000 Mailing-List: contact user-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hbase.apache.org Delivered-To: mailing list user@hbase.apache.org Received: (qmail 6981 invoked by uid 99); 2 Jun 2015 23:16:15 -0000 Received: from Unknown (HELO spamd3-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 02 Jun 2015 23:16:15 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd3-us-west.apache.org (ASF Mail Server at spamd3-us-west.apache.org) with ESMTP id 8DA61181A4E; Tue, 2 Jun 2015 23:16:14 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd3-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: 2.879 X-Spam-Level: ** X-Spam-Status: No, score=2.879 tagged_above=-999 required=6.31 tests=[DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, HTML_MESSAGE=3, RCVD_IN_MSPIKE_H3=-0.01, RCVD_IN_MSPIKE_WL=-0.01, SPF_PASS=-0.001] autolearn=disabled Authentication-Results: spamd3-us-west.apache.org (amavisd-new); dkim=pass (2048-bit key) header.d=gmail.com Received: from mx1-eu-west.apache.org ([10.40.0.8]) by localhost (spamd3-us-west.apache.org [10.40.0.10]) (amavisd-new, port 10024) with ESMTP id 3e9jqu1EbWO9; Tue, 2 Jun 2015 23:16:12 +0000 (UTC) Received: from mail-ig0-f180.google.com (mail-ig0-f180.google.com [209.85.213.180]) by mx1-eu-west.apache.org (ASF Mail Server at mx1-eu-west.apache.org) with ESMTPS id 882DA23131; Tue, 2 Jun 2015 23:16:11 +0000 (UTC) Received: by igbhj9 with SMTP id hj9so98758602igb.1; Tue, 02 Jun 2015 16:16:10 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :cc:content-type; bh=IlRYkK0bYs7McM3DMm+umJ7RSXa8J+R5EJi6pNV7cJA=; b=fGVg92MVQlgihcyxpqd/kTAF7imW1aMStIPoLbZHau/e1WzcWdzhPLUC1xkcLTjI01 VIycmozEwMNuuuZtRciKxUj0uGSy+d/GpxLm6ejfBa1+Wb+jkC0cX+Q003/CDX5wUNHe DXPrVmKot02kIhdTzuZ4GGRcIwFPr4R199oBXkuoQu1hD40tkqr7E/ah+bO0bMYXz/W/ HGt9OYAcCMcBiZWQxkCRtV0RwCkf3kOdyXfqXVXAJBaYM6m1YbDSw0DqLTwQUtU65dBH fxnx8k20+HAHbQMbn33M4kyiUE3GM32rwtaYszYT+2/Y04QCuJK85wUrTpj5bjPw2TP/ h/Vg== MIME-Version: 1.0 X-Received: by 10.50.79.196 with SMTP id l4mr23638596igx.48.1433286970528; Tue, 02 Jun 2015 16:16:10 -0700 (PDT) Received: by 10.107.154.144 with HTTP; Tue, 2 Jun 2015 16:16:10 -0700 (PDT) In-Reply-To: References: Date: Tue, 2 Jun 2015 16:16:10 -0700 Message-ID: Subject: Re: PhoenixIOException resolved only after compaction, is there a way to avoid it? From: Siva To: user@phoenix.apache.org Cc: "user@hbase.apache.org" , dev , hbase-dev Content-Type: multipart/alternative; boundary=089e01183f48bd7ebd0517912043 --089e01183f48bd7ebd0517912043 Content-Type: text/plain; charset=UTF-8 Thanks Vlad for your response. After the changing the below parameters from their default values in hbase-site.xml, queries are working fine. hbase.regionserver.lease.period 1200000 hbase.rpc.timeout 1200000 hbase.client.scanner.caching 1000 Still there are few quires taking a lot of time. Join between two tables is taking more than 5 mins with filter condition, if i omit the filter condition query is failing at all. table1 - 5.5M records -- 2 GB of compressed data table2 - 8.5M records -- 2 GB of compressed data. We have 2 GB of heap space on 4 region servers. 2GB of heap space on master. No activity is going on the cluster when I was running the queries. Do you recommend any of the parameters to tune memory and GC for Phoenix and Hbase? Thanks, Siva. On Mon, Jun 1, 2015 at 1:14 PM, Vladimir Rodionov wrote: > >> Is IO exception is because of Phoenix > >> not able to read from multiple regions since error was resolved after > the > >> compaction? or Any other thoughts? > > Compaction does not decrease # of regions - it sorts/merges data into a > single file (in case of a major compaction) for every > region/column_family. SocketTimeout exception is probably because Phoenix > must read data from multiple files in > every region before compaction - this requires more CPU, more RAM and > produces more temp garbage. > Excessive GC activity, in turn, results in socket timeouts. Check GC logs > in RS and check RS logs for other errors - > they will probably give you a clue on what is going on during a query > execution. > > -Vlad > > > > On Mon, Jun 1, 2015 at 11:10 AM, Siva wrote: > >> Hi Everyone, >> >> We load the data to Hbase tables through BulkImports. >> >> If the data set is small, we can query the imported data from phoenix with >> no issues. >> >> If data size is huge (with respect to our cluster, we have very small >> cluster), I m encountering the following error >> (org.apache.phoenix.exception.PhoenixIOException). >> >> 0: jdbc:phoenix:172.31.45.176:2181:/hbase> select count(*) >> . . . . . . . . . . . . . . . . . . . . .> from "ldll_compression" >> ldll join "ds_compression" ds on (ds."statusid" = ldll."statusid") >> . . . . . . . . . . . . . . . . . . . . .> where ldll."logdate" >= >> '2015-02-04' >> . . . . . . . . . . . . . . . . . . . . .> and ldll."logdate" <= >> '2015-02-06' >> . . . . . . . . . . . . . . . . . . . . .> and ldll."dbname" = >> 'lmguaranteedrate'; >> +------------------------------------------+ >> | COUNT(1) | >> +------------------------------------------+ >> java.lang.RuntimeException: >> org.apache.phoenix.exception.PhoenixIOException: >> org.apache.phoenix.exception.PhoenixIOException: Failed after attempts=36, >> exceptions: >> Mon Jun 01 13:50:57 EDT 2015, null, java.net.SocketTimeoutException: >> callTimeout=60000, callDuration=62358: row '' on table 'ldll_compression' >> at >> region=ldll_compression,,1432851434288.1a8b511def7d0c9e69a5491c6330d715., >> hostname=ip-172-31-32-181.us-west-2.compute.internal,60020,1432768597149, >> seqNum=16566 >> >> at sqlline.SqlLine$IncrementalRows.hasNext(SqlLine.java:2440) >> at sqlline.SqlLine$TableOutputFormat.print(SqlLine.java:2074) >> at sqlline.SqlLine.print(SqlLine.java:1735) >> at sqlline.SqlLine$Commands.execute(SqlLine.java:3683) >> at sqlline.SqlLine$Commands.sql(SqlLine.java:3584) >> at sqlline.SqlLine.dispatch(SqlLine.java:821) >> at sqlline.SqlLine.begin(SqlLine.java:699) >> at sqlline.SqlLine.mainWithInputRedirection(SqlLine.java:441) >> at sqlline.SqlLine.main(SqlLine.java:424) >> >> I did the major compaction for "ldll_compression" through Hbase >> shell(major_compact 'ldll_compression'). Same query ran successfully after >> the compaction. >> >> 0: jdbc:phoenix:172.31.45.176:2181:/hbase> select count(*) >> . . . . . . . . . . . . . . . . . . . . .> from "ldll_compression" >> ldll join "ds_compression" ds on (ds."statusid" = ldll."statusid") >> . . . . . . . . . . . . . . . . . . . . .> where ldll."logdate" >= >> '2015-02-04' >> . . . . . . . . . . . . . . . . . . . . .> and ldll."logdate" <= >> '2015-02-06' >> . . . . . . . . . . . . . . . . . . . . .> and ldll."dbname" = >> 'lmguaranteedrate' >> . . . . . . . . . . . . . . . . . . . . .> ; >> +------------------------------------------+ >> | COUNT(1) | >> +------------------------------------------+ >> | 13480 | >> +------------------------------------------+ >> 1 row selected (72.36 seconds) >> >> Did anyone face the similar issue? Is IO exception is because of Phoenix >> not able to read from multiple regions since error was resolved after the >> compaction? or Any other thoughts? >> >> Thanks, >> Siva. >> > > --089e01183f48bd7ebd0517912043--