Return-Path: X-Original-To: apmail-hadoop-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 6D53D101AD for ; Fri, 14 Jun 2013 16:50:23 +0000 (UTC) Received: (qmail 17737 invoked by uid 500); 14 Jun 2013 16:50:18 -0000 Delivered-To: apmail-hadoop-user-archive@hadoop.apache.org Received: (qmail 17556 invoked by uid 500); 14 Jun 2013 16:50:18 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 17549 invoked by uid 99); 14 Jun 2013 16:50:17 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 14 Jun 2013 16:50:17 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of ed.serrano@gmail.com designates 209.85.220.175 as permitted sender) Received: from [209.85.220.175] (HELO mail-vc0-f175.google.com) (209.85.220.175) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 14 Jun 2013 16:50:11 +0000 Received: by mail-vc0-f175.google.com with SMTP id hr11so605828vcb.6 for ; Fri, 14 Jun 2013 09:49:51 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type; bh=GjgI35XWmO6uwqFj/otqgs/7SkiuLLUYDuUJZThERoM=; b=m0eCw7X/hP48SQyybMki1YfYxjrAOxgGnJlt4qi/CLmIQx1qe8Gmy1BX9WPBIhHq91 771vmG5ibdeIKC6Wd5WQWRTbM1nAQGY+See+E6AigiRCmIvNQBgWLRwudWW68su61BFL JJf4NE9ON91CGicArwLbll6AFnv4ErpsQ489kC0ofw6pPdZVj65j6kUPwWLs8NJVYtf/ GnwSwCj2bDLKJSH7TRpLNp7PnYFkWRvCiROQjJAcZ5lpbr2cuIfgmsPJlzuFiLdb+4JX /nhh3ia1tHHK4cuFyX+nNjbyhbNg76357+lrWTlDdBOhrVYS+VASzJS/a0Ta3fV4gN4a XTqg== MIME-Version: 1.0 X-Received: by 10.58.22.36 with SMTP id a4mr1230472vef.28.1371228590885; Fri, 14 Jun 2013 09:49:50 -0700 (PDT) Received: by 10.221.12.144 with HTTP; Fri, 14 Jun 2013 09:49:50 -0700 (PDT) In-Reply-To: References: Date: Fri, 14 Jun 2013 11:49:50 -0500 Message-ID: Subject: Re: webhdfs read error after successful pig job From: Ed Serrano To: user@hadoop.apache.org Content-Type: multipart/alternative; boundary=089e013a062410d27f04df2009e9 X-Virus-Checked: Checked by ClamAV on apache.org --089e013a062410d27f04df2009e9 Content-Type: text/plain; charset=ISO-8859-1 You might want to investigate if your issue is aways on the same node. On Fri, Jun 14, 2013 at 11:43 AM, Adam Silberstein wrote: > Hi, > I'm having some trouble with webhdfs read after running a Pig job that > completed successfully. > > Here are some details: > > -I am using Hadoop CDH-4.1.3 and the compatible Pig that goes with this > (0.10.0 I think) > > -The Pig job writes out about 10 files. I'm programmatically attempting > to read each of these with webhdfs soon after pig notifies me the job is > complete. The reads often all succeed. And even in the failure case, most > of the reads still succeed, but one may fail. > > -I wondered if I was facing a race condition where Pig was reporting > success before the file was truly ready to read. However, when I run > WebHDFS read with curl even hours later, the request hangs. In contrast, I > can run 'cat' from the DFS command line and the file is output correctly. > > -I ran fsck over the problem file and it report back totally normal. > > -I looked at the namenode to see why my curl request hangs. I get this > error: > ERROR org.apache.hadoop.security.UserGroupInformation: > PriviledgedActionException as:ubuntu (auth:SIMPLE) > cause:java.io.IOException: Could not reach the block containing the data. > Please try again > (I'm guessing the permissions aren't really the important thing here, the > underlying cause of not reaching the block seems more reasonable). > > -I have a 4 node cluster with replication set to 1. > > > If anyone has seen this, has diagnostic tips, or best of all, a solution, > please let me know! > > Thanks, > Adam > > > -- ------------------------------------- *Ed Serrano* Mobile: 972-897-5443 --089e013a062410d27f04df2009e9 Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable
You might want to investigate if your issue is aways on th= e same node.


On Fri, Jun 14, 2013 at 11:43 AM, Adam Silberstein &= lt;adam@trifacta.com= > wrote:
Hi,
I'm having some trouble with webhdfs read after running a Pig job that = completed successfully.

Here are some details:

-I am using Hadoop CDH-4.1.3 and the compatible Pig that goes with this (0.= 10.0 I think)

-The Pig job writes out about 10 files. =A0I'm programmatically attempt= ing to read each of these with webhdfs soon after pig notifies me the job i= s complete. =A0The reads often all succeed. =A0And even in the failure case= , most of the reads still succeed, but one may fail.

-I wondered if I was facing a race condition where Pig was reporting succes= s before the file was truly ready to read. =A0However, when I run WebHDFS r= ead with curl even hours later, the request hangs. =A0In contrast, I can ru= n 'cat' from the DFS command line and the file is output correctly.=

-I ran fsck over the problem file and it report back totally normal.

-I looked at the namenode to see why my curl request hangs. =A0I get this e= rror:
ERROR org.apache.hadoop.security.UserGroupInformation: PriviledgedActionExc= eption as:ubuntu (auth:SIMPLE) cause:java.io.IOException: Could not reach t= he block containing the data. Please try again
(I'm guessing the permissions aren't really the important thing her= e, the underlying cause of not reaching the block seems more reasonable).
-I have a 4 node cluster with replication set to 1.


If anyone has seen this, has diagnostic tips, or best of all, a solution, p= lease let me know!

Thanks,
Adam





--
----------= ---------------------------
Ed Serrano
Mobile: 972-897= -5443
--089e013a062410d27f04df2009e9--