Return-Path: X-Original-To: apmail-hadoop-mapreduce-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-mapreduce-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 36C4FD6B8 for ; Mon, 1 Oct 2012 17:12:55 +0000 (UTC) Received: (qmail 6554 invoked by uid 500); 1 Oct 2012 17:12:50 -0000 Delivered-To: apmail-hadoop-mapreduce-user-archive@hadoop.apache.org Received: (qmail 6455 invoked by uid 500); 1 Oct 2012 17:12:50 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 6447 invoked by uid 99); 1 Oct 2012 17:12:50 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 01 Oct 2012 17:12:50 +0000 X-ASF-Spam-Status: No, hits=2.2 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_NEUTRAL X-Spam-Check-By: apache.org Received-SPF: neutral (nike.apache.org: local policy) Received: from [209.85.220.176] (HELO mail-vc0-f176.google.com) (209.85.220.176) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 01 Oct 2012 17:12:43 +0000 Received: by vcbgb22 with SMTP id gb22so7466675vcb.35 for ; Mon, 01 Oct 2012 10:12:23 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20120113; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type:x-gm-message-state; bh=8ZSpvt4kcyAdkUIMCONmtbUWkChJgrSRI8Pk0UtLp+o=; b=YfI5kTBmbIFMVcYKMPACsPNRDxEvI0gGMGJHDQcHDgIg6ESS+GK4UEqT/s72IBNm79 lXsdPFD+8fyf464Yf8bFA7Pu6yYkDFUXEWXp29G5j/XJ9ysCVLAEgqxJ19PlHri5DUJ/ 9WGJmb6I6866SIauFRl8ggBSWefnAmJK6XO6xlAG3yJ2tFKoRjevDaO/O0Fj51XYmqIt QfP48+6P5FwHFTV/DnZmUpe/2pLy/aOzeIYrpgEEyRnZCb4giy1svIwT64ORHqt7pinG diytZCtKKojiToGBPs0TGZliT88R5DEO4Qa27KI5IEMebQ7gqX19BPxKp9exC6mioIZ3 Jkng== MIME-Version: 1.0 Received: by 10.52.89.35 with SMTP id bl3mr6646596vdb.87.1349111542995; Mon, 01 Oct 2012 10:12:22 -0700 (PDT) Received: by 10.58.124.227 with HTTP; Mon, 1 Oct 2012 10:12:22 -0700 (PDT) In-Reply-To: <5069C0B0.2090201@cs.uni-kassel.de> References: <5069C0B0.2090201@cs.uni-kassel.de> Date: Mon, 1 Oct 2012 10:12:22 -0700 Message-ID: Subject: Re: HDFS "file" missing a part-file From: Robert Molina To: user@hadoop.apache.org Content-Type: multipart/alternative; boundary=20cf307d034e485a0f04cb0282ac X-Gm-Message-State: ALoCoQlcXiGLdsjvsVmfxi2ud9AH77A1DwN9bFXV6drHHmPhVvEMhSsVk7BfxU4Xud8AWX39CLdu --20cf307d034e485a0f04cb0282ac Content-Type: text/plain; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable Hi Bjorn, Can you post the exception you are getting during the map phase? On Mon, Oct 1, 2012 at 9:11 AM, Bj=F6rn-Elmar Macek wrote: > Hi, > > i am kind of unsure where to post this problem, but i think it is more > related to hadoop than to pig. > > By successfully executing a pig script i created a new file in my hdfs. > Sadly though, i cannot use it for further processing except for "dump"ing > and viewing the data: every data-manipulation script-command just as > "foreach" gives exceptions during the map phase. > Since there was no problem executing the same script on the first 100 > lines of my data (LIMIT statement),i copied it to my local fs folder. > What i realized is, that one of the files namely part-r-000001 was empty > and contained within the _temporary folder. > > Is there any reason for this? How can i fix this issue? Did the job (whic= h > created the file we are talking about) NOT run properly til its end, > although the tasktracker worked til the very end and the file was created= ? > > Best regards, > Bj=F6rn > --20cf307d034e485a0f04cb0282ac Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable Hi Bjorn,=A0
Can you post the exception you are getting during the map = phase?



On Mon, = Oct 1, 2012 at 9:11 AM, Bj=F6rn-Elmar Macek <macek@cs.uni-kassel.de= > wrote:
Hi,

i am kind of unsure where to post this problem, but i think it is more rela= ted to hadoop than to pig.

By successfully executing a pig script i created a new file in my hdfs. Sad= ly though, i cannot use it for further processing except for "dump&quo= t;ing and viewing the data: every data-manipulation script-command just as = "foreach" gives exceptions during the map phase.
Since there was no problem executing the same script on the first 100 lines= of my data (LIMIT statement),i copied it to my local fs folder.
What i realized is, that one of the files namely part-r-000001 was empty an= d contained within the _temporary folder.

Is there any reason for this? How can i fix this issue? Did the job (which = created the file we are talking about) NOT run properly til its end, althou= gh the tasktracker worked til the very end and the file was created?

Best regards,
Bj=F6rn

--20cf307d034e485a0f04cb0282ac--