Return-Path: X-Original-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 17836F36F for ; Fri, 19 Apr 2013 07:08:58 +0000 (UTC) Received: (qmail 70314 invoked by uid 500); 19 Apr 2013 07:07:25 -0000 Delivered-To: apmail-hadoop-hdfs-user-archive@hadoop.apache.org Received: (qmail 69498 invoked by uid 500); 19 Apr 2013 07:07:22 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 63685 invoked by uid 99); 19 Apr 2013 07:02:59 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 19 Apr 2013 07:02:59 +0000 X-ASF-Spam-Status: No, hits=1.3 required=5.0 tests=FORGED_YAHOO_RCVD,FREEMAIL_ENVFROM_END_DIGIT,RCVD_IN_DNSWL_NONE,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: local policy) Received: from [77.238.189.197] (HELO nm9-vm0.bullet.mail.ird.yahoo.com) (77.238.189.197) by apache.org (qpsmtpd/0.29) with SMTP; Fri, 19 Apr 2013 07:02:51 +0000 Received: from [77.238.189.49] by nm9.bullet.mail.ird.yahoo.com with NNFMP; 19 Apr 2013 07:02:30 -0000 Received: from [217.146.189.111] by tm2.bullet.mail.ird.yahoo.com with NNFMP; 19 Apr 2013 07:02:30 -0000 Received: from [127.0.0.1] by smtp127.mail.ird.yahoo.com with NNFMP; 19 Apr 2013 07:02:30 -0000 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=yahoo.com; s=s1024; t=1366354950; bh=5+1fazdtTYAxVdSGF/M7w4InTwTmZ3ZeEX4f1ebKivU=; h=X-Yahoo-Newman-Id:X-Yahoo-Newman-Property:X-YMail-OSG:X-Yahoo-SMTP:X-Rocket-Received:From:To:References:In-Reply-To:Subject:Date:Message-ID:MIME-Version:Content-Type:Content-Transfer-Encoding:X-Mailer:Thread-Index:Content-Language; b=nt3NaieIklD6v7Dz+oq8TXQ5LYQFiZvPHsT/yvfBHcSxeRmu/Syh87M4qJYufmoS0SUVtaSrN43EZb7+Ra8XZzqSccR7Qk1SWOy9AEEz9wHLV6Dsyz2FXpRoR+PlPRDThJ2ma03KWDwKptF8/mYCRhGA5+ZxnEBCOi5tafvjqfM= X-Yahoo-Newman-Id: 306463.7917.bm@smtp127.mail.ird.yahoo.com X-Yahoo-Newman-Property: ymail-3 X-YMail-OSG: vpjaBxoVM1lwdHNhdyGJkUIJonOOWSZ3dje2WyJQqKNnK29 aEN58fRZ27KZOXZDkHB7xNvkN.1N0qTs29lAIYaQ8e44q_hTp32uD8ZM96PO fLQ7SUtpo7BOykZSX7RKLm2dFuBHfJnz8az39bk5FLsk_LSDgbqipRbIbGzD 4NNNoEsWfGrujYwz3X6ZJTVr1lTkTYy.1r4xhe9_PjuvA20Gli32WoUWri8k g_Ks4ozRpnIWwnBixkEhRYp3H3U064N_1u_Yrbkq9wedaojO8zidCD2ijMaK flk7irUNCFVsP_K3dH3YVyEFdgZyZwLp.nosmTIakS_ynEOEL9NPo1SQbUGW zBW6A0Go1F8wLtO05o4R2DSFRSXbnurXrqNx9KgSYh6a4AWYB1_EeXBOgqda Fwfkc8oS87LYqNnCZzH1wknL.lv2iGd4awuDnqCckdGuZ7QMD_9S8Nw8v900 U8uA3T7qlZqGmn1Fbb2X3FPQprGTNN7PMuzv40XALrRt8XMj034fIbqvGypD QTa7b_8uWacS4VG0U.zvbVBbii_099LMndtfbJeez.60PHhRY87fDiHJJnkr 8xkIVe9kl8TyKiA-- X-Yahoo-SMTP: k2gD1GeswBAV_JFpZm8dmpTCwr4ufTKOyA-- X-Rocket-Received: from sattelite (davidparks21@113.161.75.108 with login) by smtp127.mail.ird.yahoo.com with SMTP; 19 Apr 2013 00:02:30 -0700 PDT From: "David Parks" To: References: In-Reply-To: Subject: RE: Uploading file to HDFS Date: Fri, 19 Apr 2013 14:02:15 +0700 Message-ID: <118701ce3ccb$d5a1c720$80e55560$@yahoo.com> MIME-Version: 1.0 Content-Type: text/plain; charset="GB2312" Content-Transfer-Encoding: quoted-printable X-Mailer: Microsoft Outlook 14.0 Thread-Index: AQExkd0iOiq3R7BZItIyRlktUC87pwKlyGOFmgFDFLA= Content-Language: en-us X-Virus-Checked: Checked by ClamAV on apache.org I think the problem here is that he doesn't have Hadoop installed on = this other location so there's no Hadoop DFS client to do the put directly = into HDFS on, he would normally copy the file to one of the nodes in the = cluster where the client files are installed. I've had the same problem = recently. I've tried setting up dfs-hdfs-proxy, though I must say that it's been crashing when I try to put modest to large files through it (but I've = got a thread going with the developer on that issue). That, or one of the = other remote mount options might work well.=20 https://www.google.com/url?sa=3Dt&rct=3Dj&q=3D&esrc=3Ds&source=3Dweb&cd=3D= 1&cad=3Drja&ved=3D 0CDQQFjAA&url=3Dhttp%3A%2F%2Fwiki.apache.org%2Fhadoop%2FMountableHDFS&ei=3D= T-pwU Y74A8jPrQfYooHoBw&usg=3DAFQjCNEQbxmrMGKAETj3FPEw3Lr1PBHz-w&sig2=3D4JpEzZ_= 8IAyJ-N PofSRmMg&bvm=3Dbv.45373924,d.bmk You could also install Hadoop on the box that has the 2TB file (I = realize that you might not control it or want to do that depending on the configuration). A remote NFS mount that you can access from one of the Hadoop boxes... ? Split up the file into smaller pieces? There are some ideas. I'd love to hear your final solution as I've also = been having fits getting into HDFS from outside the Hadoop environment. I = wish it natively supported NFS mounts or some light weight/easy to install = remote DFS tools. Dave -----Original Message----- From: Harsh J [mailto:harsh@cloudera.com]=20 Sent: Friday, April 19, 2013 1:40 PM To: Subject: Re: Uploading file to HDFS Can you not simply do a fs -put from the location where the 2 TB file currently resides? HDFS should be able to consume it just fine, as the client chunks them into fixed size blocks. On Fri, Apr 19, 2013 at 10:05 AM, =B3=AC=BC=B6=C8=FB=D1=C7=C8=CB = wrote: > I have a problem. Our cluster has 32 nodes. Each disk is 1TB. I wanna=20 > upload 2TB file to HDFS.How can I put the file to the namenode and = upload to HDFS? -- Harsh J