Return-Path: X-Original-To: apmail-hadoop-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 446E79DBB for ; Wed, 6 Mar 2013 16:21:07 +0000 (UTC) Received: (qmail 38488 invoked by uid 500); 6 Mar 2013 16:21:01 -0000 Delivered-To: apmail-hadoop-user-archive@hadoop.apache.org Received: (qmail 38394 invoked by uid 500); 6 Mar 2013 16:21:01 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 38376 invoked by uid 99); 6 Mar 2013 16:21:00 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 06 Mar 2013 16:21:00 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of gsmsteve@gmail.com designates 209.85.223.169 as permitted sender) Received: from [209.85.223.169] (HELO mail-ie0-f169.google.com) (209.85.223.169) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 06 Mar 2013 16:20:53 +0000 Received: by mail-ie0-f169.google.com with SMTP id 13so9796606iea.28 for ; Wed, 06 Mar 2013 08:20:32 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=x-received:mime-version:in-reply-to:references:from:date:message-id :subject:to:content-type; bh=q9zhdBdFThwzZZ+06KqImBFCMhbDghdE326EAUvKzRA=; b=RTaCP12UKbb6o63v8WSaXAIf5FtERuugdjeIi84mkqpKl/fgNrrnRjXQrrx2j+2gRQ lctuiYAebhnVuDPXslEoNCgtN4naF3nec6FQ3HeF2/FebZEhk03vvkOYJowKQuUz/D9S 0hkJW2+ysIFW4B6ITIlIkvJXL4kdkT2HukW/QC2tUCpoDGZcP8NdBFF2yirGBWKMsDXB wKSFKBK7KVrPXcxQCVLG+HxM9yAvQ9T11ST0ck4SVQYv6vhHCwINWZvtQ9yrHg4hAuV6 LfGoU50hAW+eL4uNGCCyO+kdcERhPnuJNLSpu6DwYWFQ3oVZyVFrHmK5WvktAISmwSKk mmRw== X-Received: by 10.50.170.69 with SMTP id ak5mr11464725igc.56.1362586830562; Wed, 06 Mar 2013 08:20:30 -0800 (PST) MIME-Version: 1.0 Received: by 10.64.50.202 with HTTP; Wed, 6 Mar 2013 08:20:10 -0800 (PST) In-Reply-To: References: From: Shumin Guo Date: Wed, 6 Mar 2013 11:20:10 -0500 Message-ID: Subject: Re: S3N copy creating recursive folders To: user@hadoop.apache.org Content-Type: multipart/alternative; boundary=e89a8f2343350342e704d743f898 X-Virus-Checked: Checked by ClamAV on apache.org --e89a8f2343350342e704d743f898 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: quoted-printable I used to have similar problem. Looks like there is a recursive folder creation bug. How about you try remove the srcData from the , for example use the following command: *hadoop fs -cp s3n://acessKey:acessSecret@bucket.name/srcData" /test/* Or with distcp: *hadoop distcp s3n://acessKey:acessSecret@bucket.name/srcData" /test/* HTH. Shumin On Wed, Mar 6, 2013 at 5:44 AM, Subroto wrote: > Hi Mike, > > I have tries distcp as well and it ended up with exception: > 13/03/06 05:41:13 INFO tools.DistCp: srcPaths=3D[ > s3n://acessKey:acessSecret@dm.test.bucket/srcData] > 13/03/06 05:41:13 INFO tools.DistCp: destPath=3D/test/srcData > 13/03/06 05:41:18 INFO tools.DistCp: /test/srcData does not exist. > org.apache.hadoop.tools.DistCp$DuplicationException: Invalid input, there > are duplicated files in the sources: > s3n://acessKey:acessSecret@dm.test.bucket/srcData/compressed, > s3n://acessKey:acessSecret@dm.test.bucket/srcData/compressed > at org.apache.hadoop.tools.DistCp.checkDuplication(DistCp.java:1368) > at org.apache.hadoop.tools.DistCp.setup(DistCp.java:1176) > at org.apache.hadoop.tools.DistCp.copy(DistCp.java:666) > at org.apache.hadoop.tools.DistCp.run(DistCp.java:881) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79) > at org.apache.hadoop.tools.DistCp.main(DistCp.java:908) > > One more interesting stuff to notice is that same thing works nicely with > hadoop 2.0 > > Cheers, > Subroto Sanyal > > On Mar 6, 2013, at 11:12 AM, Michel Segel wrote: > > Have you tried using distcp? > > Sent from a remote device. Please excuse any typos... > > Mike Segel > > On Mar 5, 2013, at 8:37 AM, Subroto wrote: > > Hi, > > Its not because there are too many recursive folders in S3 bucket; in-fac= t > there is no recursive folder in the source. > If I list the S3 bucket with Native S3 tools I can find a file srcData > with size 0 in the folder srcData. > The copy command keeps on creating folder /test/srcData/srcData/srcData > (keep on appending srcData). > > Cheers, > Subroto Sanyal > > On Mar 5, 2013, at 3:32 PM, =E5=8D=96=E6=8A=A5=E7=9A=84=E5=B0=8F=E8=A1=8C= =E5=AE=B6 wrote: > > Hi Subroto, > > I didn't use the s3n filesystem.But from the output "cp: > java.io.IOException: mkdirs: Pathname too long. Limit 8000 characters, > 1000 levels.", I think this is because the problem of the path. Is the > path longer than 8000 characters or the level is more than 1000? > You only have 998 folders.Maybe the last one is more than 8000 > characters.Why not count the last one's length? > > BRs//Julian > > > > > > ------------------ Original ------------------ > *From: * "Subroto"; > *Date: * Tue, Mar 5, 2013 10:22 PM > *To: * "user"; ** > *Subject: * S3N copy creating recursive folders > > Hi, > > I am using Hadoop 1.0.3 and trying to execute: > hadoop fs -cp s3n://acessKey:acessSecret@bucket.name/srcData" > /test/srcData > > This ends up with: > cp: java.io.IOException: mkdirs: Pathname too long. Limit 8000 > characters, 1000 levels. > > When I try to list the folder recursively /test/srcData: it lists 998 > folders like: > drwxr-xr-x - root supergroup 0 2013-03-05 08:49 > /test/srcData/srcData > drwxr-xr-x - root supergroup 0 2013-03-05 08:49 > /test/srcData/srcData/srcData > drwxr-xr-x - root supergroup 0 2013-03-05 08:49 > /test/srcData/srcData/srcData/srcData > drwxr-xr-x - root supergroup 0 2013-03-05 08:49 > /test/srcData/srcData/srcData/srcData/srcData > drwxr-xr-x - root supergroup 0 2013-03-05 08:49 > /test/srcData/srcData/srcData/srcData/srcData/srcData > > Is there a problem with s3n filesystem ?? > > Cheers, > Subroto Sanyal > > > > --e89a8f2343350342e704d743f898 Content-Type: text/html; charset=UTF-8 Content-Transfer-Encoding: quoted-printable I used to have similar problem. Looks like there is a recursive folder crea= tion bug. How about you try remove the srcData from the <dst>, for ex= ample use the following command:=C2=A0

hadoop fs -cp s3n://acessKey:acessSecret@bucket.= name/srcData&quo= t; /test/

Or w= ith distcp:=C2=A0

hadoop distcp=C2=A0s3n://acessKey:acessSecret@bucket.name/srcData" /test/

HTH.=C2=A0
=
Shumin

On Wed, Mar 6, 2013 at 5:44 AM, Subroto <ssanyal@datameer.com> wrote:
Hi Mike,

I have trie= s distcp as well and it ended up with exception:
13/03/06 05= :41:13 INFO tools.DistCp: srcPaths=3D[s3n://acessKey:acessSecret@dm.test= .bucket/srcData]
13/03/06 05:41:13 INFO tools.DistCp: destPath=3D/test/srcData
13/03/06 05:41:18 INFO tools.DistCp: /test/srcData does not exist.
<= div>org.apache.hadoop.tools.DistCp$DuplicationException: Invalid input, the= re are duplicated files in the sources: s3n://acessKey:acessSecret@dm.te= st.bucket/srcData/compressed, s3n://acessKey:acessSecret@dm.test.buc= ket/srcData/compressed
at org.apache.hadoop.tool= s.DistCp.checkDuplication(DistCp.java:1368)
at org.apache.hadoop.tools.DistCp.setup(DistCp.java= :1176)
at org.apache.hadoop.tool= s.DistCp.copy(DistCp.java:666)
at org.apache.hadoop.tools.DistCp.run(DistCp.java:881)
at org.apache.hadoop.util.Tool= Runner.run(ToolRunner.java:65)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79)
at org.apache.hadoop.tool= s.DistCp.main(DistCp.java:908)

One more interestin= g stuff to notice is that same thing works nicely with hadoop 2.0

Cheers,
Subroto Sanyal

On Mar 6, 2013, at 11:12 AM, Michel Segel wrot= e:

Have you tried= using distcp?

Sent from a remote device. Please excuse any typos...

Mike Segel

On Mar 5, 2013, at 8:37 AM, Subroto <ssanyal@datameer.co= m> wrote:

Hi,

Its not bec= ause there are too many recursive folders in S3 bucket; in-fact there is no= recursive folder in the source.
If I list the S3 bucket with Nat= ive S3 tools I can find a file srcData with size 0 in the folder srcData.= =C2=A0
The copy command keeps on creating folder=C2=A0=C2=A0/test/srcData/src= Data/srcData (keep on appending srcData).

Cheers,<= /div>
Subroto Sanyal

On Mar 5, 2013, at 3:32 PM= , =E5=8D=96=E6=8A=A5=E7=9A=84=E5=B0=8F=E8=A1=8C=E5=AE=B6 wrote:

Hi Subroto,

I didn't use the s3n filesystem.But =C2=A0from the output "cp: java.io.IOException: mkdirs: Pathname too lo= ng.=C2=A0 Limit 8000 characters, 1000 levels.", I think this is because the problem of the path. Is the pa= th longer than 8000 characters or the level is more than 1000?
You only have 998 folders.Maybe the la= st one is more than 8000 characters.Why not count the last one's length= ?

BRs//Julian



<= /span>


=
= ------------------=C2=A0Original=C2=A0------------------
From: =C2=A0&qu= ot;Subroto"<ssanyal@datameer.com>;
Date: =C2=A0Tue, Mar 5, 2013 10:22 PM
To: =C2= =A0"user"<user@hadoop.apache.org>;
Sub= ject: =C2=A0S3N copy creating recursive folders

Hi,

I am using Hadoop 1.0.3 and trying to execu= te:
hadoop fs -cp s3n://acessKey:acessSecret@bucket.name/srcData&= quot; /test/srcData

This ends up with:
cp: java.io.IOException: m= kdirs: Pathname too long.=C2=A0 Limit 8000 characters, 1000 levels.

When I try to list the folder recursively /test/srcData: it lists 998 f= olders like:
drwxr-xr-x=C2=A0=C2=A0 - root supergroup=C2=A0=C2=A0=C2=A0= =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 0 2013-03-05 08:49 /test/srcData/srcDa= ta
drwxr-xr-x=C2=A0=C2=A0 - root supergroup=C2=A0=C2=A0=C2=A0=C2=A0=C2= =A0=C2=A0=C2=A0=C2=A0=C2=A0 0 2013-03-05 08:49 /test/srcData/srcData/srcDat= a
drwxr-xr-x=C2=A0=C2=A0 - root supergroup=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2= =A0=C2=A0=C2=A0=C2=A0 0 2013-03-05 08:49 /test/srcData/srcData/srcData/srcD= ata
drwxr-xr-x=C2=A0=C2=A0 - root supergroup=C2=A0=C2=A0=C2=A0=C2=A0=C2= =A0=C2=A0=C2=A0=C2=A0=C2=A0 0 2013-03-05 08:49 /test/srcData/srcData/srcDat= a/srcData/srcData
drwxr-xr-x=C2=A0=C2=A0 - root supergroup=C2=A0=C2=A0= =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 0 2013-03-05 08:49 /test/srcData= /srcData/srcData/srcData/srcData/srcData

Is there a problem with s3n filesystem ??

Cheers,
Subroto San= yal



--e89a8f2343350342e704d743f898--