Return-Path: X-Original-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 1CE59109C2 for ; Mon, 2 Dec 2013 23:32:54 +0000 (UTC) Received: (qmail 96419 invoked by uid 500); 2 Dec 2013 23:32:49 -0000 Delivered-To: apmail-hadoop-hdfs-user-archive@hadoop.apache.org Received: (qmail 96301 invoked by uid 500); 2 Dec 2013 23:32:49 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 96292 invoked by uid 99); 2 Dec 2013 23:32:48 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 02 Dec 2013 23:32:48 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of acm@hortonworks.com designates 209.85.160.43 as permitted sender) Received: from [209.85.160.43] (HELO mail-pb0-f43.google.com) (209.85.160.43) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 02 Dec 2013 23:32:42 +0000 Received: by mail-pb0-f43.google.com with SMTP id rq2so20132965pbb.30 for ; Mon, 02 Dec 2013 15:32:21 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:from:message-id:mime-version:subject:date :references:to:in-reply-to:content-type; bh=YXPpQeAnB4110VaUJaQAq89HY9Q2ZhEos9KSiLr8MyE=; b=WD9OXUg2UCZqmldjpSFJBfaKu/fiRNFquTvlYuh0pV/zkk8ZPG6/T6ZZrCoxJ1qW1j UfhRi4AYr7GVF4TusW2zokCPTBESK1zQhkgR6UzsD+uEmoG5Ef14oDvVcPkBh3zzrB/G OYYBI44bOwrNeklPtNS/l8R9FcokL9YGomamxqecXRcEnoKkdLrtE1xYANlPhrgL93hA GRjBKnUk7VZ0OuuJ7YGZ30DGEnXLly4HPxUawVeXZEvRWTgxiE7CIbKyZE0V1Zqv4UQi WyVOzndiijLsnKq7Tb1/v5UU5Er06p1MQMa1KLmEbqdGV8SJgzYbvfu75V8MlFVE5SHx EQ7g== X-Gm-Message-State: ALoCoQnpUsOysJZIwlWhxwbLhX677YQQQCJAQsOtVD23JeZdnVgQpzN5rMg0VdydZ4H9jSOrR+jaoLp/JbUi/ulFHHkqj0P8wGzNe8OtQ6Vrg7B5xBgrWJw= X-Received: by 10.68.139.100 with SMTP id qx4mr6948827pbb.144.1386027141150; Mon, 02 Dec 2013 15:32:21 -0800 (PST) Received: from [10.11.3.119] ([192.175.27.2]) by mx.google.com with ESMTPSA id jk16sm15101643pbb.34.2013.12.02.15.32.14 for (version=TLSv1 cipher=ECDHE-RSA-RC4-SHA bits=128/128); Mon, 02 Dec 2013 15:32:15 -0800 (PST) From: Arun C Murthy Message-Id: <3F7FF87F-2894-403E-BF1C-837FECBA2875@hortonworks.com> Mime-Version: 1.0 (Mac OS X Mail 6.6 \(1510\)) Subject: Re: Hadoop 2.2.0 from source configuration Date: Mon, 2 Dec 2013 15:32:12 -0800 References: To: user@hadoop.apache.org In-Reply-To: X-Mailer: Apple Mail (2.1510) Content-Type: multipart/alternative; boundary="Apple-Mail=_B6DC9E5E-8E6A-4D2A-900C-B4B54AAF71A9" X-Virus-Checked: Checked by ClamAV on apache.org --Apple-Mail=_B6DC9E5E-8E6A-4D2A-900C-B4B54AAF71A9 Content-Transfer-Encoding: quoted-printable Content-Type: text/plain; charset=ISO-8859-1 Daniel, Apologies if you had a bad experience. If you can point them out to us, we= 'd be more than happy to fix it - alternately, we'd *love* it if you could = help us improve docs too. Now, for the problem at hand: http://wiki.apache.org/hadoop/CouldOnlyBeRep= licatedTo is one place to look. Basically NN cannot find any datanodes. Any= thing in your NN logs to indicate trouble? Also, pls feel free to open liras with issues you find and we'll help. thanks, Arun On Dec 2, 2013, at 8:44 AM, Daniel Savard wrote: > Andr=E9, >=20 > good for you that greedy instructions on the reference page were enough t= o setup your cluster. However, read them again and see how many assumptions= are made into them about what you are supposed to already know and should = come without saying more about it. >=20 > I did try the single node setup, it is worst than the cluster setup regar= ding the instructions. You are supposed to already have a near working syst= em as far as I understand the instructions. It is assumed the HDFS is alrea= dy setup and working properly. Try to find the instructions to setup HDFS f= or version 2.2.0 and you will end up with a lot of inappropriate instructio= ns about previous version (some properties were renamed). >=20 > It may appear hard at people to say this is toxic, but it is. The first p= lace a newcomer will go is setup a single node. This will be his starting p= oint and he will be left with a bunch of a priori and no clue. >=20 > To go back to my very problem at this point:=20 >=20 > 13/12/02 11:34:07 WARN hdfs.DFSClient: DataStreamer Exception > org.apache.hadoop.ipc.RemoteException(java.io.IOException): File /test._C= OPYING_ could only be replicated to 0 nodes instead of minReplication (=3D1= ). There are 1 datanode(s) running and no node(s) are excluded in this ope= ration. > at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseT= arget(BlockManager.java:1384) > at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalB= lock(FSNamesystem.java:2477) > at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(= NameNodeRpcServer.java:555) > at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSide= TranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:387= ) > at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos= $ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.j= ava:59582) > at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.= call(ProtobufRpcEngine.java:585) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2048) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2044) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInfo= rmation.java:1491) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2042) >=20 > at org.apache.hadoop.ipc.Client.call(Client.java:1347) > at org.apache.hadoop.ipc.Client.call(Client.java:1300) > at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpc= Engine.java:206) > at com.sun.proxy.$Proxy9.addBlock(Unknown Source) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorIm= pl.java:57) > at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAc= cessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(Ret= ryInvocationHandler.java:186) > at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvo= cationHandler.java:102) > at com.sun.proxy.$Proxy9.addBlock(Unknown Source) > at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslator= PB.addBlock(ClientNamenodeProtocolTranslatorPB.java:330) > at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowin= gBlock(DFSOutputStream.java:1226) > at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutpu= tStream(DFSOutputStream.java:1078) > at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputS= tream.java:514) >=20 > I can copy an empty file, but as soon as its content is non-zero I am get= ting this message. Searching on the message is of no help so far. >=20 > And I skimmed through the cluster instructions and found nothing there th= at could help in any way neither. >=20 >=20 > ----------------- > Daniel Savard >=20 >=20 > 2013/12/2 Andre Kelpe > Hi Daniel, >=20 > first of all, before posting to a mailing list, take a deep breath and > let your frustrations out. Then write the email. Using words like > "crappy", "toxicware", "nightmare" are not going to help you getting > useful responses. >=20 > While I agree that the docs can be confusing, we should try to stay > constructive. You haven't mentioned which documentation you are > using. I found the cluster tutorial sufficient to get me started: > http://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-common/Cl= usterSetup.html >=20 > If you are looking for an easy way to spin up a small cluster with > hadoop 2.2, try the hadoop2 branch of this vagrant setup: >=20 > https://github.com/fs111/vagrant-hadoop-cluster/tree/hadoop2 >=20 > - Andr=E9 >=20 > On Mon, Dec 2, 2013 at 5:34 AM, Daniel Savard w= rote: > > I am trying to configure hadoop 2.2.0 from source code and I found the > > instructions really crappy and incomplete. It is like they were written= to > > avoid someone can do the job himself and must contract someone else to = do it > > or buy a packaged version. > > > > It is about three days I am struggling with this stuff with partial suc= cess. > > The documentation is less than clear and most of the stuff out there ap= ply > > to earlier version and they haven't been updated for version 2.2.0. > > > > I was able to setup HDFS, however I am still unable to use it. I am doi= ng a > > single node installation and the instruction page doesn't explain anyth= ing > > beside telling you to do this and that without documenting what each th= ing > > is doing and what choices are available and what guidelines you should > > follow. There is even environment variables you are told to set, but no= thing > > is said about what they mean and to which value they should be set. It = seems > > it assumes prior knowledge of everything about hadoop. > > > > Anyone knows a site with proper documentation about hadoop or it's hope= less > > and this whole thing is just a piece of toxicware? > > > > I am already looking for alternate solutions to hadoop which for sure w= ill > > be a nightmare to manage and install each time a new version, release w= ill > > become available. > > > > TIA > > ----------------- > > Daniel Savard >=20 >=20 >=20 > -- > Andr=E9 Kelpe > andre@concurrentinc.com > http://concurrentinc.com >=20 -- Arun C. Murthy Hortonworks Inc. http://hortonworks.com/ --=20 CONFIDENTIALITY NOTICE NOTICE: This message is intended for the use of the individual or entity to= =20 which it is addressed and may contain information that is confidential,=20 privileged and exempt from disclosure under applicable law. If the reader= =20 of this message is not the intended recipient, you are hereby notified that= =20 any printing, copying, dissemination, distribution, disclosure or=20 forwarding of this communication is strictly prohibited. If you have=20 received this communication in error, please contact the sender immediately= =20 and delete it from your system. Thank You. --Apple-Mail=_B6DC9E5E-8E6A-4D2A-900C-B4B54AAF71A9 Content-Transfer-Encoding: quoted-printable Content-Type: text/html; charset=ISO-8859-1 Daniel,

<= div> Apologies if you had a bad experience. If you can point them out = to us, we'd be more than happy to fix it - alternately, we'd *love* it if y= ou could help us improve docs too.

 Now, for = the problem at hand: http://wiki.apache.org/hadoop/CouldOnlyBeReplicatedTo&= nbsp;is one place to look. Basically NN cannot find any datanodes. Anything= in your NN logs to indicate trouble?

 Also, = pls feel free to open liras with issues you find and we'll help.
=
thanks,
Arun

On Dec 2, 201= 3, at 8:44 AM, Daniel Savard <daniel.savard@gmail.com> wrote:

Andr=E9,

good for you that greedy instructions on the = reference page were enough to setup your cluster. However, read them again = and see how many assumptions are made into them about what you are supposed= to already know and should come without saying more about it.

I did try the single node setup, it is worst than the cluster set= up regarding the instructions. You are supposed to already have a near work= ing system as far as I understand the instructions. It is assumed the HDFS = is already setup and working properly. Try to find the instructions to setu= p HDFS for version 2.2.0 and you will end up with a lot of inappropriate in= structions about previous version (some properties were renamed).

It may appear hard at people to say this is toxic, but it is. The= first place a newcomer will go is setup a single node. This will be his st= arting point and he will be left with a bunch of a priori and no clue.

To go back to my very problem at this point:

13/12/02 11:= 34:07 WARN hdfs.DFSClient: DataStreamer Exception
org.apache.hadoop.ipc.= RemoteException(java.io.IOException): File /test._COPYING_ could only be re= plicated to 0 nodes instead of minReplication (=3D1).  There are 1 dat= anode(s) running and no node(s) are excluded in this operation.
    at org.apache.hadoop.hdfs.server.blockmanagement.BlockMa= nager.chooseTarget(BlockManager.java:1384)
    at org.apa= che.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesyste= m.java:2477)
    at org.apache.hadoop.hdfs.server.namenod= e.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:555)
    at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProto= colServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslat= orPB.java:387)
    at org.apache.hadoop.hdfs.protocol.pro= to.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod= (ClientNamenodeProtocolProtos.java:59582)
    at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoB= ufRpcInvoker.call(ProtobufRpcEngine.java:585)
    at org.= apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
    at or= g.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2048)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.jav= a:2044)
    at java.security.AccessController.doPrivilege= d(Native Method)
    at javax.security.auth.Subject.doAs(= Subject.java:415)
    at org.apache.hadoop.security.UserG= roupInformation.doAs(UserGroupInformation.java:1491)
    at org.apache.hadoop.ipc.Server$Handler.run(Server.java:= 2042)

    at org.apache.hadoop.ipc.Client.call(Client= .java:1347)
    at org.apache.hadoop.ipc.Client.call(Clie= nt.java:1300)
    at org.apache.hadoop.ipc.ProtobufRpcEng= ine$Invoker.invoke(ProtobufRpcEngine.java:206)
    at com.sun.proxy.$Proxy9.addBlock(Unknown Source)
&nb= sp;   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Meth= od)
    at sun.reflect.NativeMethodAccessorImpl.invoke(Na= tiveMethodAccessorImpl.java:57)
    at sun.reflect.Delega= tingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:606)
&= nbsp;   at org.apache.hadoop.io.retry.RetryInvocationHandler.invo= keMethod(RetryInvocationHandler.java:186)
    at org.apac= he.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.jav= a:102)
    at com.sun.proxy.$Proxy9.addBlock(Unknown Source)
&nb= sp;   at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocol= TranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:330)
 = ;   at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locate= FollowingBlock(DFSOutputStream.java:1226)
    at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.n= extBlockOutputStream(DFSOutputStream.java:1078)
    at or= g.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:= 514)

I can copy an empty file, but as soon as its content is n= on-zero I am getting this message. Searching on the message is of no help s= o far.

And I skimmed through the cluster instructions and found nothing = there that could help in any way neither.


-----------------
Daniel Savard


2013/12/2 Andre Kelpe = <akelpe@co= ncurrentinc.com>
Hi Daniel,

first of all, before posting to a mailing list, take a deep breath and
let your frustrations out. Then write the email. Using words like
"crappy", "toxicware", "nightmare" are not going to help you getting
useful responses.

While I agree that the docs can be confusing, we should try to stay
constructive. You haven't  mentioned which documentation you are
using. I found the cluster tutorial sufficient to get me started:
http://hadoop.apache.org/docs/s= table/hadoop-project-dist/hadoop-common/ClusterSetup.html

If you are looking for an easy way to spin up a small cluster with
hadoop 2.2, try the hadoop2 branch of this vagrant setup:

https://github.com/fs111/vagrant-hadoop-cluster/tree/hadoop= 2

- Andr=E9

On Mon, Dec 2, 2013 at 5:34 AM, Daniel Savard <daniel.savard@gmail.com> wrote:
> I am trying to configure hadoop 2.2.0 from source code and I found the=
> instructions really crappy and incomplete. It is like they were writte= n to
> avoid someone can do the job himself and must contract someone else to= do it
> or buy a packaged version.
>
> It is about three days I am struggling with this stuff with partial su= ccess.
> The documentation is less than clear and most of the stuff out there a= pply
> to earlier version and they haven't been updated for version 2.2.0. >
> I was able to setup HDFS, however I am still unable to use it. I am do= ing a
> single node installation and the instruction page doesn't explain anyt= hing
> beside telling you to do this and that without documenting what each t= hing
> is doing and what choices are available and what guidelines you should=
> follow. There is even environment variables you are told to set, but n= othing
> is said about what they mean and to which value they should be set. It= seems
> it assumes prior knowledge of everything about hadoop.
>
> Anyone knows a site with proper documentation about hadoop or it's hop= eless
> and this whole thing is just a piece of toxicware?
>
> I am already looking for alternate solutions to hadoop which for sure = will
> be a nightmare to manage and install each time a new version, release = will
> become available.
>
> TIA
> -----------------
> Daniel Savard



--
Andr=E9 Kelpe
andre@concurrentinc.com
http://concurrentin= c.com


--
Arun C. Murthy
Ho= rtonworks Inc.
http://hortonworks.co= m/



CONFIDENTIALITY NOTICE
NOTICE: This message is = intended for the use of the individual or entity to which it is addressed a= nd may contain information that is confidential, privileged and exempt from= disclosure under applicable law. If the reader of this message is not the = intended recipient, you are hereby notified that any printing, copying, dis= semination, distribution, disclosure or forwarding of this communication is= strictly prohibited. If you have received this communication in error, ple= ase contact the sender immediately and delete it from your system. Thank Yo= u. --Apple-Mail=_B6DC9E5E-8E6A-4D2A-900C-B4B54AAF71A9--