hadoop-mapreduce-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Daniel Savard <daniel.sav...@gmail.com>
Subject Re: Hadoop 2.2.0 from source configuration
Date Mon, 02 Dec 2013 17:10:55 GMT
Here is additional information about the HDFS:

$ hdfs dfsadmin -report
Configured Capacity: 3208335360 (2.99 GB)
Present Capacity: 534454272 (509.70 MB)
DFS Remaining: 534450176 (509.69 MB)
DFS Used: 4096 (4 KB)
DFS Used%: 0.00%
Under replicated blocks: 0
Blocks with corrupt replicas: 0
Missing blocks: 0

-------------------------------------------------
Datanodes available: 1 (1 total, 0 dead)

Live datanodes:
Name: 127.0.0.1:50010 (hosta.subdom1.tld1)
Hostname: hosta.subdom1.tld1
Decommission Status : Normal
Configured Capacity: 3208335360 (2.99 GB)
DFS Used: 4096 (4 KB)
Non DFS Used: 2673881088 (2.49 GB)
DFS Remaining: 534450176 (509.69 MB)
DFS Used%: 0.00%
DFS Remaining%: 16.66%
Last contact: Mon Dec 02 12:07:28 EST 2013


I see nothing that could explain the error. I can mkdir, put empty files,
list content.

-----------------
Daniel Savard


2013/12/2 Daniel Savard <daniel.savard@gmail.com>

> André,
>
> good for you that greedy instructions on the reference page were enough to
> setup your cluster. However, read them again and see how many assumptions
> are made into them about what you are supposed to already know and should
> come without saying more about it.
>
> I did try the single node setup, it is worst than the cluster setup
> regarding the instructions. You are supposed to already have a near working
> system as far as I understand the instructions. It is assumed the HDFS is
> already setup and working properly. Try to find the instructions to setup
> HDFS for version 2.2.0 and you will end up with a lot of inappropriate
> instructions about previous version (some properties were renamed).
>
> It may appear hard at people to say this is toxic, but it is. The first
> place a newcomer will go is setup a single node. This will be his starting
> point and he will be left with a bunch of a priori and no clue.
>
> To go back to my very problem at this point:
>
> 13/12/02 11:34:07 WARN hdfs.DFSClient: DataStreamer Exception
> org.apache.hadoop.ipc.RemoteException(java.io.IOException): File
> /test._COPYING_ could only be replicated to 0 nodes instead of
> minReplication (=1).  There are 1 datanode(s) running and no node(s) are
> excluded in this operation.
>     at
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1384)
>     at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2477)
>     at
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:555)
>     at
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:387)
>     at
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:59582)
>     at
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2048)
>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2044)
>     at java.security.AccessController.doPrivileged(Native Method)
>     at javax.security.auth.Subject.doAs(Subject.java:415)
>     at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2042)
>
>     at org.apache.hadoop.ipc.Client.call(Client.java:1347)
>     at org.apache.hadoop.ipc.Client.call(Client.java:1300)
>     at
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
>     at com.sun.proxy.$Proxy9.addBlock(Unknown Source)
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>     at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>     at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>     at java.lang.reflect.Method.invoke(Method.java:606)
>     at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:186)
>     at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
>     at com.sun.proxy.$Proxy9.addBlock(Unknown Source)
>     at
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:330)
>     at
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1226)
>     at
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1078)
>     at
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:514)
>
> I can copy an empty file, but as soon as its content is non-zero I am
> getting this message. Searching on the message is of no help so far.
>
> And I skimmed through the cluster instructions and found nothing there
> that could help in any way neither.
>
>
> -----------------
> Daniel Savard
>
>
> 2013/12/2 Andre Kelpe <akelpe@concurrentinc.com>
>
>> Hi Daniel,
>>
>> first of all, before posting to a mailing list, take a deep breath and
>> let your frustrations out. Then write the email. Using words like
>> "crappy", "toxicware", "nightmare" are not going to help you getting
>> useful responses.
>>
>> While I agree that the docs can be confusing, we should try to stay
>> constructive. You haven't  mentioned which documentation you are
>> using. I found the cluster tutorial sufficient to get me started:
>>
>> http://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-common/ClusterSetup.html
>>
>> If you are looking for an easy way to spin up a small cluster with
>> hadoop 2.2, try the hadoop2 branch of this vagrant setup:
>>
>> https://github.com/fs111/vagrant-hadoop-cluster/tree/hadoop2
>>
>> - André
>>
>> On Mon, Dec 2, 2013 at 5:34 AM, Daniel Savard <daniel.savard@gmail.com>
>> wrote:
>> > I am trying to configure hadoop 2.2.0 from source code and I found the
>> > instructions really crappy and incomplete. It is like they were written
>> to
>> > avoid someone can do the job himself and must contract someone else to
>> do it
>> > or buy a packaged version.
>> >
>> > It is about three days I am struggling with this stuff with partial
>> success.
>> > The documentation is less than clear and most of the stuff out there
>> apply
>> > to earlier version and they haven't been updated for version 2.2.0.
>> >
>> > I was able to setup HDFS, however I am still unable to use it. I am
>> doing a
>> > single node installation and the instruction page doesn't explain
>> anything
>> > beside telling you to do this and that without documenting what each
>> thing
>> > is doing and what choices are available and what guidelines you should
>> > follow. There is even environment variables you are told to set, but
>> nothing
>> > is said about what they mean and to which value they should be set. It
>> seems
>> > it assumes prior knowledge of everything about hadoop.
>> >
>> > Anyone knows a site with proper documentation about hadoop or it's
>> hopeless
>> > and this whole thing is just a piece of toxicware?
>> >
>> > I am already looking for alternate solutions to hadoop which for sure
>> will
>> > be a nightmare to manage and install each time a new version, release
>> will
>> > become available.
>> >
>> > TIA
>> > -----------------
>> > Daniel Savard
>>
>>
>>
>> --
>> André Kelpe
>> andre@concurrentinc.com
>> http://concurrentinc.com
>>
>
>

Mime
View raw message