hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Xu, Richard " <richard...@citi.com>
Subject RE: Unable to start hadoop-0.20.2 but able to start hadoop-0.20.203 cluster
Date Fri, 27 May 2011 21:32:23 GMT
Add more to that:

I also tried start 0.20.2 on a linux machine in distributed mode, same error.

I had successfully started 0.20.203 on this linux machine with same config.

Seems that it is not related to Solaris.

Could it caused by port? I checked a few, did not find anyone blocked.



-----Original Message-----
From: Xu, Richard [ICG-IT]
Sent: Friday, May 27, 2011 4:18 PM
To: 'Allen Wittenauer'; 'common-user@hadoop.apache.org'
Subject: RE: Unable to start hadoop-0.20.2 but able to start hadoop-0.20.203 cluster

Hi Allen,

Thanks a lot for your response.

I agree with you that it does not matter with replication settings.

What really bothered me is same environment, same configures, hadoop 0.20.203 takes us 3 mins,
why 0.20.2 took 3 days.

Can you pls. shed more light on how "to make Hadoop's broken username detection to work properly"?

-----Original Message-----
From: Allen Wittenauer [mailto:aw@apache.org]
Sent: Friday, May 27, 2011 11:42 AM
To: common-user@hadoop.apache.org
Cc: Xu, Richard [ICG-IT]
Subject: Re: Unable to start hadoop-0.20.2 but able to start hadoop-0.20.203 cluster


On May 27, 2011, at 7:26 AM, DAN wrote:
> You see you have "2 Solaris servers for now", and dfs.replication is setted as 3.
> These don't match.


        That doesn't matter.  HDFS will basically flag any files written with a warning that
they are under-replicated.

        The problem is that the datanode processes aren't running and/or aren't communicating
to the namenode. That's what the "java.io.IOException: File /tmp/hadoop-cfadm/mapred/system/jobtracker.info
could only be replicated to 0 nodes, instead of 1" means.

        It should also be pointed out that writing to /tmp (the default) is a bad idea.  This
should get changed.

        Also, since you are running Solaris, check the FAQ on some settings you'll need to
do in order to make Hadoop's broken username detection to work properly, amongst other things.

Mime
View raw message