hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Bing Li <lbl...@gmail.com>
Subject Re: ERROR: org.apache.hadoop.hbase.MasterNotRunningException: Retried 7 Times
Date Tue, 14 Feb 2012 08:32:58 GMT
Dear Jimmy,

After changing my Linux from RedHat 9 to Ubuntu 10, all the problems are
solved.

Maybe RedHat 9 is too old to fit HBase and Hadoop?

Thanks,
Bing

On Tue, Feb 14, 2012 at 5:23 AM, Bing Li <lblabs@gmail.com> wrote:

> Dear Jimmy,
>
> I noticed that my name node was not started. That might be the reason? I
> still tried to figure out why the name node was not started.
>
> Thanks so much!
>
> Bing
>
>
> On Tue, Feb 14, 2012 at 5:11 AM, Jimmy Xiang <jxiang@cloudera.com> wrote:
>
>> Which HDFS/Hadoop are you using?  The name node configuration for
>> fs.default.name
>> should be hdfs://localhost:9000 if you want your hbase.rootdir to be
>> hdfs://localhost:9000/
>> hbase.  They need to match.
>>
>>
>>
>>
>> On Mon, Feb 13, 2012 at 11:58 AM, Bing Li <lblabs@gmail.com> wrote:
>>
>>> Dear Jimmy,
>>>
>>> Thanks so much for your instant reply!
>>>
>>> My hbase-site.xml is like the following.
>>>
>>>   <property>
>>>     <name>hbase.rootdir</name>
>>>     <value>hdfs://localhost:9000/hbase</value>
>>>   </property>
>>>   <property>
>>>     <name>dfs.replication</name>
>>>     <value>1</value>
>>>   </property>
>>>   <property>
>>>     <name>hbase.master</name>
>>>     <value>localhost:60000</value>
>>>   </property>
>>>   <property>
>>>     <name>hbase.cluster.distributed</name>
>>>     <value>true</value>
>>>   </property>
>>>   <property>
>>>     <name>hbase.zookeeper.quorum</name>
>>>     <value>localhost</value>
>>>   </property>
>>>
>>> When I run hadoop fs -ls /, the directories and files under the linux
>>> root are displayed.
>>>
>>> Best,
>>> Bing
>>>
>>> On Tue, Feb 14, 2012 at 3:48 AM, Jimmy Xiang <jxiang@cloudera.com>wrote:
>>>
>>>> Which port does your HDFS listen to? It is not 9000, right?
>>>>
>>>> <name>hbase.rootdir</name>
>>>>     <value>hdfs://localhost:9000/hbase</value>
>>>>
>>>> You need to fix this and make sure your HDFS is working, for example,
>>>> the following command should work for you.
>>>>
>>>> hadoop fs -ls /
>>>>
>>>>
>>>>
>>>> On Mon, Feb 13, 2012 at 11:44 AM, Bing Li <lblabs@gmail.com> wrote:
>>>>
>>>>> Dear Jimmy,
>>>>>
>>>>> I configured the standalone mode successfully. But I wonder why the
>>>>> pseudo distributed one does work.
>>>>>
>>>>> I checked in logs and got the following exceptions. Does the
>>>>> information give you some hints?
>>>>>
>>>>> Thanks so much for your help again!
>>>>>
>>>>> Best,
>>>>> Bing
>>>>>
>>>>> 2012-02-13 18:25:49,782 FATAL org.apache.hadoop.hbase.master.HMaster:
>>>>> Unhandled exception. Starting shutdown.
>>>>> java.net.ConnectException: Call to localhost/127.0.0.1:9000 failed on
>>>>> connection exception: java.net.ConnectException: Connection refuse
>>>>> d
>>>>> at org.apache.hadoop.ipc.Client.wrapException(Client.java:1095)
>>>>>  at org.apache.hadoop.ipc.Client.call(Client.java:1071)
>>>>> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
>>>>>  at $Proxy10.getProtocolVersion(Unknown Source)
>>>>> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:396)
>>>>>  at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:379)
>>>>> at
>>>>> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:119)
>>>>>  at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:238)
>>>>> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:203)
>>>>>  at
>>>>> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89)
>>>>> at
>>>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1386)
>>>>>  at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
>>>>> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1404)
>>>>>  at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:254)
>>>>> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:187)
>>>>>  at org.apache.hadoop.hbase.util.FSUtils.getRootDir(FSUtils.java:471)
>>>>> at
>>>>> org.apache.hadoop.hbase.master.MasterFileSystem.<init>(MasterFileSystem.java:94)
>>>>>  at
>>>>> org.apache.hadoop.hbase.master.HMaster.finishInitialization(HMaster.java:448)
>>>>> at org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:326)
>>>>>  at java.lang.Thread.run(Thread.java:662)
>>>>> Caused by: java.net.ConnectException: Connection refused
>>>>> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>>>>>  at
>>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
>>>>> at
>>>>> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
>>>>>  at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:656)
>>>>> at
>>>>> org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:434)
>>>>>  at
>>>>> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:560)
>>>>> at org.apache.hadoop.ipc.Client$Connection.access$2000(Client.java:184)
>>>>>  at org.apache.hadoop.ipc.Client.getConnection(Client.java:1202)
>>>>> at org.apache.hadoop.ipc.Client.call(Client.java:1046)
>>>>>  ... 18 more
>>>>> 2012-02-13 18:25:49,787 INFO org.apache.hadoop.hbase.master.HMaster:
>>>>> Aborting
>>>>> 2012-02-13 18:25:49,787 DEBUG org.apache.hadoop.hbase.master.HMaster:
>>>>> Stopping service threads
>>>>>
>>>>>
>>>>> Thanks so much!
>>>>> Bing
>>>>>
>>>>>
>>>>> On Tue, Feb 14, 2012 at 3:35 AM, Jimmy Xiang <jxiang@cloudera.com>wrote:
>>>>>
>>>>>> In this case, you may just use the standalone mode.  You can follow
>>>>>> the quick start step by step.
>>>>>>
>>>>>> The default zookeeper port is 2181, you don't need to configure it.
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Mon, Feb 13, 2012 at 11:28 AM, Bing Li <lblabs@gmail.com>
wrote:
>>>>>>
>>>>>>> Dear Jimmy,
>>>>>>>
>>>>>>> I am a new user of HBase. My experiences in HBase and Hadoop
is very
>>>>>>> limited. I just tried to follow some books, such as Hadoop/HBase
the
>>>>>>> Definitive Guide. However, I still got some problems.
>>>>>>>
>>>>>>> What I am trying to do is just to set up a pseudo distributed
HBase
>>>>>>> environment on a single node. After that, I will start my system
>>>>>>> programming in Java. I hope I could deploy the system in fully
distributed
>>>>>>> mode when my system is done.
>>>>>>>
>>>>>>> So what I am configuring is very simple. Do I need to set up
the
>>>>>>> zookeeper port in hbase-site.xml?
>>>>>>>
>>>>>>> Thanks so much!
>>>>>>>
>>>>>>> Best,
>>>>>>> Bing
>>>>>>>
>>>>>>>
>>>>>>> On Tue, Feb 14, 2012 at 3:16 AM, Jimmy Xiang <jxiang@cloudera.com>wrote:
>>>>>>>
>>>>>>>> Have you restarted your HBase after the change?  What's the
>>>>>>>> zookeeper port does your HMaster use?
>>>>>>>>
>>>>>>>> Can you run the following to checkout where is your HMaster
as
>>>>>>>> below?
>>>>>>>>
>>>>>>>> hbase zkcli
>>>>>>>>   then:  get /hbase/master
>>>>>>>>  It should show you master location.
>>>>>>>>
>>>>>>>> It seems you have a distributed installation.  How many
>>>>>>>> regionservers do you have?  Can you check your
>>>>>>>> master web UI to make sure all look fine.
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>> Jimmy
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mon, Feb 13, 2012 at 10:51 AM, Bing Li <lblabs@gmail.com>
wrote:
>>>>>>>>
>>>>>>>>> Dear Jimmy,
>>>>>>>>>
>>>>>>>>> Thanks so much for your reply!
>>>>>>>>>
>>>>>>>>> I didn't set up the zookeeper.quorom. After getting your
email, I
>>>>>>>>> made a change. Now my hbase-site.xml is as follows.
>>>>>>>>>
>>>>>>>>> <configuration>
>>>>>>>>>   <property>
>>>>>>>>>     <name>hbase.rootdir</name>
>>>>>>>>>     <value>hdfs://localhost:9000/hbase</value>
>>>>>>>>>   </property>
>>>>>>>>>   <property>
>>>>>>>>>     <name>dfs.replication</name>
>>>>>>>>>     <value>1</value>
>>>>>>>>>   </property>
>>>>>>>>>   <property>
>>>>>>>>>     <name>hbase.cluster.distributed</name>
>>>>>>>>>     <value>true</value>
>>>>>>>>>   </property>
>>>>>>>>>   <property>
>>>>>>>>>     <name>hbase.zookeeper.quorum</name>
>>>>>>>>>     <value>localhost</value>
>>>>>>>>>   </property>
>>>>>>>>> </configuration>
>>>>>>>>>
>>>>>>>>> The previous error is still existed. I feel weird why
HBase
>>>>>>>>> developers cannot provide a reliable description about
their work.
>>>>>>>>>
>>>>>>>>> Best,
>>>>>>>>> Bing
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Tue, Feb 14, 2012 at 2:16 AM, Jimmy Xiang <jxiang@cloudera.com>wrote:
>>>>>>>>>
>>>>>>>>>> What's your hbase.zookeeper.quorom configuration?
  You can check
>>>>>>>>>> out this quick start guide:
>>>>>>>>>>
>>>>>>>>>> http://hbase.apache.org/book/quickstart.html
>>>>>>>>>>
>>>>>>>>>> Thanks,
>>>>>>>>>> Jimmy
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Mon, Feb 13, 2012 at 10:09 AM, Bing Li <lblabs@gmail.com>wrote:
>>>>>>>>>>
>>>>>>>>>>> Dear all,
>>>>>>>>>>>
>>>>>>>>>>> After searching on the Web and asking for help
from friends, I
>>>>>>>>>>> noticed that
>>>>>>>>>>> the pseudo distributed configuration in the book,
HBase the
>>>>>>>>>>> Definitive
>>>>>>>>>>> Guide, was not complete. Now the ZooKeeper related
exception is
>>>>>>>>>>> fixed.
>>>>>>>>>>> However, I got another error when typing "status"
in the HBase
>>>>>>>>>>> shell.
>>>>>>>>>>>
>>>>>>>>>>>    ERROR: org.apache.hadoop.hbase.MasterNotRunningException:
>>>>>>>>>>> Retried 7
>>>>>>>>>>> Times
>>>>>>>>>>>
>>>>>>>>>>> I am trying to fix it myself. Your help is highly
appreciated.
>>>>>>>>>>>
>>>>>>>>>>> Thanks so much!
>>>>>>>>>>> Bing Li
>>>>>>>>>>>
>>>>>>>>>>> On Mon, Feb 13, 2012 at 5:00 AM, Bing Li <lblabs@gmail.com>
>>>>>>>>>>> wrote:
>>>>>>>>>>>
>>>>>>>>>>> > Dear all,
>>>>>>>>>>> >
>>>>>>>>>>> > I am a new learner of HBase. I tried to
set up my HBase on a
>>>>>>>>>>> > pseudo-distributed HDFS.
>>>>>>>>>>> >
>>>>>>>>>>> > After starting HDFS by running ./start-dfs.sh
and
>>>>>>>>>>> ./start-hbase.sh, I
>>>>>>>>>>> > started the HBase shell.
>>>>>>>>>>> >
>>>>>>>>>>> >    ./hbase shell
>>>>>>>>>>> >
>>>>>>>>>>> > It was started properly. However, when I
typed the command,
>>>>>>>>>>> status, as
>>>>>>>>>>> > follows.
>>>>>>>>>>> >
>>>>>>>>>>> >    hbase(main):001:0> status
>>>>>>>>>>> >
>>>>>>>>>>> > It got the following exception. Since I
had very limited
>>>>>>>>>>> experiences to
>>>>>>>>>>> > use HBase, I could not figure out what the
problem was.
>>>>>>>>>>> >
>>>>>>>>>>> > SLF4J: Class path contains multiple SLF4J
bindings.
>>>>>>>>>>> > SLF4J: Found binding in
>>>>>>>>>>> >
>>>>>>>>>>> [jar:file:/opt/hbase-0.92.0/lib/slf4j-log4j12-1.5.8.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>>>>>>> > SLF4J: Found binding in
>>>>>>>>>>> >
>>>>>>>>>>> [jar:file:/opt/hadoop-1.0.0/lib/slf4j-log4j12-1.4.3.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>>>>>>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindingsfor
an
>>>>>>>>>>> > explanation.
>>>>>>>>>>> > 12/02/13 04:34:01 ERROR zookeeper.RecoverableZooKeeper:
>>>>>>>>>>> ZooKeeper exists
>>>>>>>>>>> > failed after 3 retries
>>>>>>>>>>> > 12/02/13 04:34:01 WARN zookeeper.ZKUtil:
hconnection Unable to
>>>>>>>>>>> set watcher
>>>>>>>>>>> > on znode /hbase/master
>>>>>>>>>>> > org.apache.zookeeper.KeeperException$ConnectionLossException:
>>>>>>>>>>> > KeeperErrorCode = ConnectionLoss for /hbase/master
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.apache.zookeeper.KeeperException.create(KeeperException.java:99)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
>>>>>>>>>>> >         at
>>>>>>>>>>> org.apache.zookeeper.ZooKeeper.exists(ZooKeeper.java:1003)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.exists(RecoverableZooKeeper.java:154)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.apache.hadoop.hbase.zookeeper.ZKUtil.watchAndCheckExists(ZKUtil.java:226)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.apache.hadoop.hbase.zookeeper.ZooKeeperNodeTracker.start(ZooKeeperNodeTracker.java:76)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.setupZookeeperTrackers(HConnectionManager.java:580)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.<init>(HConnectionManager.java:569)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.apache.hadoop.hbase.client.HConnectionManager.getConnection(HConnectionManager.java:186)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.apache.hadoop.hbase.client.HBaseAdmin.<init>(HBaseAdmin.java:98)
>>>>>>>>>>> >         at
>>>>>>>>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>>>>>>>>>>> > Method)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>>>>>>>>>>> >         at
>>>>>>>>>>> java.lang.reflect.Constructor.newInstance(Constructor.java:513)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.jruby.javasupport.JavaConstructor.newInstanceDirect(JavaConstructor.java:275)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.jruby.java.invokers.ConstructorInvoker.call(ConstructorInvoker.java:91)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.jruby.java.invokers.ConstructorInvoker.call(ConstructorInvoker.java:178)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:322)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:178)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:182)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.jruby.java.proxies.ConcreteJavaProxy$2.call(ConcreteJavaProxy.java:47)
>>>>>>>>>>> >         at
>>>>>>>>>>> >
>>>>>>>>>>> org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:322)
>>>>>>>>>>> >
>>>>>>>>>>> > Could you please give me a hand? Thanks
so much!
>>>>>>>>>>> >
>>>>>>>>>>> > Best regards,
>>>>>>>>>>> > Bing
>>>>>>>>>>> >
>>>>>>>>>>> >
>>>>>>>>>>> >
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message