hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Puri, Aseem" <Aseem.P...@Honeywell.com>
Subject RE: problem in configuring hbase with hdfs
Date Wed, 18 Mar 2009 04:24:27 GMT
Hi 
	There is exception while staring HBase master. The exceptions are:

Wed Mar 18 09:10:51 IST 2009 Starting master on ie11dtxpficbfise
java version "1.6.0_11"
Java(TM) SE Runtime Environment (build 1.6.0_11-b03)
Java HotSpot(TM) Client VM (build 11.0-b16, mixed mode, sharing)
ulimit -n 256
2009-03-18 09:10:55,650 INFO org.apache.hadoop.hbase.master.HMaster: Root region dir: hdfs://localhost:9000/hbase/-ROOT-/70236052
2009-03-18 09:10:55,806 FATAL org.apache.hadoop.hbase.master.HMaster: Not starting HMaster
because:
java.io.EOFException
	at java.io.DataInputStream.readUnsignedShort(DataInputStream.java:323)
	at java.io.DataInputStream.readUTF(DataInputStream.java:572)
	at org.apache.hadoop.hbase.util.FSUtils.getVersion(FSUtils.java:101)
	at org.apache.hadoop.hbase.util.FSUtils.checkVersion(FSUtils.java:120)
	at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:203)
	at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:147)
	at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
	at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
	at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
	at org.apache.hadoop.hbase.master.HMaster.doMain(HMaster.java:784)
	at org.apache.hadoop.hbase.master.HMaster.main(HMaster.java:818)
2009-03-18 09:10:55,806 ERROR org.apache.hadoop.hbase.master.HMaster: Can not start master
java.lang.reflect.InvocationTargetException
	at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
	at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
	at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
	at org.apache.hadoop.hbase.master.HMaster.doMain(HMaster.java:784)
	at org.apache.hadoop.hbase.master.HMaster.main(HMaster.java:818)
Caused by: java.io.EOFException
	at java.io.DataInputStream.readUnsignedShort(DataInputStream.java:323)
	at java.io.DataInputStream.readUTF(DataInputStream.java:572)
	at org.apache.hadoop.hbase.util.FSUtils.getVersion(FSUtils.java:101)
	at org.apache.hadoop.hbase.util.FSUtils.checkVersion(FSUtils.java:120)
	at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:203)
	at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:147)
	... 6 more

Also when I start my start my hadoop server datanode not started it also
throw exceptions. Datanode exceptions are:

2009-03-18 09:08:47,354 INFO org.apache.hadoop.dfs.DataNode: STARTUP_MSG: 
/************************************************************
STARTUP_MSG: Starting DataNode
STARTUP_MSG:   host = ie11dtxpficbfise/199.63.66.65
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 0.18.0
STARTUP_MSG:   build = http://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.18 -r
686010; compiled by 'hadoopqa' on Thu Aug 14 19:48:33 UTC 2008
************************************************************/
2009-03-18 09:08:56,667 ERROR org.apache.hadoop.dfs.DataNode: org.apache.hadoop.dfs.IncorrectVersionException:
Unexpected version of storage directory C:\tmp\hadoop-HadoopAdmin\dfs\data. Reported: -18.
Expecting = -16.
	at org.apache.hadoop.dfs.Storage.getFields(Storage.java:584)
	at org.apache.hadoop.dfs.DataStorage.getFields(DataStorage.java:171)
	at org.apache.hadoop.dfs.Storage$StorageDirectory.read(Storage.java:164)
	at org.apache.hadoop.dfs.Storage$StorageDirectory.read(Storage.java:153)
	at org.apache.hadoop.dfs.DataStorage.doTransition(DataStorage.java:221)
	at org.apache.hadoop.dfs.DataStorage.recoverTransitionRead(DataStorage.java:141)
	at org.apache.hadoop.dfs.DataNode.startDataNode(DataNode.java:273)
	at org.apache.hadoop.dfs.DataNode.<init>(DataNode.java:190)
	at org.apache.hadoop.dfs.DataNode.makeInstance(DataNode.java:2987)
	at org.apache.hadoop.dfs.DataNode.instantiateDataNode(DataNode.java:2942)
	at org.apache.hadoop.dfs.DataNode.createDataNode(DataNode.java:2950)
	at org.apache.hadoop.dfs.DataNode.main(DataNode.java:3072)

2009-03-18 09:08:56,682 INFO org.apache.hadoop.dfs.DataNode: SHUTDOWN_MSG: 
/************************************************************
SHUTDOWN_MSG: Shutting down DataNode at ie11dtxpficbfise/199.63.66.65
************************************************************/

So is there any relation of Hbase master with hadoop datanode exceptions that's why it do
not start? 

Please tell where is the problem bcoz of it Hbase master do not start

Thanks & Regards
Aseem Puri


-----Original Message-----
From: jdcryans@gmail.com [mailto:jdcryans@gmail.com] On Behalf Of Jean-Daniel Cryans
Sent: Tuesday, March 17, 2009 5:26 PM
To: hbase-user@hadoop.apache.org
Subject: Re: problem in configuring hbase with hdfs

Aseem,

It tells you that there is no master to stop so it means that
something went wrong when it got started and shut down by itself. Can
you look in your master log and see if there are any exceptions
thrown?

Thx,

J-D

On Tue, Mar 17, 2009 at 3:32 AM, Puri, Aseem <Aseem.Puri@honeywell.com> wrote:
> Hi
>
>
>
> I am newbie working on Hadoop - HBase. I am using Hadoop-0.18.0 and
> HBase-0.18.1. There is some problem with using HBase master when HBase
> uses HDFS. My Hadoop-site configuration is:
>
>
>
> <configuration>
>
>  <property>
>
>    <name>fs.default.name</name>
>
>    <value>hdfs://localhost:9000</value>
>
>  </property>
>
>  <property>
>
>    <name>mapred.job.tracker</name>
>
>    <value>localhost:9001</value>
>
>  </property>
>
>  <property>
>
>    <name>dfs.replication</name>
>
>    <value>1</value>
>
>  </property>
>
> </configuration>
>
>
>
> And HBase-site configuration is:
>
>
>
> <configuration>
>
>  <property>
>
>    <name>hbase.rootdir</name>
>
>    <value>hdfs://localhost:9000/hbase</value>
>
>    <description>The directory shared by region servers.
>
>    </description>
>
>  </property>
>
>  <property>
>
>    <name>hbase.master</name>
>
>    <value>localhost:60000</value>
>
>    <description>The host and port that the HBase master runs at.
>
>    </description>
>
>  </property>
>
> <property>
>
>    <name>hbase.regionserver</name>
>
>    <value>localhost:60020</value>
>
>    <description>The host and port a HBase region server runs at.
>
>    </description>
>
>  </property>
>
>  </configuration>
>
>
>
> When I try command $ bin/hadoop dfs -ls / I got follwing result:
>
>
>
> $ bin/hadoop dfs -ls /
>
> Found 2 items
>
> drwxr-xr-x   - HadoopAdmin supergroup          0 2009-03-17 12:22 /hbase
>
> drwxr-xr-x   - HadoopAdmin supergroup          0 2009-03-17 12:21 /tmp
>
>
>
> But when I use HBase command list in $ bin/hbase shell I got following
> exception:
>
>
>
> hbase(main):001:0> list
>
> 09/03/17 12:24:24 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0
>
> .1:60000. Already tried 0 time(s).
>
> 09/03/17 12:24:26 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0
>
> .1:60000. Already tried 1 time(s).
>
> 09/03/17 12:24:28 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0
>
> .1:60000. Already tried 2 time(s).
>
> 09/03/17 12:24:29 INFO client.HConnectionManager$TableServers: Attempt 0
> of 5 fa
>
> iled with <java.io.IOException: Call failed on local exception>.
> Retrying after
>
> sleep of 2000
>
> 09/03/17 12:24:33 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0
>
> .1:60000. Already tried 0 time(s).
>
> 09/03/17 12:24:35 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0
>
> .1:60000. Already tried 1 time(s).
>
> 09/03/17 12:24:37 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0
>
> .1:60000. Already tried 2 time(s).
>
> 09/03/17 12:24:38 INFO client.HConnectionManager$TableServers: Attempt 1
> of 5 fa
>
> iled with <java.io.IOException: Call failed on local exception>.
> Retrying after
>
> sleep of 2000
>
> 09/03/17 12:24:42 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0
>
> .1:60000. Already tried 0 time(s).
>
> 09/03/17 12:24:44 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0
>
> .1:60000. Already tried 1 time(s).
>
> 09/03/17 12:24:46 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0
>
> .1:60000. Already tried 2 time(s).
>
> 09/03/17 12:24:48 INFO client.HConnectionManager$TableServers: Attempt 2
> of 5 fa
>
> iled with <java.io.IOException: Call failed on local exception>.
> Retrying after
>
> sleep of 2000
>
> 09/03/17 12:24:52 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0
>
> .1:60000. Already tried 0 time(s).
>
> 09/03/17 12:24:54 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0
>
> .1:60000. Already tried 1 time(s).
>
> 09/03/17 12:24:56 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0
>
> .1:60000. Already tried 2 time(s).
>
> 09/03/17 12:24:57 INFO client.HConnectionManager$TableServers: Attempt 3
> of 5 fa
>
> iled with <java.io.IOException: Call failed on local exception>.
> Retrying after
>
> sleep of 4000
>
> 09/03/17 12:25:03 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0
>
> .1:60000. Already tried 0 time(s).
>
> 09/03/17 12:25:05 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0
>
> .1:60000. Already tried 1 time(s).
>
> 09/03/17 12:25:07 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0
>
> .1:60000. Already tried 2 time(s).
>
> NativeException: org.apache.hadoop.hbase.MasterNotRunningException:
> localhost:60
>
> 000
>
>        from
> org/apache/hadoop/hbase/client/HConnectionManager.java:221:in `getM
>
> aster'
>
>        from org/apache/hadoop/hbase/client/HBaseAdmin.java:67:in
> `<init>'
>
>        from sun/reflect/NativeConstructorAccessorImpl.java:-2:in
> `newInstance0'
>
>
>
>        from sun/reflect/NativeConstructorAccessorImpl.java:39:in
> `newInstance'
>
>        from sun/reflect/DelegatingConstructorAccessorImpl.java:27:in
> `newInstan
>
> ce'
>
>        from java/lang/reflect/Constructor.java:513:in `newInstance'
>
>        from org/jruby/javasupport/JavaConstructor.java:195:in
> `new_instance'
>
>        from
> org.jruby.javasupport.JavaConstructorInvoker$new_instance_method_0_
>
> 0:-1:in `call'
>
>        from org/jruby/runtime/CallSite.java:261:in `call'
>
>        from org/jruby/evaluator/ASTInterpreter.java:670:in `callNode'
>
>        from org/jruby/evaluator/ASTInterpreter.java:324:in
> `evalInternal'
>
>        from org/jruby/evaluator/ASTInterpreter.java:2173:in `setupArgs'
>
>        from org/jruby/evaluator/ASTInterpreter.java:571:in
> `attrAssignNode'
>
>        from org/jruby/evaluator/ASTInterpreter.java:309:in
> `evalInternal'
>
>        from org/jruby/evaluator/ASTInterpreter.java:620:in `blockNode'
>
>        from org/jruby/evaluator/ASTInterpreter.java:318:in
> `evalInternal'
>
> ... 178 levels...
>
>        from
> ruby/C_3a_/Documents_20_and_20_Settings/HadoopAdmin/hbase/bin/C:\DO
>
> CUME~1\HADOOP~1\hbase\/bin/hirb.rb:-1:in `__file__'
>
>        from
> ruby/C_3a_/Documents_20_and_20_Settings/HadoopAdmin/hbase/bin/C:\DO
>
> CUME~1\HADOOP~1\hbase\/bin/hirb.rb:-1:in `load'
>
>        from org/jruby/Ruby.java:512:in `runScript'
>
>        from org/jruby/Ruby.java:432:in `runNormally'
>
>        from org/jruby/Ruby.java:312:in `runFromMain'
>
>        from org/jruby/Main.java:144:in `run'
>
>        from org/jruby/Main.java:89:in `run'
>
>        from org/jruby/Main.java:80:in `main'
>
>        from file:/C:/Documents and
> Settings/HadoopAdmin/hbase/lib/jruby-complet
>
> e-1.1.2.jar!/builtin/javasupport/proxy/concrete.rb:23:in `initialize'
>
>        from file:/C:/Documents and
> Settings/HadoopAdmin/hbase/lib/jruby-complet
>
> e-1.1.2.jar!/builtin/javasupport/proxy/concrete.rb:6:in `new'
>
>        from file:/C:/Documents and
> Settings/HadoopAdmin/hbase/lib/jruby-complet
>
> e-1.1.2.jar!/builtin/javasupport/proxy/concrete.rb:6:in `new'
>
>        from C:/DOCUME~1/HADOOP~1/hbase/bin/HBase.rb:37:in `initialize'
>
>        from C:\DOCUME~1\HADOOP~1\hbase\/bin/hirb.rb:218:in `new'
>
>        from C:\DOCUME~1\HADOOP~1\hbase\/bin/hirb.rb:218:in `admin'
>
>        from C:\DOCUME~1\HADOOP~1\hbase\/bin/hirb.rb:242:in `list'
>
>        from (hbase):2:in `binding'hbase(main):002:0>
>
>
>
>
>
> Also when I try to stop HBase with the command $ bin/stop-hbase.sh I got
> the message:
>
> no master to stop
>
>
>
> When I change my HBase-site configuration to:
>
>
>
> <configuration>
>
> <property>
>
>    <name>hbase.master</name>
>
>    <value>localhost:60000</value>
>
>    <description>The host and port that the HBase master runs at.
>
>    </description>
>
>  </property>
>
> <property>
>
>    <name>hbase.regionserver</name>
>
>    <value>localhost:60020</value>
>
>    <description>The host and port a HBase region server runs at.
>
>    </description>
>
>  </property>
>
>  </configuration>
>
>
>
> My master starts working as HBase is now using local file system.
>
>
>
> But when I want to use HDFS and change HBase-site configuration then my
> master do not starts. Please tell me how I should configure so my HBase
> master starts working and HBase use HDFS. Hope you people help me in
> this.
>
>
>
> -Aseem
>
>
>
>

Mime
View raw message