hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From jayavelu jaisenthilkumar <joysent...@gmail.com>
Subject Re: Failed to create /hbase.... KeeperErrorCode = ConnectionLoss for /hbase
Date Thu, 01 Apr 2010 18:39:02 GMT
Hi Daniel,
                   I removed the property tags from the hbase-site.xml.

Same error occurs.

Also one strange behaviour,  If i give ./stop-hbase.sh , the terminal says
stopping master ....................................
and never stopped.

I couldnt able to ./jps to check the java in this scenario, it didnt work
aswell.  So I killed the Hmaster start ( ps -ef | grep java)

Also manually need to kill Hregionserver both on master, slave1 and slave2.

Any suggestions please.......

Regs,
senthil
On 31 March 2010 19:15, Jean-Daniel Cryans <jdcryans@apache.org> wrote:

> You set the tick time like this:
>
>  <property>
>   <name>hbase.zookeeper.property.tickTime</name>
>   <value>1</value>
>   <description>Property from ZooKeeper's config zoo.cfg.
>   The number of milliseconds of each tick.  See
>   zookeeper.session.timeout description.
>   </description>
>
> 1 means HBase has to report to zookeeper every 1 millisecond and if
> for any reason it doesn't after 20ms, the session is expired (!!). I
> recommend using the default value.
>
> Also you should keep the same config on every node, rsync can do wonders.
>
> J-D
>
> On Wed, Mar 31, 2010 at 9:24 AM, jayavelu jaisenthilkumar
> <joysenthil@gmail.com> wrote:
> > Hi,
> >            I am using 1 master and 2 slaves one has password for ssh.
> >
> > I am using hadoop0.20.1 and hbase0.20.3(direct one not upgraded one)
> >
> > 1)The slave one with password is could not be disabled, i removed the
> whole
> > .ssh directory try to ssh-keygen with passwordless phrase, still i am
> asked
> > for the password  when i
> > "ssh localhost"
> >
> > 2) I am able to run hadoop and successfuly run the Mapreduce in the
> hadoop
> > environment as per the Running Hadoop On Ubuntu Linux (Multi-Node
> Cluster)
> > by noel
> >
> > 3) I am now following the tutorial hbase: overview HBase 0.20.3 API
> >
> > Its not clearly stated as the mulitnode cluster hadoop for the
> distributed
> > mode hbase.
> >
> > I ran the hdfs and the hbase using start-dfs.sh and start-hbase.sh
> > respectively.
> >
> > The master log indicates connection loss on the /hbase :  ( is this hbase
> is
> > created by Hbase or should we do to create it again
> >
> > 2010-03-31 16:45:57,850 INFO org.apache.zookeeper.
> > ClientCnxn: Attempting connection to server Hadoopserver/
> 192.168.1.65:2222
> > 2010-03-31 16:45:57,858 INFO org.apache.zookeeper.ClientCnxn: Priming
> > connection to java.nio.channels.SocketChannel[connected local=/
> > 192.168.1.65:43017 remote=Hadoopserver/192.168.1.65:2222]
> > 2010-03-31 16:45:57,881 INFO org.apache.zookeeper.ClientCnxn: Server
> > connection successful
> > 2010-03-31 16:45:57,883 WARN org.apache.zookeeper.ClientCnxn: Exception
> > closing session 0x0 to sun.nio.ch.SelectionKeyImpl@11c2b67
> > java.io.IOException: Read error rc = -1 java.nio.DirectByteBuffer[pos=0
> > lim=4 cap=4]
> >    at
> org.apache.zookeeper.ClientCnxn$SendThread.doIO(ClientCnxn.java:701)
> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:945)
> > 2010-03-31 16:45:57,885 WARN org.apache.zookeeper.ClientCnxn: Ignoring
> > exception during shutdown input
> > java.net.SocketException: Transport endpoint is not connected
> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
> >    at
> > sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:640)
> >    at sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360)
> >    at
> > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:999)
> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
> > 2010-03-31 16:45:57,885 WARN org.apache.zookeeper.ClientCnxn: Ignoring
> > exception during shutdown output
> > java.net.SocketException: Transport endpoint is not connected
> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
> >    at
> > sun.nio.ch.SocketChannelImpl.shutdownOutput(SocketChannelImpl.java:651)
> >    at sun.nio.ch.SocketAdaptor.shutdownOutput(SocketAdaptor.java:368)
> >    at
> > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:1004)
> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
> > 2010-03-31 16:45:57,933 INFO
> org.apache.hadoop.hbase.master.RegionManager:
> > -ROOT- region unset (but not set to be reassigned)
> > 2010-03-31 16:45:57,934 INFO
> org.apache.hadoop.hbase.master.RegionManager:
> > ROOT inserted into regionsInTransition
> > 2010-03-31 16:45:58,024 DEBUG
> > org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper: Failed to read:
> > org.apache.zookeeper.KeeperException$ConnectionLossException:
> > KeeperErrorCode = ConnectionLoss for /hbase/master
> > 2010-03-31 16:45:58,422 INFO org.apache.zookeeper.ClientCnxn: Attempting
> > connection to server Hadoopclient1/192.168.1.2:2222
> > 2010-03-31 16:45:58,423 INFO org.apache.zookeeper.ClientCnxn: Priming
> > connection to java.nio.channels.SocketChannel[connected local=/
> > 192.168.1.65:51219 remote=Hadoopclient1/192.168.1.2:2222]
> > 2010-03-31 16:45:58,423 INFO org.apache.zookeeper.ClientCnxn: Server
> > connection successful
> > 2010-03-31 16:45:58,436 WARN org.apache.zookeeper.ClientCnxn: Exception
> > closing session 0x0 to sun.nio.ch.SelectionKeyImpl@17b6643
> > java.io.IOException: Read error rc = -1 java.nio.DirectByteBuffer[pos=0
> > lim=4 cap=4]
> >    at
> org.apache.zookeeper.ClientCnxn$SendThread.doIO(ClientCnxn.java:701)
> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:945)
> > 2010-03-31 16:45:58,437 WARN org.apache.zookeeper.ClientCnxn: Ignoring
> > exception during shutdown input
> > java.net.SocketException: Transport endpoint is not connected
> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
> >    at
> > sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:640)
> >    at sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360)
> >    at
> > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:999)
> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
> > 2010-03-31 16:45:58,437 WARN org.apache.zookeeper.ClientCnxn: Ignoring
> > exception during shutdown output
> > java.net.SocketException: Transport endpoint is not connected
> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
> >    at
> > sun.nio.ch.SocketChannelImpl.shutdownOutput(SocketChannelImpl.java:651)
> >    at sun.nio.ch.SocketAdaptor.shutdownOutput(SocketAdaptor.java:368)
> >    at
> > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:1004)
> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
> > 2010-03-31 16:45:58,537 WARN
> > org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper: Failed to create
> /hbase
> > -- check quorum servers,
> > currently=Hadoopclient1:2222,Hadoopclient:2222,Hadoopserver:2222
> > org.apache.zookeeper.KeeperException$ConnectionLossException:
> > KeeperErrorCode = ConnectionLoss for /hbase
> >    at
> org.apache.zookeeper.KeeperException.create(KeeperException.java:90)
> >    at
> org.apache.zookeeper.KeeperException.create(KeeperException.java:42)
> >    at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:608)
> >    at
> >
> org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.ensureExists(ZooKeeperWrapper.java:405)
> >    at
> >
> org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.ensureParentExists(ZooKeeperWrapper.java:428)
> >    at
> >
> org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.writeMasterAddress(ZooKeeperWrapper.java:516)
> >    at
> >
> org.apache.hadoop.hbase.master.HMaster.writeAddressToZooKeeper(HMaster.java:263)
> >    at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:245)
> >    at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> Method)
> >    at
> >
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
> >    at
> >
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
> >    at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
> >    at org.apache.hadoop.hbase.master.HMaster.doMain(HMaster.java:1241)
> >    at org.apache.hadoop.hbase.master.HMaster.main(HMaster.java:1282)
> > 2010-03-31 16:45:58,549 INFO org.apache.zookeeper.ClientCnxn: Attempting
> > connection to server Hadoopclient/192.168.1.3:2222
> > 2010-03-31 16:45:58,550 INFO org.apache.zookeeper.ClientCnxn: Priming
> > connection to java.nio.channels.SocketChannel[connected local=/
> > 192.168.1.65:56142 remote=Hadoopclient/192.168.1.3:2222]
> > 2010-03-31 16:45:58,550 INFO org.apache.zookeeper.ClientCnxn: Server
> > connection successful
> > 2010-03-31 16:45:58,577 WARN org.apache.zookeeper.ClientCnxn: Exception
> > closing session 0x0 to sun.nio.ch.SelectionKeyImpl@76e8a7
> > java.io.IOException: Read error rc = -1 java.nio.DirectByteBuffer[pos=0
> > lim=4 cap=4]
> >    at
> org.apache.zookeeper.ClientCnxn$SendThread.doIO(ClientCnxn.java:701)
> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:945)
> > 2010-03-31 16:45:58,577 WARN org.apache.zookeeper.ClientCnxn: Ignoring
> > exception during shutdown input
> > java.net.SocketException: Transport endpoint is not connected
> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
> >    at
> > sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:640)
> >    at sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360)
> >    at
> > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:999)
> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
> > 2010-03-31 16:45:58,577 WARN org.apache.zookeeper.ClientCnxn: Ignoring
> > exception during shutdown output
> > java.net.SocketException: Transport endpoint is not connected
> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
> >    at
> > sun.nio.ch.SocketChannelImpl.shutdownOutput(SocketChannelImpl.java:651)
> >    at sun.nio.ch.SocketAdaptor.shutdownOutput(SocketAdaptor.java:368)
> >    at
> > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:1004)
> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
> > 2010-03-31 16:45:58,678 DEBUG
> > org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper: Failed to read:
> > org.apache.zookeeper.KeeperException$ConnectionLossException:
> > KeeperErrorCode = ConnectionLoss for /hbase/master
> >
> > my hbase-site.xml ( Master)
> > <configuration>
> >  <property>
> >    <name>hbase.rootdir</name>
> >    <value>hdfs://Hadoopserver:54310/hbase</value>
> >    <description>The directory shared by region servers.
> >    Should be fully-qualified to include the filesystem to use.
> >    E.g: hdfs://NAMENODE_SERVER:PORT/HBASE_ROOTDIR
> >    </description>
> >  </property>
> >  <property>
> >    <name>hbase.cluster.distributed</name>
> >    <value>true</value>
> >    <description>The mode the cluster will be in. Possible values are
> >      false: standalone and pseudo-distributed setups with managed
> Zookeeper
> >      true: fully-distributed with unmanaged Zookeeper Quorum (see
> > hbase-env.sh)
> >    </description>
> >  </property>
> >  <property>
> >    <name>hbase.zookeeper.quorum</name>
> >    <value>Hadoopserver,Hadoopclient1,Hadoopclient</value>
> >    <description>Comma separated list of servers in the ZooKeeper Quorum.
> >    For example, "host1.mydomain.com,host2.mydomain.com,
> host3.mydomain.com".
> >    By default this is set to localhost for local and pseudo-distributed
> > modes
> >    of operation. For a fully-distributed setup, this should be set to a
> > full
> >    list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in
> > hbase-env.sh
> >    this is the list of servers which we will start/stop ZooKeeper on.
> >    </description>
> >  </property>
> >  <property>
> >    <name>hbase.zookeeper.property.tickTime</name>
> >    <value>1</value>
> >    <description>Property from ZooKeeper's config zoo.cfg.
> >    The number of milliseconds of each tick.  See
> >    zookeeper.session.timeout description.
> >    </description>
> >  </property>
> > <property>
> >    <name>zookeeper.retries</name>
> >    <value>5</value>
> >    <description>How many times to retry connections to ZooKeeper. Used
> for
> >      reading/writing root region location, checking/writing out of safe
> > mode.
> >      Used together with ${zookeeper.pause} in an exponential backoff
> > fashion
> >      when making queries to ZooKeeper.
> >    </description>
> >  </property>
> > <property>
> > <name>hbase.zookeeper.property.clientPort</name>
> > <value>2222</value>
> > <description>Property from ZooKeeper's config zoo.cfg.
> > the port at which the clients will connect.
> > </description>
> > </property>
> >
> > <property>
> >  <name>dfs.replication</name>
> > <value>3</value>
> > <description>Default block replication.
> > The actual number of replications can be specified when the file is
> created.
> > The default is used if replication is not specified in create time.
> > </description>
> > </property>
> > </configuration>
> >
> >
> > The hbase-site.xml in slave1, slave2
> > <configuration>
> >  <property>
> >    <name>hbase.rootdir</name>
> >    <value>hdfs://Hadoopserver:54310/hbase</value>
> >    <description>The directory shared by region servers.
> >    Should be fully-qualified to include the filesystem to use.
> >    E.g: hdfs://NAMENODE_SERVER:PORT/HBASE_ROOTDIR
> >    </description>
> >  </property>
> >  <property>
> >    <name>hbase.cluster.distributed</name>
> >    <value>true</value>
> >    <description>The mode the cluster will be in. Possible values are
> >      false: standalone and pseudo-distributed setups with managed
> Zookeeper
> >      true: fully-distributed with unmanaged Zookeeper Quorum (see
> > hbase-env.sh)
> >    </description>
> >  </property>
> >  <property>
> >    <name>hbase.zookeeper.quorum</name>
> >    <value>Hadoopserver,Hadoopclient1,Hadoopclient</value>
> >    <description>Comma separated list of servers in the ZooKeeper Quorum.
> >    For example, "host1.mydomain.com,host2.mydomain.com,
> host3.mydomain.com".
> >    By default this is set to localhost for local and pseudo-distributed
> > modes
> >    of operation. For a fully-distributed setup, this should be set to a
> > full
> >    list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in
> > hbase-env.sh
> >    this is the list of servers which we will start/stop ZooKeeper on.
> >    </description>
> >  </property>
> >
> > <property>
> > <name>hbase.zookeeper.property.clientPort</name>
> > <value>2222</value>
> > <description>Property from ZooKeeper's config zoo.cfg.
> > the port at which the clients will connect.
> > </description>
> > </property>
> >
> > <property>
> >  <name>dfs.replication</name>
> > <value>3</value>
> > <description>Default block replication.
> > The actual number of replications can be specified when the file is
> created.
> > The default is used if replication is not specified in create time.
> > </description>
> > </property>
> > </configuration>
> >
> > Regionservers ( only on Master in my case Hadoopserver)
> >
> > Hadoopserver
> > Hadoopclient1
> > Hadoopclient
> >
> >
> > Regionservers ( on slaves)
> > localhost
> >
> > Blocked with this error for the past one week. Googled so much didnt get
> any
> > solution.
> >
> >
> > REgs.
> > senthil
> >
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message