hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Ted Yu <yuzhih...@gmail.com>
Subject Re: Failed to create /hbase.... KeeperErrorCode = ConnectionLoss for /hbase
Date Thu, 01 Apr 2010 19:55:24 GMT
Please check the following entry in hbase-env.sh:

hbase-env.sh:# The directory where pid files are stored. /tmp by default.
hbase-env.sh:# export HBASE_PID_DIR=/var/hadoop/pids

If pid file is stored under /tmp, it might have been cleaned up.

On Thu, Apr 1, 2010 at 11:44 AM, Jean-Daniel Cryans <jdcryans@apache.org>wrote:

> If the master doesn't shut down, it means it's waiting on something...
> you looked at the logs?
>
> You say you ran ./jps ... did you install that in the local directory?
> Also what do you mean "it didn't work as well"? What didn't work? The
> command didn't return anything or the HMaster process wasn't listed?
>
> Also did you check the zookeeper logs like Patrick said? You should
> see in there when the master tries to connect, and you should see why
> it wasn't able to do so.
>
> To help you I need more data about your problem.
>
> J-D
>
> On Thu, Apr 1, 2010 at 11:39 AM, jayavelu jaisenthilkumar
> <joysenthil@gmail.com> wrote:
> > Hi Daniel,
> >                   I removed the property tags from the hbase-site.xml.
> >
> > Same error occurs.
> >
> > Also one strange behaviour,  If i give ./stop-hbase.sh , the terminal
> says
> > stopping master ....................................
> > and never stopped.
> >
> > I couldnt able to ./jps to check the java in this scenario, it didnt work
> > aswell.  So I killed the Hmaster start ( ps -ef | grep java)
> >
> > Also manually need to kill Hregionserver both on master, slave1 and
> slave2.
> >
> > Any suggestions please.......
> >
> > Regs,
> > senthil
> > On 31 March 2010 19:15, Jean-Daniel Cryans <jdcryans@apache.org> wrote:
> >
> >> You set the tick time like this:
> >>
> >>  <property>
> >>   <name>hbase.zookeeper.property.tickTime</name>
> >>   <value>1</value>
> >>   <description>Property from ZooKeeper's config zoo.cfg.
> >>   The number of milliseconds of each tick.  See
> >>   zookeeper.session.timeout description.
> >>   </description>
> >>
> >> 1 means HBase has to report to zookeeper every 1 millisecond and if
> >> for any reason it doesn't after 20ms, the session is expired (!!). I
> >> recommend using the default value.
> >>
> >> Also you should keep the same config on every node, rsync can do
> wonders.
> >>
> >> J-D
> >>
> >> On Wed, Mar 31, 2010 at 9:24 AM, jayavelu jaisenthilkumar
> >> <joysenthil@gmail.com> wrote:
> >> > Hi,
> >> >            I am using 1 master and 2 slaves one has password for ssh.
> >> >
> >> > I am using hadoop0.20.1 and hbase0.20.3(direct one not upgraded one)
> >> >
> >> > 1)The slave one with password is could not be disabled, i removed the
> >> whole
> >> > .ssh directory try to ssh-keygen with passwordless phrase, still i am
> >> asked
> >> > for the password  when i
> >> > "ssh localhost"
> >> >
> >> > 2) I am able to run hadoop and successfuly run the Mapreduce in the
> >> hadoop
> >> > environment as per the Running Hadoop On Ubuntu Linux (Multi-Node
> >> Cluster)
> >> > by noel
> >> >
> >> > 3) I am now following the tutorial hbase: overview HBase 0.20.3 API
> >> >
> >> > Its not clearly stated as the mulitnode cluster hadoop for the
> >> distributed
> >> > mode hbase.
> >> >
> >> > I ran the hdfs and the hbase using start-dfs.sh and start-hbase.sh
> >> > respectively.
> >> >
> >> > The master log indicates connection loss on the /hbase :  ( is this
> hbase
> >> is
> >> > created by Hbase or should we do to create it again
> >> >
> >> > 2010-03-31 16:45:57,850 INFO org.apache.zookeeper.
> >> > ClientCnxn: Attempting connection to server Hadoopserver/
> >> 192.168.1.65:2222
> >> > 2010-03-31 16:45:57,858 INFO org.apache.zookeeper.ClientCnxn: Priming
> >> > connection to java.nio.channels.SocketChannel[connected local=/
> >> > 192.168.1.65:43017 remote=Hadoopserver/192.168.1.65:2222]
> >> > 2010-03-31 16:45:57,881 INFO org.apache.zookeeper.ClientCnxn: Server
> >> > connection successful
> >> > 2010-03-31 16:45:57,883 WARN org.apache.zookeeper.ClientCnxn:
> Exception
> >> > closing session 0x0 to sun.nio.ch.SelectionKeyImpl@11c2b67
> >> > java.io.IOException: Read error rc = -1
> java.nio.DirectByteBuffer[pos=0
> >> > lim=4 cap=4]
> >> >    at
> >> org.apache.zookeeper.ClientCnxn$SendThread.doIO(ClientCnxn.java:701)
> >> >    at
> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:945)
> >> > 2010-03-31 16:45:57,885 WARN org.apache.zookeeper.ClientCnxn: Ignoring
> >> > exception during shutdown input
> >> > java.net.SocketException: Transport endpoint is not connected
> >> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
> >> >    at
> >> > sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:640)
> >> >    at sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360)
> >> >    at
> >> >
> org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:999)
> >> >    at
> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
> >> > 2010-03-31 16:45:57,885 WARN org.apache.zookeeper.ClientCnxn: Ignoring
> >> > exception during shutdown output
> >> > java.net.SocketException: Transport endpoint is not connected
> >> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
> >> >    at
> >> >
> sun.nio.ch.SocketChannelImpl.shutdownOutput(SocketChannelImpl.java:651)
> >> >    at sun.nio.ch.SocketAdaptor.shutdownOutput(SocketAdaptor.java:368)
> >> >    at
> >> >
> org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:1004)
> >> >    at
> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
> >> > 2010-03-31 16:45:57,933 INFO
> >> org.apache.hadoop.hbase.master.RegionManager:
> >> > -ROOT- region unset (but not set to be reassigned)
> >> > 2010-03-31 16:45:57,934 INFO
> >> org.apache.hadoop.hbase.master.RegionManager:
> >> > ROOT inserted into regionsInTransition
> >> > 2010-03-31 16:45:58,024 DEBUG
> >> > org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper: Failed to read:
> >> > org.apache.zookeeper.KeeperException$ConnectionLossException:
> >> > KeeperErrorCode = ConnectionLoss for /hbase/master
> >> > 2010-03-31 16:45:58,422 INFO org.apache.zookeeper.ClientCnxn:
> Attempting
> >> > connection to server Hadoopclient1/192.168.1.2:2222
> >> > 2010-03-31 16:45:58,423 INFO org.apache.zookeeper.ClientCnxn: Priming
> >> > connection to java.nio.channels.SocketChannel[connected local=/
> >> > 192.168.1.65:51219 remote=Hadoopclient1/192.168.1.2:2222]
> >> > 2010-03-31 16:45:58,423 INFO org.apache.zookeeper.ClientCnxn: Server
> >> > connection successful
> >> > 2010-03-31 16:45:58,436 WARN org.apache.zookeeper.ClientCnxn:
> Exception
> >> > closing session 0x0 to sun.nio.ch.SelectionKeyImpl@17b6643
> >> > java.io.IOException: Read error rc = -1
> java.nio.DirectByteBuffer[pos=0
> >> > lim=4 cap=4]
> >> >    at
> >> org.apache.zookeeper.ClientCnxn$SendThread.doIO(ClientCnxn.java:701)
> >> >    at
> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:945)
> >> > 2010-03-31 16:45:58,437 WARN org.apache.zookeeper.ClientCnxn: Ignoring
> >> > exception during shutdown input
> >> > java.net.SocketException: Transport endpoint is not connected
> >> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
> >> >    at
> >> > sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:640)
> >> >    at sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360)
> >> >    at
> >> >
> org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:999)
> >> >    at
> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
> >> > 2010-03-31 16:45:58,437 WARN org.apache.zookeeper.ClientCnxn: Ignoring
> >> > exception during shutdown output
> >> > java.net.SocketException: Transport endpoint is not connected
> >> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
> >> >    at
> >> >
> sun.nio.ch.SocketChannelImpl.shutdownOutput(SocketChannelImpl.java:651)
> >> >    at sun.nio.ch.SocketAdaptor.shutdownOutput(SocketAdaptor.java:368)
> >> >    at
> >> >
> org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:1004)
> >> >    at
> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
> >> > 2010-03-31 16:45:58,537 WARN
> >> > org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper: Failed to create
> >> /hbase
> >> > -- check quorum servers,
> >> > currently=Hadoopclient1:2222,Hadoopclient:2222,Hadoopserver:2222
> >> > org.apache.zookeeper.KeeperException$ConnectionLossException:
> >> > KeeperErrorCode = ConnectionLoss for /hbase
> >> >    at
> >> org.apache.zookeeper.KeeperException.create(KeeperException.java:90)
> >> >    at
> >> org.apache.zookeeper.KeeperException.create(KeeperException.java:42)
> >> >    at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:608)
> >> >    at
> >> >
> >>
> org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.ensureExists(ZooKeeperWrapper.java:405)
> >> >    at
> >> >
> >>
> org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.ensureParentExists(ZooKeeperWrapper.java:428)
> >> >    at
> >> >
> >>
> org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.writeMasterAddress(ZooKeeperWrapper.java:516)
> >> >    at
> >> >
> >>
> org.apache.hadoop.hbase.master.HMaster.writeAddressToZooKeeper(HMaster.java:263)
> >> >    at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:245)
> >> >    at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> >> Method)
> >> >    at
> >> >
> >>
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
> >> >    at
> >> >
> >>
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
> >> >    at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
> >> >    at org.apache.hadoop.hbase.master.HMaster.doMain(HMaster.java:1241)
> >> >    at org.apache.hadoop.hbase.master.HMaster.main(HMaster.java:1282)
> >> > 2010-03-31 16:45:58,549 INFO org.apache.zookeeper.ClientCnxn:
> Attempting
> >> > connection to server Hadoopclient/192.168.1.3:2222
> >> > 2010-03-31 16:45:58,550 INFO org.apache.zookeeper.ClientCnxn: Priming
> >> > connection to java.nio.channels.SocketChannel[connected local=/
> >> > 192.168.1.65:56142 remote=Hadoopclient/192.168.1.3:2222]
> >> > 2010-03-31 16:45:58,550 INFO org.apache.zookeeper.ClientCnxn: Server
> >> > connection successful
> >> > 2010-03-31 16:45:58,577 WARN org.apache.zookeeper.ClientCnxn:
> Exception
> >> > closing session 0x0 to sun.nio.ch.SelectionKeyImpl@76e8a7
> >> > java.io.IOException: Read error rc = -1
> java.nio.DirectByteBuffer[pos=0
> >> > lim=4 cap=4]
> >> >    at
> >> org.apache.zookeeper.ClientCnxn$SendThread.doIO(ClientCnxn.java:701)
> >> >    at
> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:945)
> >> > 2010-03-31 16:45:58,577 WARN org.apache.zookeeper.ClientCnxn: Ignoring
> >> > exception during shutdown input
> >> > java.net.SocketException: Transport endpoint is not connected
> >> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
> >> >    at
> >> > sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:640)
> >> >    at sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360)
> >> >    at
> >> >
> org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:999)
> >> >    at
> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
> >> > 2010-03-31 16:45:58,577 WARN org.apache.zookeeper.ClientCnxn: Ignoring
> >> > exception during shutdown output
> >> > java.net.SocketException: Transport endpoint is not connected
> >> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
> >> >    at
> >> >
> sun.nio.ch.SocketChannelImpl.shutdownOutput(SocketChannelImpl.java:651)
> >> >    at sun.nio.ch.SocketAdaptor.shutdownOutput(SocketAdaptor.java:368)
> >> >    at
> >> >
> org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:1004)
> >> >    at
> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
> >> > 2010-03-31 16:45:58,678 DEBUG
> >> > org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper: Failed to read:
> >> > org.apache.zookeeper.KeeperException$ConnectionLossException:
> >> > KeeperErrorCode = ConnectionLoss for /hbase/master
> >> >
> >> > my hbase-site.xml ( Master)
> >> > <configuration>
> >> >  <property>
> >> >    <name>hbase.rootdir</name>
> >> >    <value>hdfs://Hadoopserver:54310/hbase</value>
> >> >    <description>The directory shared by region servers.
> >> >    Should be fully-qualified to include the filesystem to use.
> >> >    E.g: hdfs://NAMENODE_SERVER:PORT/HBASE_ROOTDIR
> >> >    </description>
> >> >  </property>
> >> >  <property>
> >> >    <name>hbase.cluster.distributed</name>
> >> >    <value>true</value>
> >> >    <description>The mode the cluster will be in. Possible values
are
> >> >      false: standalone and pseudo-distributed setups with managed
> >> Zookeeper
> >> >      true: fully-distributed with unmanaged Zookeeper Quorum (see
> >> > hbase-env.sh)
> >> >    </description>
> >> >  </property>
> >> >  <property>
> >> >    <name>hbase.zookeeper.quorum</name>
> >> >    <value>Hadoopserver,Hadoopclient1,Hadoopclient</value>
> >> >    <description>Comma separated list of servers in the ZooKeeper
> Quorum.
> >> >    For example, "host1.mydomain.com,host2.mydomain.com,
> >> host3.mydomain.com".
> >> >    By default this is set to localhost for local and
> pseudo-distributed
> >> > modes
> >> >    of operation. For a fully-distributed setup, this should be set to
> a
> >> > full
> >> >    list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in
> >> > hbase-env.sh
> >> >    this is the list of servers which we will start/stop ZooKeeper on.
> >> >    </description>
> >> >  </property>
> >> >  <property>
> >> >    <name>hbase.zookeeper.property.tickTime</name>
> >> >    <value>1</value>
> >> >    <description>Property from ZooKeeper's config zoo.cfg.
> >> >    The number of milliseconds of each tick.  See
> >> >    zookeeper.session.timeout description.
> >> >    </description>
> >> >  </property>
> >> > <property>
> >> >    <name>zookeeper.retries</name>
> >> >    <value>5</value>
> >> >    <description>How many times to retry connections to ZooKeeper.
Used
> >> for
> >> >      reading/writing root region location, checking/writing out of
> safe
> >> > mode.
> >> >      Used together with ${zookeeper.pause} in an exponential backoff
> >> > fashion
> >> >      when making queries to ZooKeeper.
> >> >    </description>
> >> >  </property>
> >> > <property>
> >> > <name>hbase.zookeeper.property.clientPort</name>
> >> > <value>2222</value>
> >> > <description>Property from ZooKeeper's config zoo.cfg.
> >> > the port at which the clients will connect.
> >> > </description>
> >> > </property>
> >> >
> >> > <property>
> >> >  <name>dfs.replication</name>
> >> > <value>3</value>
> >> > <description>Default block replication.
> >> > The actual number of replications can be specified when the file is
> >> created.
> >> > The default is used if replication is not specified in create time.
> >> > </description>
> >> > </property>
> >> > </configuration>
> >> >
> >> >
> >> > The hbase-site.xml in slave1, slave2
> >> > <configuration>
> >> >  <property>
> >> >    <name>hbase.rootdir</name>
> >> >    <value>hdfs://Hadoopserver:54310/hbase</value>
> >> >    <description>The directory shared by region servers.
> >> >    Should be fully-qualified to include the filesystem to use.
> >> >    E.g: hdfs://NAMENODE_SERVER:PORT/HBASE_ROOTDIR
> >> >    </description>
> >> >  </property>
> >> >  <property>
> >> >    <name>hbase.cluster.distributed</name>
> >> >    <value>true</value>
> >> >    <description>The mode the cluster will be in. Possible values
are
> >> >      false: standalone and pseudo-distributed setups with managed
> >> Zookeeper
> >> >      true: fully-distributed with unmanaged Zookeeper Quorum (see
> >> > hbase-env.sh)
> >> >    </description>
> >> >  </property>
> >> >  <property>
> >> >    <name>hbase.zookeeper.quorum</name>
> >> >    <value>Hadoopserver,Hadoopclient1,Hadoopclient</value>
> >> >    <description>Comma separated list of servers in the ZooKeeper
> Quorum.
> >> >    For example, "host1.mydomain.com,host2.mydomain.com,
> >> host3.mydomain.com".
> >> >    By default this is set to localhost for local and
> pseudo-distributed
> >> > modes
> >> >    of operation. For a fully-distributed setup, this should be set to
> a
> >> > full
> >> >    list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in
> >> > hbase-env.sh
> >> >    this is the list of servers which we will start/stop ZooKeeper on.
> >> >    </description>
> >> >  </property>
> >> >
> >> > <property>
> >> > <name>hbase.zookeeper.property.clientPort</name>
> >> > <value>2222</value>
> >> > <description>Property from ZooKeeper's config zoo.cfg.
> >> > the port at which the clients will connect.
> >> > </description>
> >> > </property>
> >> >
> >> > <property>
> >> >  <name>dfs.replication</name>
> >> > <value>3</value>
> >> > <description>Default block replication.
> >> > The actual number of replications can be specified when the file is
> >> created.
> >> > The default is used if replication is not specified in create time.
> >> > </description>
> >> > </property>
> >> > </configuration>
> >> >
> >> > Regionservers ( only on Master in my case Hadoopserver)
> >> >
> >> > Hadoopserver
> >> > Hadoopclient1
> >> > Hadoopclient
> >> >
> >> >
> >> > Regionservers ( on slaves)
> >> > localhost
> >> >
> >> > Blocked with this error for the past one week. Googled so much didnt
> get
> >> any
> >> > solution.
> >> >
> >> >
> >> > REgs.
> >> > senthil
> >> >
> >>
> >
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message