hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Dalia Sobhy <dalia.mohso...@hotmail.com>
Subject RE: Important "Undefined Error"
Date Fri, 18 May 2012 10:50:53 GMT

Hi Shashwat,
I am deploying my cluster on 4 PCs, one master and 3 slave nodes, not all in a single machine...

> From: dalia.mohsobhy@hotmail.com
> To: user@hbase.apache.org
> Subject: RE: Important "Undefined Error"
> Date: Fri, 18 May 2012 12:48:58 +0200
> 
> 
> /etc/hosts
> 127.0.0.1 localhost
> 10.0.2.3	namenode.dalia.com10.0.2.5	datanode3.dalia.com10.0.2.6	datanode1.dalia.com10.0.2.42	datanode2.dalia.com
> /etc/hostnamenamenode.dalia.com
> And I am always receiving this error:
> INFO org.apache.hadoop.ipc.Client: Retrying connect to server: namenode/10.0.2.3:8020. Already tried 0 time(s).
> Note that I have already disabled the firewall and I opened the port : ufw allow 8020
> But when I run : telnet 10.0.2.3 8020 => connection refused ....
> So the problem is that I cannot open the port..... :(
> Note that I have tried it with other ports as 54310 and 9000 but same error occurs...
> > Date: Fri, 18 May 2012 01:52:48 +0530
> > Subject: Re: Important "Undefined Error"
> > From: dwivedishashwat@gmail.com
> > To: user@hbase.apache.org
> > 
> > Please send the content of /etc/hosts and /etc/hostname file
> > 
> > try this link
> > 
> > http://helpmetocode.blogspot.in/2012/05/hadoop-fully-distributed-cluster.html
> > 
> > 
> > for hadoop configuration
> > 
> > On Mon, May 14, 2012 at 10:15 PM, Dalia Sobhy <dalia.mohsobhy@hotmail.com>wrote:
> > 
> > > Yeasss
> > >
> > > Sent from my iPhone
> > >
> > > On 2012-05-14, at 5:28 PM, "N Keywal" <nkeywal@gmail.com> wrote:
> > >
> > > > In core-file.xml, do you have this?
> > > >
> > > > <configuration>
> > > > <property>
> > > > <name>fs.default.name</name>
> > > > <value>hdfs://namenode:8020/hbase</value>
> > > > </property>
> > > >
> > > > If you want hbase to connect to 8020 you must have hdfs listening on
> > > > 8020 as well.
> > > >
> > > >
> > > > On Mon, May 14, 2012 at 5:17 PM, Dalia Sobhy <dalia.mohsobhy@hotmail.com>
> > > wrote:
> > > >> Hiiii
> > > >>
> > > >> I have tried to make both ports the same.
> > > >> But the prob is the hbase cannot connect to port 8020.
> > > >> When i run nmap hostname, port 8020 wasnt with the list of open ports.
> > > >> I have tried what harsh told me abt.
> > > >> I used the same port he used but same error occurred.
> > > >> Another aspect in cloudera doc it says that i have to canonical name
> > > for the host ex: namenode.example.com as the hostname, but i didnt find
> > > it in any tutorial. No one makes it.
> > > >> Note that i am deploying my cluster in fully distributed mode i.e am
> > > using 4 machines..
> > > >>
> > > >> So any ideas??!!
> > > >>
> > > >> Sent from my iPhone
> > > >>
> > > >> On 2012-05-14, at 4:07 PM, "N Keywal" <nkeywal@gmail.com> wrote:
> > > >>
> > > >>> Hi,
> > > >>>
> > > >>> There could be multiple issues, but it's strange to have in
> > > hbase-site.xml
> > > >>>
> > > >>>  <value>hdfs://namenode:9000/hbase</value>
> > > >>>
> > > >>> while the core-site.xml says:
> > > >>>
> > > >>> <value>hdfs://namenode:54310/</value>
> > > >>>
> > > >>> The two entries should match.
> > > >>>
> > > >>> I would recommend to:
> > > >>> - use netstat to check the ports (netstat -l)
> > > >>> - do the check recommended by Harsh J previously.
> > > >>>
> > > >>> N.
> > > >>>
> > > >>>
> > > >>> On Mon, May 14, 2012 at 3:21 PM, Dalia Sobhy <
> > > dalia.mohsobhy@hotmail.com> wrote:
> > > >>>>
> > > >>>>
> > > >>>> pleaseeeeeeeeeeee helpppppppppppppppppppp
> > > >>>>
> > > >>>>> From: dalia.mohsobhy@hotmail.com
> > > >>>>> To: user@hbase.apache.org
> > > >>>>> Subject: RE: Important "Undefined Error"
> > > >>>>> Date: Mon, 14 May 2012 12:20:18 +0200
> > > >>>>>
> > > >>>>>
> > > >>>>>
> > > >>>>> Hi,
> > > >>>>> I tried what you told me, but nothing worked:(((
> > > >>>>> First when I run this command:dalia@namenode:~$ host -v -t A
> > > `hostname`Output:Trying "namenode"Host namenode not found:
> > > 3(NXDOMAIN)Received 101 bytes from 10.0.2.1#53 in 13 ms My
> > > core-site.xml:<configuration><property>        <name>fs.default.name</name>
> > > <!--<value>hdfs://namenode:8020</value>-->
> > >  <value>hdfs://namenode:54310/</value></property></configuration>
> > > >>>>> My
> > > hdfs-site.xml<configuration><property><name>dfs.name.dir</name><value>/data/1/dfs/nn,/nfsmount/dfs/nn</value></property><!--<property><name>dfs.data.dir</name><value>/data/1/dfs/dn,/data/2/dfs/dn,/data/3/dfs/dn</value></property>--><property><name>dfs.datanode.max.xcievers</name><value>4096</value></property><property><name>dfs.replication</name><value>3</value></property><property>
> > > <name>dfs.permissions.superusergroup</name> <value>hadoop</value></property>
> > > >>>>> My
> > > Mapred-site.xml<configuration><name>mapred.local.dir</name><value>/data/1/mapred/local,/data/2/mapred/local,/data/3/mapred/local</value></configuration>
> > > >>>>> My
> > > Hbase-site.xml<configuration><property><name>hbase.cluster.distributed</name>
> > >  <value>true</value></property><property>  <name>hbase.rootdir</name>
> > > <value>hdfs://namenode:9000/hbase</value></property><property><name>hbase.zookeeper.quorun</name>
> > > <value>namenode</value></property><property><name>hbase.regionserver.port</name><value>60020</value><description>The
> > > host and port that the HBase master runs
> > > at.</description></property><property><name>dfs.replication</name><value>1</value></property><property><name>hbase.zookeeper.property.clientPort</name><value>2181</value><description>Property
> > > from ZooKeeper's config zoo.cfg.The port at which the clients will
> > > connect.</description></property></configuration>
> > > >>>>> Please Help I am really disappointed I have been through all that
> > > for two weeks !!!!
> > > >>>>>
> > > >>>>>
> > > >>>>>
> > > >>>>>> From: dwivedishashwat@gmail.com
> > > >>>>>> To: user@hbase.apache.org
> > > >>>>>> Subject: RE: Important "Undefined Error"
> > > >>>>>> Date: Sat, 12 May 2012 23:31:49 +0530
> > > >>>>>>
> > > >>>>>> The problem is your hbase is not able to connect to Hadoop, can you
> > > put your
> > > >>>>>> hbase-site.xml content >> here.. have you specified localhost
> > > somewhere, if
> > > >>>>>> so remove localhost from everywhere and put your hdfsl namenode
> > > address
> > > >>>>>> suppose your namenode is running on master:9000 then put your hbase
> > > file
> > > >>>>>> system setting as master:9000/hbase here I am sending you the
> > > configuration
> > > >>>>>> which I am using in hbase and is working
> > > >>>>>>
> > > >>>>>>
> > > >>>>>> My hbase-site.xml content is
> > > >>>>>>
> > > >>>>>> <?xml version="1.0"?>
> > > >>>>>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> > > >>>>>> <!--
> > > >>>>>> /**
> > > >>>>>>  * Copyright 2010 The Apache Software Foundation
> > > >>>>>>  *
> > > >>>>>>  * Licensed to the Apache Software Foundation (ASF) under one
> > > >>>>>>  * or more contributor license agreements.  See the NOTICE file
> > > >>>>>>  * distributed with this work for additional information
> > > >>>>>>  * regarding copyright ownership.  The ASF licenses this file
> > > >>>>>>  * to you under the Apache License, Version 2.0 (the
> > > >>>>>>  * "License"); you may not use this file except in compliance
> > > >>>>>>  * with the License.  You may obtain a copy of the License at
> > > >>>>>>  *
> > > >>>>>>  *     http://www.apache.org/licenses/LICENSE-2.0
> > > >>>>>>  *
> > > >>>>>>  * Unless required by applicable law or agreed to in writing,
> > > software
> > > >>>>>>  * distributed under the License is distributed on an "AS IS" BASIS,
> > > >>>>>>  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
> > > implied.
> > > >>>>>>  * See the License for the specific language governing permissions
> > > and
> > > >>>>>>  * limitations under the License.
> > > >>>>>>  */
> > > >>>>>> -->
> > > >>>>>> <configuration>
> > > >>>>>> <property>
> > > >>>>>> <name>hbase.rootdir</name>
> > > >>>>>> <value>hdfs://master:9000/hbase</value>
> > > >>>>>> </property>
> > > >>>>>> <property>
> > > >>>>>> <name>hbase.master</name>
> > > >>>>>> <value>master:60000</value>
> > > >>>>>> <description>The host and port that the HBase master runs
> > > at.</description>
> > > >>>>>> </property>
> > > >>>>>> <property>
> > > >>>>>> <name>hbase.regionserver.port</name>
> > > >>>>>> <value>60020</value>
> > > >>>>>> <description>The host and port that the HBase master runs
> > > at.</description>
> > > >>>>>> </property>
> > > >>>>>> <!--<property>
> > > >>>>>> <name>hbase.master.port</name>
> > > >>>>>> <value>60000</value>
> > > >>>>>> <description>The host and port that the HBase master runs
> > > at.</description>
> > > >>>>>> </property>-->
> > > >>>>>> <property>
> > > >>>>>> <name>hbase.cluster.distributed</name>
> > > >>>>>> <value>true</value>
> > > >>>>>> </property>
> > > >>>>>> <property>
> > > >>>>>> <name>hbase.tmp.dir</name>
> > > >>>>>> <value>/home/shashwat/Hadoop/hbase-0.90.4/temp</value>
> > > >>>>>> </property>
> > > >>>>>> <property>
> > > >>>>>> <name>hbase.zookeeper.quorum</name>
> > > >>>>>> <value>master</value>
> > > >>>>>> </property>
> > > >>>>>> <property>
> > > >>>>>> <name>dfs.replication</name>
> > > >>>>>> <value>1</value>
> > > >>>>>> </property>
> > > >>>>>> <property>
> > > >>>>>> <name>hbase.zookeeper.property.clientPort</name>
> > > >>>>>> <value>2181</value>
> > > >>>>>> <description>Property from ZooKeeper's config zoo.cfg.
> > > >>>>>> The port at which the clients will connect.
> > > >>>>>> </description>
> > > >>>>>> </property>
> > > >>>>>> <property>
> > > >>>>>> <name>hbase.zookeeper.property.dataDir</name>
> > > >>>>>> <value>/home/shashwat/zookeeper</value>
> > > >>>>>> <description>Property from ZooKeeper's config zoo.cfg.
> > > >>>>>> The directory where the snapshot is stored.
> > > >>>>>> </description>
> > > >>>>>> </property>
> > > >>>>>>
> > > >>>>>> </configuration>
> > > >>>>>>
> > > >>>>>>
> > > >>>>>>
> > > >>>>>>
> > > >>>>>> Check this out, and also stop hbase, If its not stopping kill all
> > > the
> > > >>>>>> processes, and after putting your  hdfs-site.xml, mapred-site.xml
> > > and
> > > >>>>>> core-site.sml to hbase conf directory try to restart, and also
> > > delete the
> > > >>>>>> folders created by hbase ,,, like temp directory or other then try
> > > to start.
> > > >>>>>>
> > > >>>>>> Regards
> > > >>>>>> ∞
> > > >>>>>> Shashwat Shriparv
> > > >>>>>>
> > > >>>>>>
> > > >>>>>> -----Original Message-----
> > > >>>>>> From: Dalia Sobhy [mailto:dalia.mohsobhy@hotmail.com]
> > > >>>>>> Sent: 12 May 2012 22:48
> > > >>>>>> To: user@hbase.apache.org
> > > >>>>>> Subject: RE: Important "Undefined Error"
> > > >>>>>>
> > > >>>>>>
> > > >>>>>> Hi Shashwat,
> > > >>>>>> I want to tell you about my configurations:
> > > >>>>>> I am using 4 nodesOne "Master": Namenode, SecondaryNamenode, Job
> > > Tracker,
> > > >>>>>> Zookeeper, HMasterThree "Slaves": datanodes, tasktrackers,
> > > regionservers In
> > > >>>>>> both master and slaves, all the hadoop daemons are working well,
> > > but as for
> > > >>>>>> the hbase master service it is not working..
> > > >>>>>> As for region server here is the error:12/05/12 14:42:13 INFO
> > > >>>>>> util.ServerCommandLine: vmName=Java HotSpot(TM) 64-Bit Server VM,
> > > >>>>>> vmVendor=Sun Microsystems Inc., vmVersion=20.1-b0212/05/12 14:42:13
> > > INFO
> > > >>>>>> util.ServerCommandLine: vmInputArguments=[-Xmx1000m, -ea,
> > > >>>>>> -XX:+UseConcMarkSweepGC, -XX:+CMSIncrementalMode,
> > > >>>>>> -Dhbase.log.dir=/usr/lib/hbase/bin/../logs,
> > > -Dhbase.log.file=hbase.log,
> > > >>>>>> -Dhbase.home.dir=/usr/lib/hbase/bin/.., -Dhbase.id.str=,
> > > >>>>>> -Dhbase.root.logger=INFO,console,
> > > >>>>>>
> > > -Djava.library.path=/usr/lib/hadoop-0.20/lib/native/Linux-amd64-64:/usr/lib/
> > > >>>>>> hbase/bin/../lib/native/Linux-amd64-64]12/05/12 14:42:13 INFO
> > > >>>>>> ipc.HBaseRpcMetrics: Initializing RPC Metrics with
> > > hostName=HRegionServer,
> > > >>>>>> port=6002012/05/12 14:42:14 FATAL zookeeper.ZKConfig: The server in
> > > zoo.cfg
> > > >>>>>> cannot be set to localhost in a fully-distributed setup because it
> > > won't be
> > > >>>>>> reachable. See "Getting Started" for more information.12/05/12
> > > 14:42:14 WARN
> > > >>>>>> zookeeper.ZKConfig: Cannot read zoo.cfg, loading from XML
> > > >>>>>> filesjava.io.IOException:
> > > >>>>>>  The server in zoo.cfg cannot be set to localhost in a
> > > fully-distributed
> > > >>>>>> setup because it won't be reachable. See "Getting Started" for more
> > > >>>>>> information.        at
> > > >>>>>>
> > > org.apache.hadoop.hbase.zookeeper.ZKConfig.parseZooCfg(ZKConfig.java:172)
> > > >>>>>> at
> > > org.apache.hadoop.hbase.zookeeper.ZKConfig.makeZKProps(ZKConfig.java:68)
> > > >>>>>> at
> > > >>>>>>
> > > org.apache.hadoop.hbase.zookeeper.ZKConfig.getZKQuorumServersString(ZKConfig
> > > >>>>>> .java:249)  at
> > > >>>>>>
> > > org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher.<init>(ZooKeeperWatcher.j
> > > >>>>>> ava:117)    at
> > > >>>>>>
> > > org.apache.hadoop.hbase.regionserver.HRegionServer.initializeZooKeeper(HRegi
> > > >>>>>> onServer.java:489)  at
> > > >>>>>>
> > > org.apache.hadoop.hbase.regionserver.HRegionServer.initialize(HRegionServer.
> > > >>>>>> java:465)   at
> > > >>>>>>
> > > org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:56
> > > >>>>>> 4)  at java.lang.Thread.run(Thread.java:662)12/05/12 14:42:14 INFO
> > > >>>>>> zookeeper.ZooKeeper: Client
> > > environment:zookeeper.version=3.3.5-cdh3u4--1,
> > > >>>>>> built on 05/07/2012 21:12 GMT12/05/12 14:42:14 INFO
> > > zookeeper.ZooKeeper:
> > > >>>>>> Client environment:host.name=data
> > > >>>>>>  node212/05/12 14:42:14 INFO zookeeper.ZooKeeper: Client
> > > >>>>>> environment:java.version=1.6.0_2612/05/12 14:42:14 INFO
> > > zookeeper.ZooKeeper:
> > > >>>>>> Client environment:java.vendor=Sun Microsystems Inc.12/05/12
> > > 14:42:14 INFO
> > > >>>>>> zookeeper.ZooKeeper: Client
> > > environment:java.home=/usr/lib/jvm/java-6-sun-1.
> > > >>>>>> 6.0.26/jre12/05/12 14:42:14 INFO zookeeper.ZooKeeper: Client
> > > >>>>>>
> > > environment:java.class.path=/usr/lib/hbase/bin/../conf:/usr/lib/jvm/java-6-s
> > > >>>>>>
> > > un/lib/tools.jar:/usr/lib/hbase/bin/..:/usr/lib/hbase/bin/../hbase-0.90.6-cd
> > > >>>>>>
> > > h3u4.jar:/usr/lib/hbase/bin/../hbase-0.90.6-cdh3u4-tests.jar:/usr/lib/hbase/
> > > >>>>>>
> > > bin/../lib/activation-1.1.jar:/usr/lib/hbase/bin/../lib/asm-3.1.jar:/usr/lib
> > > >>>>>>
> > > /hbase/bin/../lib/avro-1.5.4.jar:/usr/lib/hbase/bin/../lib/avro-ipc-1.5.4.ja
> > > >>>>>>
> > > r:/usr/lib/hbase/bin/../lib/commons-cli-1.2.jar:/usr/lib/hbase/bin/../lib/co
> > > >>>>>>
> > > mmons-codec-1.4.jar:/usr/lib/hbase/bin/../lib/commons-el-1.0.jar:/usr/lib/hb
> > > >>>>>>
> > > ase/bin/../lib/commons-httpclient-3.1.jar:/usr/lib/hbase/bin/../lib/commons-
> > > >>>>>> lang-2.5.jar:/usr/lib/hbase/bin/../lib/commo
> > > >>>>>>
> > > >>>>>>
> > > ns-logging-1.1.1.jar:/usr/lib/hbase/bin/../lib/commons-net-1.4.1.jar:/usr/li
> > > >>>>>>
> > > b/hbase/bin/../lib/core-3.1.1.jar:/usr/lib/hbase/bin/../lib/guava-r06.jar:/u
> > > >>>>>>
> > > sr/lib/hbase/bin/../lib/guava-r09-jarjar.jar:/usr/lib/hbase/bin/../lib/hadoo
> > > >>>>>>
> > > p-core.jar:/usr/lib/hbase/bin/../lib/jackson-core-asl-1.5.2.jar:/usr/lib/hba
> > > >>>>>>
> > > se/bin/../lib/jackson-jaxrs-1.5.5.jar:/usr/lib/hbase/bin/../lib/jackson-mapp
> > > >>>>>>
> > > er-asl-1.5.2.jar:/usr/lib/hbase/bin/../lib/jackson-xc-1.5.5.jar:/usr/lib/hba
> > > >>>>>>
> > > se/bin/../lib/jamon-runtime-2.3.1.jar:/usr/lib/hbase/bin/../lib/jasper-compi
> > > >>>>>>
> > > ler-5.5.23.jar:/usr/lib/hbase/bin/../lib/jasper-runtime-5.5.23.jar:/usr/lib/
> > > >>>>>>
> > > hbase/bin/../lib/jaxb-api-2.1.jar:/usr/lib/hbase/bin/../lib/jaxb-impl-2.1.12
> > > >>>>>>
> > > .jar:/usr/lib/hbase/bin/../lib/jersey-core-1.4.jar:/usr/lib/hbase/bin/../lib
> > > >>>>>>
> > > /jersey-json-1.4.jar:/usr/lib/hbase/bin/../lib/jersey-server-1.4.jar:/usr/li
> > > >>>>>>
> > > b/hbase/bin/../lib/jettison-1.1.jar:/usr/lib/hbase/bin/../lib/jetty-6.1.26.j
> > > >>>>>>
> > > ar:/usr/lib/hbase/bin/../lib/jetty-util-6.1.26.jar:/usr/lib/hbase/bin/../lib
> > > >>>>>> /jruby-co
> > > >>>>>>
> > > >>>>>>
> > > mplete-1.6.0.jar:/usr/lib/hbase/bin/../lib/jsp-2.1-6.1.14.jar:/usr/lib/hbase
> > > >>>>>>
> > > /bin/../lib/jsp-api-2.1-6.1.14.jar:/usr/lib/hbase/bin/../lib/jsp-api-2.1.jar
> > > >>>>>>
> > > :/usr/lib/hbase/bin/../lib/jsr311-api-1.1.1.jar:/usr/lib/hbase/bin/../lib/lo
> > > >>>>>>
> > > g4j-1.2.16.jar:/usr/lib/hbase/bin/../lib/netty-3.2.4.Final.jar:/usr/lib/hbas
> > > >>>>>>
> > > e/bin/../lib/protobuf-java-2.3.0.jar:/usr/lib/hbase/bin/../lib/servlet-api-2
> > > >>>>>>
> > > .5-6.1.14.jar:/usr/lib/hbase/bin/../lib/servlet-api-2.5.jar:/usr/lib/hbase/b
> > > >>>>>>
> > > in/../lib/slf4j-api-1.5.8.jar:/usr/lib/hbase/bin/../lib/slf4j-log4j12-1.5.8.
> > > >>>>>>
> > > jar:/usr/lib/hbase/bin/../lib/snappy-java-1.0.3.2.jar:/usr/lib/hbase/bin/../
> > > >>>>>>
> > > lib/stax-api-1.0.1.jar:/usr/lib/hbase/bin/../lib/thrift-0.2.0.jar:/usr/lib/h
> > > >>>>>>
> > > base/bin/../lib/velocity-1.5.jar:/usr/lib/hbase/bin/../lib/xmlenc-0.52.jar:/
> > > >>>>>>
> > > usr/lib/hbase/bin/../lib/zookeeper.jar:/etc/zookeeper:/etc/hadoop-0.20/conf:
> > > >>>>>>
> > > /usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u4-ant.jar:/usr/lib/hadoop-0.20/hadoo
> > > >>>>>>
> > > p-0.20.2-cdh3u4-tools.jar:/usr/lib/hadoop-0.20/hadoop-tools.jar:/usr/lib/had
> > > >>>>>> oop-0.20/
> > > >>>>>>
> > > >>>>>>
> > > hadoop-examples-0.20.2-cdh3u4.jar:/usr/lib/hadoop-0.20/hadoop-ant-0.20.2-cdh
> > > >>>>>>
> > > 3u4.jar:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u4-examples.jar:/usr/lib/hado
> > > >>>>>>
> > > op-0.20/hadoop-ant.jar:/usr/lib/hadoop-0.20/hadoop-core-0.20.2-cdh3u4.jar:/u
> > > >>>>>>
> > > sr/lib/hadoop-0.20/hadoop-core.jar:/usr/lib/hadoop-0.20/hadoop-tools-0.20.2-
> > > >>>>>>
> > > cdh3u4.jar:/usr/lib/hadoop-0.20/hadoop-examples.jar:/usr/lib/hadoop-0.20/had
> > > >>>>>>
> > > oop-0.20.2-cdh3u4-core.jar:/usr/lib/hadoop-0.20/hadoop-test-0.20.2-cdh3u4.ja
> > > >>>>>>
> > > r:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u4-test.jar:/usr/lib/hadoop-0.20/ha
> > > >>>>>>
> > > doop-test.jar:/usr/lib/hadoop-0.20/lib/jetty-6.1.26.cloudera.1.jar:/usr/lib/
> > > >>>>>>
> > > hadoop-0.20/lib/jets3t-0.6.1.jar:/usr/lib/hadoop-0.20/lib/junit-4.5.jar:/usr
> > > >>>>>>
> > > /lib/hadoop-0.20/lib/commons-daemon-1.0.1.jar:/usr/lib/hadoop-0.20/lib/slf4j
> > > >>>>>>
> > > -api-1.4.3.jar:/usr/lib/hadoop-0.20/lib/commons-codec-1.4.jar:/usr/lib/hadoo
> > > >>>>>>
> > > p-0.20/lib/log4j-1.2.15.jar:/usr/lib/hadoop-0.20/lib/jasper-compiler-5.5.12.
> > > >>>>>>
> > > jar:/usr/lib/hadoop-0.20/lib/guava-r09-jarjar.jar:/usr/lib/hadoop-0.20/lib/j
> > > >>>>>> ackson-ma
> > > >>>>>>
> > > >>>>>>
> > > pper-asl-1.5.2.jar:/usr/lib/hadoop-0.20/lib/commons-logging-1.0.4.jar:/usr/l
> > > >>>>>>
> > > ib/hadoop-0.20/lib/ant-contrib-1.0b3.jar:/usr/lib/hadoop-0.20/lib/commons-la
> > > >>>>>>
> > > ng-2.4.jar:/usr/lib/hadoop-0.20/lib/kfs-0.2.2.jar:/usr/lib/hadoop-0.20/lib/j
> > > >>>>>>
> > > etty-util-6.1.26.cloudera.1.jar:/usr/lib/hadoop-0.20/lib/commons-httpclient-
> > > >>>>>>
> > > 3.1.jar:/usr/lib/hadoop-0.20/lib/jsch-0.1.42.jar:/usr/lib/hadoop-0.20/lib/sl
> > > >>>>>>
> > > f4j-log4j12-1.4.3.jar:/usr/lib/hadoop-0.20/lib/servlet-api-2.5-20081211.jar:
> > > >>>>>>
> > > /usr/lib/hadoop-0.20/lib/jasper-runtime-5.5.12.jar:/usr/lib/hadoop-0.20/lib/
> > > >>>>>>
> > > servlet-api-2.5-6.1.14.jar:/usr/lib/hadoop-0.20/lib/jetty-servlet-tester-6.1
> > > >>>>>>
> > > .26.cloudera.1.jar:/usr/lib/hadoop-0.20/lib/aspectjtools-1.6.5.jar:/usr/lib/
> > > >>>>>>
> > > hadoop-0.20/lib/xmlenc-0.52.jar:/usr/lib/hadoop-0.20/lib/hadoop-fairschedule
> > > >>>>>>
> > > r-0.20.2-cdh3u4.jar:/usr/lib/hadoop-0.20/lib/jackson-core-asl-1.5.2.jar:/usr
> > > >>>>>>
> > > /lib/hadoop-0.20/lib/mockito-all-1.8.2.jar:/usr/lib/hadoop-0.20/lib/commons-
> > > >>>>>>
> > > el-1.0.jar:/usr/lib/hadoop-0.20/lib/commons-logging-api-1.0.4.jar:/usr/lib/h
> > > >>>>>> adoop-0.2
> > > >>>>>>
> > > >>>>>>
> > > 0/lib/commons-net-3.1.jar:/usr/lib/hadoop-0.20/lib/commons-cli-1.2.jar:/usr/
> > > >>>>>>
> > > lib/hadoop-0.20/lib/aspectjrt-1.6.5.jar:/usr/lib/hadoop-0.20/lib/core-3.1.1.
> > > >>>>>>
> > > jar:/usr/lib/hadoop-0.20/lib/hsqldb-1.8.0.10.jar:/usr/lib/hadoop-0.20/lib/or
> > > >>>>>>
> > > o-2.0.8.jar:/usr/lib/zookeeper/zookeeper.jar:/usr/lib/zookeeper/zookeeper-3.
> > > >>>>>>
> > > 3.5-cdh3u4.jar:/usr/lib/zookeeper/lib/log4j-1.2.15.jar:/usr/lib/zookeeper/li
> > > >>>>>>
> > > b/jline-0.9.94.jar::/usr/lib/hadoop-0.20/conf:/usr/lib/hadoop-0.20/hadoop-co
> > > >>>>>>
> > > re-0.20.2-cdh3u4.jar:/usr/lib/hadoop-0.20/lib/ant-contrib-1.0b3.jar:/usr/lib
> > > >>>>>>
> > > /hadoop-0.20/lib/aspectjrt-1.6.5.jar:/usr/lib/hadoop-0.20/lib/aspectjtools-1
> > > >>>>>>
> > > .6.5.jar:/usr/lib/hadoop-0.20/lib/commons-cli-1.2.jar:/usr/lib/hadoop-0.20/l
> > > >>>>>>
> > > ib/commons-codec-1.4.jar:/usr/lib/hadoop-0.20/lib/commons-daemon-1.0.1.jar:/
> > > >>>>>>
> > > usr/lib/hadoop-0.20/lib/commons-el-1.0.jar:/usr/lib/hadoop-0.20/lib/commons-
> > > >>>>>>
> > > httpclient-3.1.jar:/usr/lib/hadoop-0.20/lib/commons-lang-2.4.jar:/usr/lib/ha
> > > >>>>>>
> > > doop-0.20/lib/commons-logging-1.0.4.jar:/usr/lib/hadoop-0.20/lib/commons-log
> > > >>>>>> ging-api-
> > > >>>>>>
> > > >>>>>>
> > > 1.0.4.jar:/usr/lib/hadoop-0.20/lib/commons-net-3.1.jar:/usr/lib/hadoop-0.20/
> > > >>>>>>
> > > lib/core-3.1.1.jar:/usr/lib/hadoop-0.20/lib/guava-r09-jarjar.jar:/usr/lib/ha
> > > >>>>>>
> > > doop-0.20/lib/hadoop-fairscheduler-0.20.2-cdh3u4.jar:/usr/lib/hadoop-0.20/li
> > > >>>>>>
> > > b/hsqldb-1.8.0.10.jar:/usr/lib/hadoop-0.20/lib/jackson-core-asl-1.5.2.jar:/u
> > > >>>>>>
> > > sr/lib/hadoop-0.20/lib/jackson-mapper-asl-1.5.2.jar:/usr/lib/hadoop-0.20/lib
> > > >>>>>>
> > > /jasper-compiler-5.5.12.jar:/usr/lib/hadoop-0.20/lib/jasper-runtime-5.5.12.j
> > > >>>>>>
> > > ar:/usr/lib/hadoop-0.20/lib/jets3t-0.6.1.jar:/usr/lib/hadoop-0.20/lib/jetty-
> > > >>>>>>
> > > 6.1.26.cloudera.1.jar:/usr/lib/hadoop-0.20/lib/jetty-servlet-tester-6.1.26.c
> > > >>>>>>
> > > loudera.1.jar:/usr/lib/hadoop-0.20/lib/jetty-util-6.1.26.cloudera.1.jar:/usr
> > > >>>>>>
> > > /lib/hadoop-0.20/lib/jsch-0.1.42.jar:/usr/lib/hadoop-0.20/lib/junit-4.5.jar:
> > > >>>>>>
> > > /usr/lib/hadoop-0.20/lib/kfs-0.2.2.jar:/usr/lib/hadoop-0.20/lib/log4j-1.2.15
> > > >>>>>>
> > > .jar:/usr/lib/hadoop-0.20/lib/mockito-all-1.8.2.jar:/usr/lib/hadoop-0.20/lib
> > > >>>>>>
> > > /oro-2.0.8.jar:/usr/lib/hadoop-0.20/lib/servlet-api-2.5-20081211.jar:/usr/li
> > > >>>>>> b/hadoop-
> > > >>>>>>
> > >  0.20/lib/servlet-api-2.5-6.1.14.jar:/usr/lib/hadoop-0.20/lib/slf4j-api-1.4.
> > > >>>>>>
> > > 3.jar:/usr/lib/hadoop-0.20/lib/slf4j-log4j12-1.4.3.jar:/usr/lib/hadoop-0.20/
> > > >>>>>> lib/xmlenc-0.52.jar12/05/12 14:42:14 INFO zookeeper.ZooKeeper:
> > > Client
> > > >>>>>>
> > > environment:java.library.path=/usr/lib/hadoop-0.20/lib/native/Linux-amd64-64
> > > >>>>>> :/usr/lib/hbase/bin/../lib/native/Linux-amd64-6412/05/12 14:42:14
> > > INFO
> > > >>>>>> zookeeper.ZooKeeper: Client environment:java.io.tmpdir=/tmp12/05/12
> > > 14:42:14
> > > >>>>>> INFO zookeeper.ZooKeeper: Client
> > > environment:java.compiler=<NA>12/05/12
> > > >>>>>> 14:42:14 INFO zookeeper.ZooKeeper: Client environment:os.name
> > > =Linux12/05/12
> > > >>>>>> 14:42:14 INFO zookeeper.ZooKeeper: Client
> > > environment:os.arch=amd6412/05/12
> > > >>>>>> 14:42:14 INFO zookeeper.ZooKeeper: Client
> > > >>>>>> environment:os.version=2.6.35-22-server12/05/12 14:42:14 INFO
> > > >>>>>> zookeeper.ZooKeeper: Client environment:user.name=dalia12/05/12
> > > 14:42:14
> > > >>>>>> INFO zookeeper.ZooKeeper: Client
> > > environment:user.home=/home/dalia12/05/12
> > > >>>>>> 14:42:14 INFO zookeeper.ZooKeeper: Client
> > > environment:user.dir=/home/dalia12
> > > >>>>>>  /05/12 14:42:14 INFO zookeeper.ZooKeeper: Initiating client
> > > connection,
> > > >>>>>> connectString=localhost:2181 sessionTimeout=180000
> > > >>>>>> watcher=regionserver:6002012/05/12 14:42:14 INFO
> > > zookeeper.ClientCnxn:
> > > >>>>>> Opening socket connection to server
> > > localhost/0:0:0:0:0:0:0:1:218112/05/12
> > > >>>>>> 14:42:14 WARN zookeeper.ClientCnxn: Session 0x0 for server null,
> > > unexpected
> > > >>>>>> error, closing socket connection and attempting
> > > >>>>>> reconnectjava.net.ConnectException: Connection refused      at
> > > >>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)    at
> > > >>>>>>
> > > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
> > >  at
> > > >>>>>>
> > > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
> > > >>>>>> 14:42:14 INFO zookeeper.ClientCnxn: Opening socket connection to
> > > server
> > > >>>>>> localhost/127.0.0.1:218112/05/12 14:42:14 WARN
> > > zookeeper.ClientCnxn: Session
> > > >>>>>> 0x0 for server null, unexpected error, closing socket connection and
> > > >>>>>> attempting reconnectjava.net.ConnectException: Connection refused
> > > at
> > > >>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native
> > > >>>>>>   Method)   at
> > > >>>>>>
> > > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
> > >  at
> > > >>>>>>
> > > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
> > > >>>>>> 14:42:15 INFO ipc.Client: Retrying connect to server:
> > > >>>>>> namenode/10.0.2.3:8020. Already tried 0 time(s).12/05/12 14:42:16
> > > INFO
> > > >>>>>> zookeeper.ClientCnxn: Opening socket connection to server
> > > localhost/0:0:0:0:
> > > >>>>>> 0:0:0:1:218112/05/12 14:42:16 WARN zookeeper.ClientCnxn: Session
> > > 0x0 for
> > > >>>>>> server null, unexpected error, closing socket connection and
> > > attempting
> > > >>>>>> reconnectjava.net.ConnectException: Connection refused      at
> > > >>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)    at
> > > >>>>>>
> > > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
> > >  at
> > > >>>>>>
> > > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
> > > >>>>>> 14:42:16 INFO ipc.Client: Retrying connect to server:
> > > >>>>>> namenode/10.0.2.3:8020. Already tried 1 time(s).12/05/12 14:42:16
> > > INFO
> > > >>>>>> zookeeper.ClientCnxn: Opening socket connection to server localhost/
> > > 127.0.0.
> > > >>>>>> 1:218112/05/12 14:
> > > >>>>>>  42:16 WARN zookeeper.ClientCnxn: Session 0x0 for server null,
> > > unexpected
> > > >>>>>> error, closing socket connection and attempting
> > > >>>>>> reconnectjava.net.ConnectException: Connection refused      at
> > > >>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)    at
> > > >>>>>>
> > > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
> > >  at
> > > >>>>>>
> > > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
> > > >>>>>> 14:42:17 INFO ipc.Client: Retrying connect to server:
> > > >>>>>> namenode/10.0.2.3:8020. Already tried 2 time(s).12/05/12 14:42:18
> > > INFO
> > > >>>>>> ipc.Client: Retrying connect to server: namenode/10.0.2.3:8020.
> > > Already
> > > >>>>>> tried 3 time(s).12/05/12 14:42:18 INFO zookeeper.ClientCnxn:
> > > Opening socket
> > > >>>>>> connection to server localhost/0:0:0:0:0:0:0:1:218112/05/12
> > > 14:42:18 WARN
> > > >>>>>> zookeeper.ClientCnxn: Session 0x0 for server null, unexpected
> > > error, closing
> > > >>>>>> socket connection and attempting reconnectjava.net.ConnectException:
> > > >>>>>> Connection refused  at
> > > sun.nio.ch.SocketChannelImpl.checkConnect(Native
> > > >>>>>> Method)     at sun.nio.ch.SocketChannelImpl.fin
> > > >>>>>>  ishConnect(SocketChannelImpl.java:567)     at
> > > >>>>>>
> > > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
> > > >>>>>> 14:42:18 INFO zookeeper.ClientCnxn: Opening socket connection to
> > > server
> > > >>>>>> localhost/127.0.0.1:218112/05/12 14:42:18 WARN
> > > zookeeper.ClientCnxn: Session
> > > >>>>>> 0x0 for server null, unexpected error, closing socket connection and
> > > >>>>>> attempting reconnectjava.net.ConnectException: Connection refused
> > > at
> > > >>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)    at
> > > >>>>>>
> > > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
> > >  at
> > > >>>>>>
> > > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
> > > >>>>>> 14:42:19 INFO ipc.Client: Retrying connect to server:
> > > >>>>>> namenode/10.0.2.3:8020. Already tried 4 time(s).12/05/12 14:42:19
> > > INFO
> > > >>>>>> zookeeper.ClientCnxn: Opening socket connection to server
> > > localhost/0:0:0:0:
> > > >>>>>> 0:0:0:1:218112/05/12 14:42:19 WARN zookeeper.ClientCnxn: Session
> > > 0x0 for
> > > >>>>>> server null, unexpected error, closing socket connection and
> > > attempting
> > > >>>>>> reconnectjava.net.ConnectException
> > > >>>>>>  : Connection refused       at
> > > sun.nio.ch.SocketChannelImpl.checkConnect(Native
> > > >>>>>> Method)     at
> > > >>>>>>
> > > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
> > >  at
> > > >>>>>>
> > > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
> > > >>>>>> 14:42:20 INFO ipc.Client: Retrying connect to server:
> > > >>>>>> namenode/10.0.2.3:8020. Already tried 5 time(s).12/05/12 14:42:20
> > > INFO
> > > >>>>>> zookeeper.ClientCnxn: Opening socket connection to server localhost/
> > > 127.0.0.
> > > >>>>>> 1:218112/05/12 14:42:20 WARN zookeeper.ClientCnxn: Session 0x0 for
> > > server
> > > >>>>>> null, unexpected error, closing socket connection and attempting
> > > >>>>>> reconnectjava.net.ConnectException: Connection refused      at
> > > >>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)    at
> > > >>>>>>
> > > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
> > >  at
> > > >>>>>>
> > > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
> > > >>>>>> 14:42:21 INFO ipc.Client: Retrying connect to server:
> > > >>>>>> namenode/10.0.2.3:8020. Already tried 6 time(s).12/05/12 14:42:22
> > > INFO
> > > >>>>>> zookeeper.ClientCnxn: Openin
> > > >>>>>>  g socket connection to server
> > > localhost/0:0:0:0:0:0:0:1:218112/05/12 14:42:
> > > >>>>>> 22 WARN zookeeper.ClientCnxn: Session 0x0 for server null,
> > > unexpected error,
> > > >>>>>> closing socket connection and attempting
> > > reconnectjava.net.ConnectException:
> > > >>>>>> Connection refused  at
> > > sun.nio.ch.SocketChannelImpl.checkConnect(Native
> > > >>>>>> Method)     at
> > > >>>>>>
> > > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
> > >  at
> > > >>>>>>
> > > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
> > > >>>>>> 14:42:22 INFO ipc.Client: Retrying connect to server:
> > > >>>>>> namenode/10.0.2.3:8020. Already tried 7 time(s).12/05/12 14:42:22
> > > INFO
> > > >>>>>> zookeeper.ClientCnxn: Opening socket connection to server localhost/
> > > 127.0.0.
> > > >>>>>> 1:218112/05/12 14:42:22 WARN zookeeper.ClientCnxn: Session 0x0 for
> > > server
> > > >>>>>> null, unexpected error, closing socket connection and attempting
> > > >>>>>> reconnectjava.net.ConnectException: Connection refused      at
> > > >>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)    at
> > > >>>>>>
> > > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
> > >  at
> > > >>>>>> org
> > > >>>>>>
> > >  .apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
> > > >>>>>> 14:42:23 INFO ipc.Client: Retrying connect to server:
> > > >>>>>> namenode/10.0.2.3:8020. Already tried 8 time(s).12/05/12 14:42:24
> > > INFO
> > > >>>>>> ipc.Client: Retrying connect to server: namenode/10.0.2.3:8020.
> > > Already
> > > >>>>>> tried 9 time(s).Exception in thread "main"
> > > java.net.ConnectException: Call
> > > >>>>>> to namenode/10.0.2.3:8020 failed on connection exception:
> > > >>>>>> java.net.ConnectException: Connection refused       at
> > > >>>>>> org.apache.hadoop.ipc.Client.wrapException(Client.java:1134)
> > >  at
> > > >>>>>> org.apache.hadoop.ipc.Client.call(Client.java:1110) at
> > > >>>>>> org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:226)      at
> > > >>>>>> $Proxy5.getProtocolVersion(Unknown Source)  at
> > > >>>>>> org.apache.hadoop.ipc.RPC.getProxy(RPC.java:398)    at
> > > >>>>>> org.apache.hadoop.ipc.RPC.getProxy(RPC.java:384)    at
> > > >>>>>>
> > > org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:129)
> > >  at
> > > >>>>>> org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:255) at
> > > >>>>>> org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:217) at
> > > >>>>>> org.apache.hadoop
> > > >>>>>>
> > >  .hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89)
> > >  at
> > > >>>>>>
> > > org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1563)
> > >  at
> > > >>>>>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:67)
> > >  at
> > > >>>>>>
> > > org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:1597)
> > > at
> > > >>>>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1579)
> > > at
> > > >>>>>> org.apache.hadoop.fs.FileSystem.get(FileSystem.java:228)    at
> > > >>>>>> org.apache.hadoop.fs.FileSystem.get(FileSystem.java:111)    at
> > > >>>>>>
> > > org.apache.hadoop.hbase.regionserver.HRegionServer.startRegionServer(HRegion
> > > >>>>>> Server.java:2785)   at
> > > >>>>>>
> > > org.apache.hadoop.hbase.regionserver.HRegionServer.startRegionServer(HRegion
> > > >>>>>> Server.java:2768)   at
> > > >>>>>>
> > > org.apache.hadoop.hbase.regionserver.HRegionServerCommandLine.start(HRegionS
> > > >>>>>> erverCommandLine.java:61)   at
> > > >>>>>>
> > > org.apache.hadoop.hbase.regionserver.HRegionServerCommandLine.run(HRegionSer
> > > >>>>>> verCommandLine.java:75)     at
> > > org.apache.hadoop.util.ToolRunner.run(ToolRunner.
> > > >>>>>> java:65)    at
> > > >>>>>> org.apache.hadoop.hbase.util.ServerCommandLine.doMain(Serve
> > > >>>>>>  rCommandLine.java:76)      at
> > > >>>>>>
> > > org.apache.hadoop.hbase.regionserver.HRegionServer.main(HRegionServer.java:2
> > > >>>>>> 829)Caused by: java.net.ConnectException: Connection refused
> > >  at
> > > >>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)    at
> > > >>>>>>
> > > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
> > >  at
> > > >>>>>>
> > > org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:2
> > > >>>>>> 06) at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:429)
> > >  at
> > > >>>>>> org.apache.hadoop.net.NetUtils.connect(NetUtils.java:394)   at
> > > >>>>>>
> > > org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:425)
> > > >>>>>> at
> > > org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:532)
> > > >>>>>> at
> > > org.apache.hadoop.ipc.Client$Connection.access$2300(Client.java:210)     at
> > > >>>>>> org.apache.hadoop.ipc.Client.getConnection(Client.java:1247)
> > >  at
> > > >>>>>> org.apache.hadoop.ipc.Client.call(Client.java:1078) ... 21
> > > more12/05/12
> > > >>>>>> 14:42:24 INFO zookeeper.ClientCnxn: Opening socket connection to
> > > server
> > > >>>>>> localhost/0:0:0:0:0:0:0:1:218112/05/12 14:42:24 WARN zookeeper
> > > >>>>>>  .ClientCnxn: Session 0x0 for server null, unexpected error,
> > > closing socket
> > > >>>>>> connection and attempting reconnectjava.net.ConnectException:
> > > Connection
> > > >>>>>> refused     at sun.nio.ch.SocketChannelImpl.checkConnect(Native
> > > Method)     at
> > > >>>>>>
> > > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
> > >  at
> > > >>>>>>
> > > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
> > > >>>>>> 14:42:25 INFO zookeeper.ClientCnxn: Opening socket connection to
> > > server
> > > >>>>>> localhost/127.0.0.1:218112/05/12 14:42:25 INFO
> > > zookeeper.ZooKeeper: Session:
> > > >>>>>> 0x0 closed12/05/12 14:42:25 INFO zookeeper.ClientCnxn: EventThread
> > > shut
> > > >>>>>> down12/05/12 14:42:25 INFO ipc.HBaseServer: Stopping server on
> > > 6002012/05/12
> > > >>>>>> 14:42:25 FATAL regionserver.HRegionServer: ABORTING region server
> > > >>>>>> serverName=datanode2,60020,1336826533870, load=(requests=0,
> > > regions=0,
> > > >>>>>> usedHeap=0, maxHeap=0): Initialization of RS failed.  Hence
> > > aborting RS.org.
> > > >>>>>> apache.hadoop.hbase.ZooKeeperConnectionException: HBase is able to
> > > connect
> > > >>>>>> to ZooKeeper but the connection closes im
> > > >>>>>>  mediately. This could be a sign that the server has too many
> > > connections
> > > >>>>>> (30 is the default). Consider inspecting your ZK server logs for
> > > that error
> > > >>>>>> and then make sure you are reusing HBaseConfiguration as often as
> > > you can.
> > > >>>>>> See HTable's javadoc for more information.  at
> > > >>>>>>
> > > org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher.<init>(ZooKeeperWatcher.j
> > > >>>>>> ava:160)    at
> > > >>>>>>
> > > org.apache.hadoop.hbase.regionserver.HRegionServer.initializeZooKeeper(HRegi
> > > >>>>>> onServer.java:489)  at
> > > >>>>>>
> > > org.apache.hadoop.hbase.regionserver.HRegionServer.initialize(HRegionServer.
> > > >>>>>> java:465)   at
> > > >>>>>>
> > > org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:56
> > > >>>>>> 4)  at java.lang.Thread.run(Thread.java:662)Caused by:
> > > >>>>>> org.apache.zookeeper.KeeperException$ConnectionLossException:
> > > >>>>>> KeeperErrorCode = ConnectionLoss for /hbase at
> > > >>>>>>
> > > org.apache.zookeeper.KeeperException.create(KeeperException.java:90)
> > >  at
> > > >>>>>>
> > > org.apache.zookeeper.KeeperException.create(KeeperException.java:42)
> > >  at
> > > >>>>>> org.apache.zookeeper.ZooKeeper.exists(ZooKeeper.java:815) at
> > > org.apac
> > > >>>>>>  he.zookeeper.ZooKeeper.exists(ZooKeeper.java:843)  at
> > > >>>>>>
> > > org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndFailSilent(ZKUtil.java:930
> > > >>>>>> )   at
> > > >>>>>>
> > > org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher.<init>(ZooKeeperWatcher.j
> > > >>>>>> ava:138)    ... 4 more12/05/12 14:42:25 INFO
> > > regionserver.HRegionServer:
> > > >>>>>> STOPPED: Initialization of RS failed.  Hence aborting RS.Exception
> > > in thread
> > > >>>>>> "regionserver60020" java.lang.NullPointerException  at
> > > >>>>>>
> > > org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:63
> > > >>>>>> 3)  at java.lang.Thread.run(Thread.java:662)
> > > >>>>>> So any help???
> > > >>>>>>> Date: Sat, 12 May 2012 20:22:03 +0530
> > > >>>>>>> Subject: Re: Important "Undefined Error"
> > > >>>>>>> From: dwivedishashwat@gmail.com
> > > >>>>>>> To: user@hbase.apache.org
> > > >>>>>>>
> > > >>>>>>> you can turn off hadoop safe mode uisng *hadoop dfsadmin -safemode
> > > >>>>>>> leave*
> > > >>>>>>>
> > > >>>>>>> On Sat, May 12, 2012 at 8:15 PM, shashwat shriparv <
> > > >>>>>>> dwivedishashwat@gmail.com> wrote:
> > > >>>>>>>
> > > >>>>>>>> First thing copy core-site.xml, dfs xml from hadoop conf directory
> > > >>>>>>>> to hbase conf dirctory. and turn of hadoop save mode and then
> > > try...
> > > >>>>>>>>
> > > >>>>>>>>
> > > >>>>>>>> On Sat, May 12, 2012 at 6:27 PM, Harsh J <harsh@cloudera.com>
> > > wrote:
> > > >>>>>>>>
> > > >>>>>>>>> Dalia,
> > > >>>>>>>>>
> > > >>>>>>>>> Is your NameNode running fine? The issue is that HBase Master has
> > > >>>>>>>>> been asked to talk to HDFS, but it can't connect to the HDFS
> > > >>>>>>>>> NameNode. Does "hadoop dfs -touchz foobar" pass or fail with
> > > similar
> > > >>>>>> retry issues?
> > > >>>>>>>>>
> > > >>>>>>>>> What's your fs.default.name's value in Hadoop's core-site.xml?
> > > And
> > > >>>>>>>>> whats the output of that fixed host command I'd posted before?
> > > >>>>>>>>>
> > > >>>>>>>>> On Sat, May 12, 2012 at 6:06 PM, Dalia Sobhy
> > > >>>>>>>>> <dalia.mohsobhy@hotmail.com>
> > > >>>>>>>>> wrote:
> > > >>>>>>>>>>
> > > >>>>>>>>>>
> > > >>>>>>>>>> Dear Harsh
> > > >>>>>>>>>> When I run $hbase master start
> > > >>>>>>>>>> I found the following errors:12/05/12 08:32:42 INFO
> > > >>>>>>>>> ipc.HBaseRpcMetrics: Initializing RPC Metrics with
> > > >>>>>>>>> hostName=HMaster,
> > > >>>>>>>>> port=6000012/05/12 08:32:42 INFO security.UserGroupInformation:
> > > >>>>>>>>> JAAS Configuration already set up for Hadoop, not
> > > >>>>>>>>> re-installing.12/05/12
> > > >>>>>>>>> 08:32:42 INFO ipc.HBaseServer: IPC Server Responder:
> > > >>>>>>>>> starting12/05/12
> > > >>>>>>>>> 08:32:42 INFO ipc.HBaseServer: IPC Server listener on 60000:
> > > >>>>>>>>> starting12/05/12 08:32:42 INFO ipc.HBaseServer: IPC Server
> > > handler
> > > >>>>>>>>> 0 on
> > > >>>>>>>>> 60000: starting12/05/12 08:32:42 INFO ipc.HBaseServer: IPC Server
> > > >>>>>>>>> handler 1 on 60000: starting12/05/12 08:32:42 INFO
> > > ipc.HBaseServer:
> > > >>>>>>>>> IPC Server handler 2 on 60000: starting12/05/12 08:32:42 INFO
> > > >>>>>>>>> ipc.HBaseServer: IPC Server handler 3 on 60000: starting12/05/12
> > > 08:32:
> > > >>>>>> 42 INFO ipc.HBaseServer:
> > > >>>>>>>>> IPC Server handler 5 on 60000: starting12/05/12 08:32:42 INFO
> > > >>>>>>>>> ipc.HBaseServer: IPC Server handler 4 on 60000: starting12/05/12
> > > >>>>>>>>> 08:32:42 INFO ipc.HBaseServer: IPC Server handler 7 on 60000:
> > > >>>>>>>>> starting12/05/12
> > > >>>>>>>>> 08:32:42 INFO ipc.HBaseServer: IPC Serv
> > > >>>>>>>>>>  er handler 6 on 60000: starting12/05/12 08:32:42 INFO
> > > >>>>>> ipc.HBaseServer:
> > > >>>>>>>>> IPC Server handler 8 on 60000: starting12/05/12 08:32:42 INFO
> > > >>>>>>>>> ipc.HBaseServer: IPC Server handler 9 on 60000: starting12/05/12
> > > >>>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client
> > > >>>>>>>>> environment:zookeeper.version=3.3.5-cdh3u4--1, built on
> > > 05/07/2012
> > > >>>>>>>>> 21:12
> > > >>>>>>>>> GMT12/05/12 08:32:42 INFO zookeeper.ZooKeeper: Client
> > > environment:
> > > >>>>>>>>> host.name=namenode12/05/12 08:32:42 INFO zookeeper.ZooKeeper:
> > > >>>>>>>>> Client
> > > >>>>>>>>> environment:java.version=1.6.0_3012/05/12 08:32:42 INFO
> > > >>>>>>>>> zookeeper.ZooKeeper: Client environment:java.vendor=Sun
> > > >>>>>>>>> Microsystems
> > > >>>>>>>>> Inc.12/05/12 08:32:42 INFO zookeeper.ZooKeeper: Client
> > > >>>>>>>>>
> > > environment:java.home=/usr/lib/jvm/java-6-sun-1.6.0.30/jre12/05/12
> > > >>>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client
> > > >>>>>>>>>
> > > environment:java.class.path=/usr/lib/hbase/bin/../conf:/usr/lib/jvm
> > > >>>>>>>>>
> > > /java-6-sun/lib/tools.jar:/usr/lib/hbase/bin/..:/usr/lib/hbase/bin/
> > > >>>>>>>>>
> > > ../hbase-0.90.4-cdh3u3.jar:/usr/lib/hbase/bin/../hbase-0.90.4-cdh3u
> > > >>>>>>>>> 3-tests.jar:/usr/lib/hbase/bin/../lib/activation-1.1.jar:/u
> > > >>>>>>>>>>
> > > >>>>>>>>>
> > > >>>>>>>>>
> > > sr/lib/hbase/bin/../lib/asm-3.1.jar:/usr/lib/hbase/bin/../lib/avro-
> > > >>>>>>>>>
> > > 1.5.4.jar:/usr/lib/hbase/bin/../lib/avro-ipc-1.5.4.jar:/usr/lib/hba
> > > >>>>>>>>>
> > > se/bin/../lib/commons-cli-1.2.jar:/usr/lib/hbase/bin/../lib/commons
> > > >>>>>>>>>
> > > -codec-1.4.jar:/usr/lib/hbase/bin/../lib/commons-el-1.0.jar:/usr/li
> > > >>>>>>>>>
> > > b/hbase/bin/../lib/commons-httpclient-3.1.jar:/usr/lib/hbase/bin/..
> > > >>>>>>>>>
> > > /lib/commons-lang-2.5.jar:/usr/lib/hbase/bin/../lib/commons-logging
> > > >>>>>>>>>
> > > -1.1.1.jar:/usr/lib/hbase/bin/../lib/commons-net-1.4.1.jar:/usr/lib
> > > >>>>>>>>>
> > > /hbase/bin/../lib/core-3.1.1.jar:/usr/lib/hbase/bin/../lib/guava-r0
> > > >>>>>>>>>
> > > 6.jar:/usr/lib/hbase/bin/../lib/guava-r09-jarjar.jar:/usr/lib/hbase
> > > >>>>>>>>>
> > > /bin/../lib/hadoop-core.jar:/usr/lib/hbase/bin/../lib/jackson-core-
> > > >>>>>>>>>
> > > asl-1.5.2.jar:/usr/lib/hbase/bin/../lib/jackson-jaxrs-1.5.5.jar:/us
> > > >>>>>>>>>
> > > r/lib/hbase/bin/../lib/jackson-mapper-asl-1.5.2.jar:/usr/lib/hbase/
> > > >>>>>>>>>
> > > bin/../lib/jackson-xc-1.5.5.jar:/usr/lib/hbase/bin/../lib/jamon-run
> > > >>>>>>>>>
> > > time-2.3.1.jar:/usr/lib/hbase/bin/../lib/jasper-compiler-5.5.23.jar
> > > >>>>>>>>> :/usr/lib/hbase/bin/../lib/jasper-runtime-5.5.23.jar:/
> > > >>>>>>  usr/l
> > > >>>>>>>>>>
> > > >>>>>>>>>
> > > >>>>>>>>>
> > > ib/hbase/bin/../lib/jaxb-api-2.1.jar:/usr/lib/hbase/bin/../lib/jaxb
> > > >>>>>>>>>
> > > -impl-2.1.12.jar:/usr/lib/hbase/bin/../lib/jersey-core-1.4.jar:/usr
> > > >>>>>>>>>
> > > /lib/hbase/bin/../lib/jersey-json-1.4.jar:/usr/lib/hbase/bin/../lib
> > > >>>>>>>>>
> > > /jersey-server-1.4.jar:/usr/lib/hbase/bin/../lib/jettison-1.1.jar:/
> > > >>>>>>>>>
> > > usr/lib/hbase/bin/../lib/jetty-6.1.26.jar:/usr/lib/hbase/bin/../lib
> > > >>>>>>>>>
> > > /jetty-util-6.1.26.jar:/usr/lib/hbase/bin/../lib/jruby-complete-1.6
> > > >>>>>>>>>
> > > .0.jar:/usr/lib/hbase/bin/../lib/jsp-2.1-6.1.14.jar:/usr/lib/hbase/
> > > >>>>>>>>>
> > > bin/../lib/jsp-api-2.1-6.1.14.jar:/usr/lib/hbase/bin/../lib/jsp-api
> > > >>>>>>>>>
> > > -2.1.jar:/usr/lib/hbase/bin/../lib/jsr311-api-1.1.1.jar:/usr/lib/hb
> > > >>>>>>>>>
> > > ase/bin/../lib/log4j-1.2.16.jar:/usr/lib/hbase/bin/../lib/netty-3.2
> > > >>>>>>>>>
> > > .4.Final.jar:/usr/lib/hbase/bin/../lib/protobuf-java-2.3.0.jar:/usr
> > > >>>>>>>>>
> > > /lib/hbase/bin/../lib/servlet-api-2.5-6.1.14.jar:/usr/lib/hbase/bin
> > > >>>>>>>>>
> > > /../lib/servlet-api-2.5.jar:/usr/lib/hbase/bin/../lib/slf4j-api-1.5
> > > >>>>>>>>>
> > > .8.jar:/usr/lib/hbase/bin/../lib/slf4j-log4j12-1.5.8.jar:/usr/lib/h
> > > >>>>>>>>> base/bin/../lib/snappy-java-1.0.3.2.jar:/usr/lib/hbase
> > > >>>>>>  /bin/
> > > >>>>>>>>>>
> > > >>>>>>>>>
> > > >>>>>>>>>
> > > ../lib/stax-api-1.0.1.jar:/usr/lib/hbase/bin/../lib/thrift-0.2.0.ja
> > > >>>>>>>>>
> > > r:/usr/lib/hbase/bin/../lib/velocity-1.5.jar:/usr/lib/hbase/bin/../
> > > >>>>>>>>>
> > > lib/xmlenc-0.52.jar:/usr/lib/hbase/bin/../lib/zookeeper.jar:/etc/zo
> > > >>>>>>>>>
> > > okeeper:/etc/hadoop-0.20/conf:/usr/lib/hadoop-0.20/hadoop-examples.
> > > >>>>>>>>>
> > > jar:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u3-core.jar:/usr/lib/had
> > > >>>>>>>>>
> > > oop-0.20/hadoop-0.20.2-cdh3u3-ant.jar:/usr/lib/hadoop-0.20/hadoop-c
> > > >>>>>>>>>
> > > ore-0.20.2-cdh3u3.jar:/usr/lib/hadoop-0.20/hadoop-test.jar:/usr/lib
> > > >>>>>>>>>
> > > /hadoop-0.20/hadoop-ant-0.20.2-cdh3u3.jar:/usr/lib/hadoop-0.20/hado
> > > >>>>>>>>>
> > > op-tools.jar:/usr/lib/hadoop-0.20/hadoop-tools-0.20.2-cdh3u3.jar:/u
> > > >>>>>>>>>
> > > sr/lib/hadoop-0.20/hadoop-test-0.20.2-cdh3u3.jar:/usr/lib/hadoop-0.
> > > >>>>>>>>>
> > > 20/hadoop-core.jar:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u3-exampl
> > > >>>>>>>>>
> > > es.jar:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u3-test.jar:/usr/lib/
> > > >>>>>>>>>
> > > hadoop-0.20/hadoop-ant.jar:/usr/lib/hadoop-0.20/hadoop-examples-0.2
> > > >>>>>>>>>
> > > 0.2-cdh3u3.jar:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u3-tools.jar:
> > > >>>>>>>>> /usr/lib/hadoop-0.20/lib/jasper-runtime-5.5.12.jar:/us
> > > >>>>>>  r/lib
> > > >>>>>>>>>>
> > > >>>>>>>>>
> > > >>>>>>>>>
> > > /hadoop-0.20/lib/hsqldb-1.8.0.10.jar:/usr/lib/hadoop-0.20/lib/jacks
> > > >>>>>>>>>
> > > on-mapper-asl-1.5.2.jar:/usr/lib/hadoop-0.20/lib/jets3t-0.6.1.jar:/
> > > >>>>>>>>>
> > > usr/lib/hadoop-0.20/lib/jetty-servlet-tester-6.1.26.cloudera.1.jar:
> > > >>>>>>>>>
> > > /usr/lib/hadoop-0.20/lib/jackson-core-asl-1.5.2.jar:/usr/lib/hadoop
> > > >>>>>>>>>
> > > -0.20/lib/oro-2.0.8.jar:/usr/lib/hadoop-0.20/lib/ant-contrib-1.0b3.
> > > >>>>>>>>>
> > > jar:/usr/lib/hadoop-0.20/lib/commons-daemon-1.0.1.jar:/usr/lib/hado
> > > >>>>>>>>>
> > > op-0.20/lib/mockito-all-1.8.2.jar:/usr/lib/hadoop-0.20/lib/aspectjr
> > > >>>>>>>>>
> > > t-1.6.5.jar:/usr/lib/hadoop-0.20/lib/commons-lang-2.4.jar:/usr/lib/
> > > >>>>>>>>>
> > > hadoop-0.20/lib/junit-4.5.jar:/usr/lib/hadoop-0.20/lib/commons-code
> > > >>>>>>>>>
> > > c-1.4.jar:/usr/lib/hadoop-0.20/lib/servlet-api-2.5-6.1.14.jar:/usr/
> > > >>>>>>>>>
> > > lib/hadoop-0.20/lib/log4j-1.2.15.jar:/usr/lib/hadoop-0.20/lib/jsch-
> > > >>>>>>>>>
> > > 0.1.42.jar:/usr/lib/hadoop-0.20/lib/core-3.1.1.jar:/usr/lib/hadoop-
> > > >>>>>>>>>
> > > 0.20/lib/jetty-6.1.26.cloudera.1.jar:/usr/lib/hadoop-0.20/lib/commo
> > > >>>>>>>>>
> > > ns-logging-1.0.4.jar:/usr/lib/hadoop-0.20/lib/jetty-util-6.1.26.clo
> > > >>>>>>>>> udera.1.jar:/usr/lib/hadoop-0.20/lib/servlet-api-2.5-2
> > > >>>>>>  00812
> > > >>>>>>>>>>
> > > >>>>>>>>>
> > > >>>>>>>>>
> > > 11.jar:/usr/lib/hadoop-0.20/lib/jasper-compiler-5.5.12.jar:/usr/lib
> > > >>>>>>>>>
> > > /hadoop-0.20/lib/kfs-0.2.2.jar:/usr/lib/hadoop-0.20/lib/commons-cli
> > > >>>>>>>>>
> > > -1.2.jar:/usr/lib/hadoop-0.20/lib/commons-net-1.4.1.jar:/usr/lib/ha
> > > >>>>>>>>>
> > > doop-0.20/lib/commons-httpclient-3.1.jar:/usr/lib/hadoop-0.20/lib/s
> > > >>>>>>>>>
> > > lf4j-api-1.4.3.jar:/usr/lib/hadoop-0.20/lib/xmlenc-0.52.jar:/usr/li
> > > >>>>>>>>>
> > > b/hadoop-0.20/lib/commons-logging-api-1.0.4.jar:/usr/lib/hadoop-0.2
> > > >>>>>>>>>
> > > 0/lib/commons-el-1.0.jar:/usr/lib/hadoop-0.20/lib/slf4j-log4j12-1.4
> > > >>>>>>>>>
> > > .3.jar:/usr/lib/hadoop-0.20/lib/aspectjtools-1.6.5.jar:/usr/lib/had
> > > >>>>>>>>>
> > > oop-0.20/lib/guava-r09-jarjar.jar:/usr/lib/hadoop-0.20/lib/hadoop-f
> > > >>>>>>>>>
> > > airscheduler-0.20.2-cdh3u3.jar:/usr/lib/zookeeper/zookeeper.jar:/us
> > > >>>>>>>>>
> > > r/lib/zookeeper/zookeeper-3.3.5-cdh3u4.jar:/usr/lib/zookeeper/lib/l
> > > >>>>>>>>>
> > > og4j-1.2.15.jar:/usr/lib/zookeeper/lib/jline-0.9.94.jar::/usr/lib/h
> > > >>>>>>>>>
> > > adoop-0.20/conf:/usr/lib/hadoop-0.20/hadoop-core-0.20.2-cdh3u3.jar:
> > > >>>>>>>>>
> > > /usr/lib/hadoop-0.20/lib/ant-contrib-1.0b3.jar:/usr/lib/hadoop-0.20
> > > >>>>>>>>> /lib/aspectjrt-1.6.5.jar:/usr/lib/hadoop-0.20/lib/aspe
> > > >>>>>>  ctjto
> > > >>>>>>>>>>
> > > >>>>>>>>>
> > > >>>>>>>>>
> > > ols-1.6.5.jar:/usr/lib/hadoop-0.20/lib/commons-cli-1.2.jar:/usr/lib
> > > >>>>>>>>>
> > > /hadoop-0.20/lib/commons-codec-1.4.jar:/usr/lib/hadoop-0.20/lib/com
> > > >>>>>>>>>
> > > mons-daemon-1.0.1.jar:/usr/lib/hadoop-0.20/lib/commons-el-1.0.jar:/
> > > >>>>>>>>>
> > > usr/lib/hadoop-0.20/lib/commons-httpclient-3.1.jar:/usr/lib/hadoop-
> > > >>>>>>>>>
> > > 0.20/lib/commons-lang-2.4.jar:/usr/lib/hadoop-0.20/lib/commons-logg
> > > >>>>>>>>>
> > > ing-1.0.4.jar:/usr/lib/hadoop-0.20/lib/commons-logging-api-1.0.4.ja
> > > >>>>>>>>>
> > > r:/usr/lib/hadoop-0.20/lib/commons-net-1.4.1.jar:/usr/lib/hadoop-0.
> > > >>>>>>>>>
> > > 20/lib/core-3.1.1.jar:/usr/lib/hadoop-0.20/lib/guava-r09-jarjar.jar
> > > >>>>>>>>>
> > > :/usr/lib/hadoop-0.20/lib/hadoop-fairscheduler-0.20.2-cdh3u3.jar:/u
> > > >>>>>>>>>
> > > sr/lib/hadoop-0.20/lib/hsqldb-1.8.0.10.jar:/usr/lib/hadoop-0.20/lib
> > > >>>>>>>>>
> > > /jackson-core-asl-1.5.2.jar:/usr/lib/hadoop-0.20/lib/jackson-mapper
> > > >>>>>>>>>
> > > -asl-1.5.2.jar:/usr/lib/hadoop-0.20/lib/jasper-compiler-5.5.12.jar:
> > > >>>>>>>>>
> > > /usr/lib/hadoop-0.20/lib/jasper-runtime-5.5.12.jar:/usr/lib/hadoop-
> > > >>>>>>>>>
> > > 0.20/lib/jets3t-0.6.1.jar:/usr/lib/hadoop-0.20/lib/jetty-6.1.26.clo
> > > >>>>>>>>> udera.1.jar:/usr/lib/hadoop-0.20/lib/jetty-servlet-tes
> > > >>>>>>  ter-6
> > > >>>>>>>>>>
> > > >>>>>>>>>
> > > >>>>>>>>>
> > > .1.26.cloudera.1.jar:/usr/lib/hadoop-0.20/lib/jetty-util-6.1.26.clo
> > > >>>>>>>>>
> > > udera.1.jar:/usr/lib/hadoop-0.20/lib/jsch-0.1.42.jar:/usr/lib/hadoo
> > > >>>>>>>>>
> > > p-0.20/lib/junit-4.5.jar:/usr/lib/hadoop-0.20/lib/kfs-0.2.2.jar:/us
> > > >>>>>>>>>
> > > r/lib/hadoop-0.20/lib/log4j-1.2.15.jar:/usr/lib/hadoop-0.20/lib/moc
> > > >>>>>>>>>
> > > kito-all-1.8.2.jar:/usr/lib/hadoop-0.20/lib/oro-2.0.8.jar:/usr/lib/
> > > >>>>>>>>>
> > > hadoop-0.20/lib/servlet-api-2.5-20081211.jar:/usr/lib/hadoop-0.20/l
> > > >>>>>>>>>
> > > ib/servlet-api-2.5-6.1.14.jar:/usr/lib/hadoop-0.20/lib/slf4j-api-1.
> > > >>>>>>>>>
> > > 4.3.jar:/usr/lib/hadoop-0.20/lib/slf4j-log4j12-1.4.3.jar:/usr/lib/h
> > > >>>>>>>>> adoop-0.20/lib/xmlenc-0.52.jar12/05/12
> > > >>>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client
> > > >>>>>>>>>
> > > environment:java.library.path=/usr/lib/hadoop-0.20/lib/native/Linux
> > > >>>>>>>>> -amd64-64:/usr/lib/hbase/bin/../lib/native/Linux-amd64-6412/05/12
> > > >>>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client
> > > >>>>>>>>> environment:java.io.tmpdir=/tmp12/05/12 08:32:42 INFO
> > > >>>>>> zookeeper.ZooKeeper:
> > > >>>>>>>>> Client environment:java.compiler=<NA>12/05/12 08:32:42 INFO
> > > >>>>>>>>> zookeeper.ZooKeeper: Client environment:os.name=Linux12/05/12
> > > >>>>>>>>> 08:32:42
> > > >>>>>>>>>>  INFO zookeeper.ZooKeeper: Client
> > > >>>>>>>>>> environment:os.arch=amd6412/05/12
> > > >>>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client
> > > >>>>>>>>> environment:os.version=2.6.35-22-server12/05/12 08:32:42 INFO
> > > >>>>>>>>> zookeeper.ZooKeeper: Client environment:user.name=dalia12/05/12
> > > >>>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client
> > > >>>>>>>>> environment:user.home=/home/dalia12/05/12
> > > >>>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client
> > > >>>>>>>>> environment:user.dir=/home/dalia12/05/12 08:32:42 INFO
> > > >>>>>> zookeeper.ZooKeeper:
> > > >>>>>>>>> Initiating client connection, connectString=namenode:2181
> > > >>>>>>>>> sessionTimeout=180000 watcher=master:6000012/05/12 08:32:42 INFO
> > > >>>>>>>>> zookeeper.ClientCnxn: Opening socket connection to server
> > > namenode/
> > > >>>>>>>>> 10.0.2.3:218112/05/12 08:32:42 INFO zookeeper.ClientCnxn: Socket
> > > >>>>>>>>> connection established to namenode/10.0.2.3:2181, initiating
> > > >>>>>>>>> session12/05/12 08:32:42 INFO zookeeper.ClientCnxn: Session
> > > >>>>>>>>> establishment complete on server namenode/10.0.2.3:2181,
> > > sessionid
> > > >>>>>>>>> = 0x13740bc4f70000c, negotiated timeout = 4000012/05/12 08:32:42
> > > >>>>>>>>> INFO
> > > >>>>>>>>> jvm.JvmMetrics: Initializing JVM Metrics with
> > > >>>>>>>>>>  processName=Master, sessionId=namenode:6000012/05/12 08:32:42
> > > >>>>>>>>>> INFO
> > > >>>>>>>>> hbase.metrics: MetricsString added: revision12/05/12 08:32:42
> > > INFO
> > > >>>>>>>>> hbase.metrics: MetricsString added: hdfsUser12/05/12 08:32:42
> > > INFO
> > > >>>>>>>>> hbase.metrics: MetricsString added: hdfsDate12/05/12 08:32:42
> > > INFO
> > > >>>>>>>>> hbase.metrics: MetricsString added: hdfsUrl12/05/12 08:32:42 INFO
> > > >>>>>>>>> hbase.metrics: MetricsString added: date12/05/12 08:32:42 INFO
> > > >>>>>>>>> hbase.metrics: MetricsString added: hdfsRevision12/05/12 08:32:42
> > > >>>>>>>>> INFO
> > > >>>>>>>>> hbase.metrics: MetricsString added: user12/05/12 08:32:42 INFO
> > > >>>>>>>>> hbase.metrics: MetricsString added: hdfsVersion12/05/12 08:32:42
> > > >>>>>>>>> INFO
> > > >>>>>>>>> hbase.metrics: MetricsString added: url12/05/12 08:32:42 INFO
> > > >>>>>>>>> hbase.metrics: MetricsString added: version12/05/12 08:32:42 INFO
> > > >>>>>>>>> hbase.metrics: new MBeanInfo12/05/12 08:32:42 INFO hbase.metrics:
> > > >>>>>>>>> new
> > > >>>>>>>>> MBeanInfo12/05/12 08:32:42 INFO metrics.MasterMetrics:
> > > >>>>>>>>> Initialized12/05/12
> > > >>>>>>>>> 08:32:42 INFO master.ActiveMasterManager:
> > > >>>>>>>>> Master=namenode:6000012/05/12
> > > >>>>>>>>> 08:32:44 INFO ipc.Client: Retrying connect to serve
> > > >>>>>>>>>>  r: namenode/10.0.2.3:8020. Already tried 0 time(s).12/05/12
> > > >>>>>>>>>> 08:32:45
> > > >>>>>>>>> INFO ipc.Client: Retrying connect to server: namenode/
> > > 10.0.2.3:8020.
> > > >>>>>>>>> Already tried 1 time(s).12/05/12 08:32:46 INFO ipc.Client:
> > > Retrying
> > > >>>>>>>>> connect to server: namenode/10.0.2.3:8020. Already tried 2
> > > >>>>>>>>> time(s).12/05/12
> > > >>>>>>>>> 08:32:47 INFO ipc.Client: Retrying connect to server: namenode/
> > > >>>>>>>>> 10.0.2.3:8020. Already tried 3 time(s).12/05/12 08:32:48 INFO
> > > >>>>>>>>> ipc.Client: Retrying connect to server: namenode/10.0.2.3:8020.
> > > >>>>>>>>> Already tried 4 time(s).12/05/12 08:32:49 INFO ipc.Client:
> > > Retrying
> > > >>>>>>>>> connect to
> > > >>>>>>>>> server: namenode/10.0.2.3:8020. Already tried 5 time(s).12/05/12
> > > >>>>>>>>> 08:32:50 INFO ipc.Client: Retrying connect to server: namenode/
> > > >>>>>>>>> 10.0.2.3:8020. Already tried 6 time(s).12/05/12 08:32:51 INFO
> > > >>>>>>>>> ipc.Client: Retrying connect to server: namenode/10.0.2.3:8020.
> > > >>>>>>>>> Already tried 7 time(s).12/05/12 08:32:52 INFO ipc.Client:
> > > Retrying
> > > >>>>>>>>> connect to
> > > >>>>>>>>> server: namenode/10.0.2.3:8020. Already tried 8 time(s).12/05/12
> > > >>>>>>>>> 08:32:53 INFO ipc.Client: Retrying connect to ser
> > > >>>>>>>>>>  ver: namenode/10.0.2.3:8020. Already tried 9 time(s).12/05/12
> > > >>>>>>>>> 08:32:53 FATAL master.HMaster: Unhandled exception. Starting
> > > >>>>>>>>> shutdown.java.net.ConnectException: Call to
> > > >>>>>> namenode/10.0.2.3:8020failed on connection exception:
> > > >>>>>> java.net.ConnectException: Connection
> > > >>>>>>>>> refused    at
> > > >>>>>> org.apache.hadoop.ipc.Client.wrapException(Client.java:1134)
> > > >>>>>>>>> at org.apache.hadoop.ipc.Client.call(Client.java:1110)  at
> > > >>>>>>>>> org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:226)       at
> > > >>>>>>>>> $Proxy6.getProtocolVersion(Unknown Source)   at
> > > >>>>>>>>> org.apache.hadoop.ipc.RPC.getProxy(RPC.java:398)     at
> > > >>>>>>>>> org.apache.hadoop.ipc.RPC.getProxy(RPC.java:384)     at
> > > >>>>>>>>>
> > > org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:123)
> > > >>>>>>>>>   at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:246)
> > > >>>>>>>>> at
> > > >>>>>>>>> org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:208)  at
> > > >>>>>>>>>
> > > >>>>>>
> > > org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSyste
> > > >>>>>> m.java:89)
> > > >>>>>>>>>       at
> > > >>>>>>>>>
> > > org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1563)
> > > >>>>>>>>>   at org.apache.hadoop.fs.FileSystem.acc
> > > >>>>>>>>>>  ess$200(FileSystem.java:67)    at
> > > >>>>>>>>>
> > > org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:1597)
> > > >>>>>>>>>  at
> > > org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1579)
> > > >>>>>> at
> > > >>>>>>>>> org.apache.hadoop.fs.FileSystem.get(FileSystem.java:228)     at
> > > >>>>>>>>> org.apache.hadoop.fs.Path.getFileSystem(Path.java:183)       at
> > > >>>>>>>>>
> > > org.apache.hadoop.hbase.util.FSUtils.getRootDir(FSUtils.java:364)    at
> > > >>>>>>>>>
> > > >>>>>>
> > > org.apache.hadoop.hbase.master.MasterFileSystem.<init>(MasterFileSystem.java
> > > >>>>>> :86)
> > > >>>>>>>>>     at
> > > >>>>>>>>>
> > > >>>>>>
> > > org.apache.hadoop.hbase.master.HMaster.finishInitialization(HMaster.java:360
> > > >>>>>> )
> > > >>>>>>>>>        at
> > > >>>>>>>>>
> > > org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:293)Caused by:
> > > >>>>>>>>> java.net.ConnectException: Connection refused at
> > > >>>>>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)     at
> > > >>>>>>>>>
> > > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
> > > >>>>>>>>>   at
> > > >>>>>>>>>
> > > >>>>>>
> > > org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:2
> > > >>>>>> 06)
> > > >>>>>>>>>      at
> > > org.apache.hadoop.net.NetUtils.connect(NetUtils.java:408)    at
> > > >>>>>>>>>
> > > >>>>>>
> > > org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:425)
> > > >>>>>>>>>   a
> > > >>>>>>>>>>  t
> > > >>>>>>>>>
> > > org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:
> > > >>>>>>>>> 532)  at
> > > >>>>>>>>>
> > > org.apache.hadoop.ipc.Client$Connection.access$2300(Client.java:210
> > > >>>>>>>>> ) at
> > > >>>>>>>>> org.apache.hadoop.ipc.Client.getConnection(Client.java:1247) at
> > > >>>>>>>>> org.apache.hadoop.ipc.Client.call(Client.java:1078)  ... 18
> > > >>>>>>>>> more12/05/12
> > > >>>>>>>>> 08:32:53 INFO master.HMaster: Aborting12/05/12 08:32:53 DEBUG
> > > >>>>>>>>> master.HMaster: Stopping service threads12/05/12 08:32:53 INFO
> > > >>>>>>>>> ipc.HBaseServer: Stopping server on 6000012/05/12 08:32:53 INFO
> > > >>>>>>>>> ipc.HBaseServer: IPC Server handler 5 on 60000: exiting12/05/12
> > > >>>>>>>>> 08:32:53 INFO ipc.HBaseServer: Stopping IPC Server listener on
> > > >>>>>>>>> 6000012/05/12
> > > >>>>>>>>> 08:32:53 INFO ipc.HBaseServer: IPC Server handler 1 on 60000:
> > > >>>>>>>>> exiting12/05/12 08:32:53 INFO ipc.HBaseServer: IPC Server
> > > handler 0
> > > >>>>>>>>> on
> > > >>>>>>>>> 60000: exiting12/05/12 08:32:53 INFO ipc.HBaseServer: IPC Server
> > > >>>>>>>>> handler 3 on 60000: exiting12/05/12 08:32:53 INFO
> > > ipc.HBaseServer:
> > > >>>>>>>>> IPC Server handler
> > > >>>>>>>>> 7 on 60000: exiting12/05/12 08:32:53 INFO ipc.HBaseServer: IPC
> > > >>>>>>>>> Server handler 9 on 60000: exiting1
> > > >>>>>>>>>>  2/05/12 08:32:53 INFO ipc.HBaseServer: IPC Server handler 6 on
> > > >>>>>> 60000:
> > > >>>>>>>>> exiting12/05/12 08:32:53 INFO ipc.HBaseServer: IPC Server
> > > handler 4
> > > >>>>>>>>> on
> > > >>>>>>>>> 60000: exiting12/05/12 08:32:53 INFO ipc.HBaseServer: IPC Server
> > > >>>>>>>>> handler 2 on 60000: exiting12/05/12 08:32:53 INFO
> > > ipc.HBaseServer:
> > > >>>>>>>>> IPC Server handler
> > > >>>>>>>>> 8 on 60000: exiting12/05/12 08:32:53 INFO ipc.HBaseServer:
> > > Stopping
> > > >>>>>>>>> IPC Server Responder12/05/12 08:32:53 INFO zookeeper.ZooKeeper:
> > > >>>>>> Session:
> > > >>>>>>>>> 0x13740bc4f70000c closed12/05/12 08:32:53 INFO
> > > zookeeper.ClientCnxn:
> > > >>>>>>>>> EventThread shut down12/05/12 08:32:53 INFO master.HMaster:
> > > HMaster
> > > >>>>>>>>> main thread exiting> From: harsh@cloudera.com
> > > >>>>>>>>>>> Date: Sat, 12 May 2012 17:28:29 +0530
> > > >>>>>>>>>>> Subject: Re: Important "Undefined Error"
> > > >>>>>>>>>>> To: user@hbase.apache.org
> > > >>>>>>>>>>>
> > > >>>>>>>>>>> Hi Dalia,
> > > >>>>>>>>>>>
> > > >>>>>>>>>>> On Sat, May 12, 2012 at 5:14 PM, Dalia Sobhy <
> > > >>>>>>>>> dalia.mohsobhy@hotmail.com> wrote:
> > > >>>>>>>>>>>>
> > > >>>>>>>>>>>> Dear all,
> > > >>>>>>>>>>>> I have first a problem with Hbase I am trying to install it on
> > > >>>>>>>>>>>> a
> > > >>>>>>>>> distributed/multinode cluster..
> > > >>>>>>>>>>>> I am using the cloudera
> > > >>>>>>>>>
> > > https://ccp.cloudera.com/display/CDH4B2/HBase+Installation#HBaseIns
> > > >>>>>>>>> tallation-StartingtheHBaseMaster
> > > >>>>>>>>>>>> But when I write this command
> > > >>>>>>>>>>>> Creating the /hbase Directory in HDFS $sudo -u hdfs hadoop fs
> > > >>>>>>>>>>>> -mkdir
> > > >>>>>>>>> /hbase
> > > >>>>>>>>>>>> I get the following error:12/05/12 07:20:42 INFO
> > > >>>>>>>>> security.UserGroupInformation: JAAS Configuration already set up
> > > >>>>>>>>> for Hadoop, not re-installing.
> > > >>>>>>>>>>>
> > > >>>>>>>>>>> This is not an error and you shouldn't be worried. It is rather
> > > >>>>>>>>>>> a noisy INFO log that should be fixed (as a DEBUG level
> > > instead)
> > > >>>>>>>>>>> in subsequent releases (Are you using CDH3 or CDH4? IIRC only
> > > >>>>>>>>>>> CDH3u3 printed these, not in anything above that.)
> > > >>>>>>>>>>>
> > > >>>>>>>>>>>> 2. Another Aspect is when I start the hbase master it closes
> > > >>>>>>>>> automatically after a while.
> > > >>>>>>>>>>>
> > > >>>>>>>>>>> Could you post us your HMaster start->crash log? You can use a
> > > >>>>>>>>>>> service like pastebin.com to send us the output.
> > > >>>>>>>>>>>
> > > >>>>>>>>>>>> 3. Also this command is not working$host -v -t A
> > > >>>>>> `namenode`namenode:
> > > >>>>>>>>> command not found
> > > >>>>>>>>>>>
> > > >>>>>>>>>>> The right command is perhaps just:
> > > >>>>>>>>>>>
> > > >>>>>>>>>>> $host -v -t A `hostname`
> > > >>>>>>>>>>>
> > > >>>>>>>>>>> --
> > > >>>>>>>>>>> Harsh J
> > > >>>>>>>>>>
> > > >>>>>>>>>
> > > >>>>>>>>>
> > > >>>>>>>>>
> > > >>>>>>>>> --
> > > >>>>>>>>> Harsh J
> > > >>>>>>>>>
> > > >>>>>>>>
> > > >>>>>>>>
> > > >>>>>>>>
> > > >>>>>>>> --
> > > >>>>>>>>
> > > >>>>>>>>
> > > >>>>>>>> ∞
> > > >>>>>>>> Shashwat Shriparv
> > > >>>>>>>>
> > > >>>>>>>>
> > > >>>>>>>>
> > > >>>>>>>
> > > >>>>>>>
> > > >>>>>>> --
> > > >>>>>>>
> > > >>>>>>>
> > > >>>>>>> ∞
> > > >>>>>>> Shashwat Shriparv
> > > >>>>>>
> > > >>>>>>
> > > >>>>>
> > > >>>>
> > >
> > 
> > 
> > 
> > -- 
> > 
> > 
> > ∞
> > Shashwat Shriparv
>  		 	   		  
 		 	   		  
Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message