hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From N Keywal <nkey...@gmail.com>
Subject Re: Important "Undefined Error"
Date Mon, 14 May 2012 15:28:15 GMT
In core-file.xml, do you have this?

<configuration>
<property>
<name>fs.default.name</name>
 <value>hdfs://namenode:8020/hbase</value>
</property>

If you want hbase to connect to 8020 you must have hdfs listening on
8020 as well.


On Mon, May 14, 2012 at 5:17 PM, Dalia Sobhy <dalia.mohsobhy@hotmail.com> wrote:
> Hiiii
>
> I have tried to make both ports the same.
> But the prob is the hbase cannot connect to port 8020.
> When i run nmap hostname, port 8020 wasnt with the list of open ports.
> I have tried what harsh told me abt.
> I used the same port he used but same error occurred.
> Another aspect in cloudera doc it says that i have to canonical name for the host ex: namenode.example.com as the hostname, but i didnt find it in any tutorial. No one makes it.
> Note that i am deploying my cluster in fully distributed mode i.e am using 4 machines..
>
> So any ideas??!!
>
> Sent from my iPhone
>
> On 2012-05-14, at 4:07 PM, "N Keywal" <nkeywal@gmail.com> wrote:
>
>> Hi,
>>
>> There could be multiple issues, but it's strange to have in hbase-site.xml
>>
>>  <value>hdfs://namenode:9000/hbase</value>
>>
>> while the core-site.xml says:
>>
>> <value>hdfs://namenode:54310/</value>
>>
>> The two entries should match.
>>
>> I would recommend to:
>> - use netstat to check the ports (netstat -l)
>> - do the check recommended by Harsh J previously.
>>
>> N.
>>
>>
>> On Mon, May 14, 2012 at 3:21 PM, Dalia Sobhy <dalia.mohsobhy@hotmail.com> wrote:
>>>
>>>
>>> pleaseeeeeeeeeeee helpppppppppppppppppppp
>>>
>>>> From: dalia.mohsobhy@hotmail.com
>>>> To: user@hbase.apache.org
>>>> Subject: RE: Important "Undefined Error"
>>>> Date: Mon, 14 May 2012 12:20:18 +0200
>>>>
>>>>
>>>>
>>>> Hi,
>>>> I tried what you told me, but nothing worked:(((
>>>> First when I run this command:dalia@namenode:~$ host -v -t A `hostname`Output:Trying "namenode"Host namenode not found: 3(NXDOMAIN)Received 101 bytes from 10.0.2.1#53 in 13 ms My core-site.xml:<configuration><property>        <name>fs.default.name</name>        <!--<value>hdfs://namenode:8020</value>-->        <value>hdfs://namenode:54310/</value></property></configuration>
>>>> My hdfs-site.xml<configuration><property><name>dfs.name.dir</name><value>/data/1/dfs/nn,/nfsmount/dfs/nn</value></property><!--<property><name>dfs.data.dir</name><value>/data/1/dfs/dn,/data/2/dfs/dn,/data/3/dfs/dn</value></property>--><property><name>dfs.datanode.max.xcievers</name><value>4096</value></property><property><name>dfs.replication</name><value>3</value></property><property> <name>dfs.permissions.superusergroup</name> <value>hadoop</value></property>
>>>> My Mapred-site.xml<configuration><name>mapred.local.dir</name><value>/data/1/mapred/local,/data/2/mapred/local,/data/3/mapred/local</value></configuration>
>>>> My Hbase-site.xml<configuration><property><name>hbase.cluster.distributed</name>  <value>true</value></property><property>  <name>hbase.rootdir</name>     <value>hdfs://namenode:9000/hbase</value></property><property><name>hbase.zookeeper.quorun</name> <value>namenode</value></property><property><name>hbase.regionserver.port</name><value>60020</value><description>The host and port that the HBase master runs at.</description></property><property><name>dfs.replication</name><value>1</value></property><property><name>hbase.zookeeper.property.clientPort</name><value>2181</value><description>Property from ZooKeeper's config zoo.cfg.The port at which the clients will connect.</description></property></configuration>
>>>> Please Help I am really disappointed I have been through all that for two weeks !!!!
>>>>
>>>>
>>>>
>>>>> From: dwivedishashwat@gmail.com
>>>>> To: user@hbase.apache.org
>>>>> Subject: RE: Important "Undefined Error"
>>>>> Date: Sat, 12 May 2012 23:31:49 +0530
>>>>>
>>>>> The problem is your hbase is not able to connect to Hadoop, can you put your
>>>>> hbase-site.xml content >> here.. have you specified localhost somewhere, if
>>>>> so remove localhost from everywhere and put your hdfsl namenode address
>>>>> suppose your namenode is running on master:9000 then put your hbase file
>>>>> system setting as master:9000/hbase here I am sending you the configuration
>>>>> which I am using in hbase and is working
>>>>>
>>>>>
>>>>> My hbase-site.xml content is
>>>>>
>>>>> <?xml version="1.0"?>
>>>>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>>>>> <!--
>>>>> /**
>>>>>  * Copyright 2010 The Apache Software Foundation
>>>>>  *
>>>>>  * Licensed to the Apache Software Foundation (ASF) under one
>>>>>  * or more contributor license agreements.  See the NOTICE file
>>>>>  * distributed with this work for additional information
>>>>>  * regarding copyright ownership.  The ASF licenses this file
>>>>>  * to you under the Apache License, Version 2.0 (the
>>>>>  * "License"); you may not use this file except in compliance
>>>>>  * with the License.  You may obtain a copy of the License at
>>>>>  *
>>>>>  *     http://www.apache.org/licenses/LICENSE-2.0
>>>>>  *
>>>>>  * Unless required by applicable law or agreed to in writing, software
>>>>>  * distributed under the License is distributed on an "AS IS" BASIS,
>>>>>  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
>>>>>  * See the License for the specific language governing permissions and
>>>>>  * limitations under the License.
>>>>>  */
>>>>> -->
>>>>> <configuration>
>>>>> <property>
>>>>> <name>hbase.rootdir</name>
>>>>> <value>hdfs://master:9000/hbase</value>
>>>>> </property>
>>>>> <property>
>>>>> <name>hbase.master</name>
>>>>> <value>master:60000</value>
>>>>> <description>The host and port that the HBase master runs at.</description>
>>>>> </property>
>>>>> <property>
>>>>> <name>hbase.regionserver.port</name>
>>>>> <value>60020</value>
>>>>> <description>The host and port that the HBase master runs at.</description>
>>>>> </property>
>>>>> <!--<property>
>>>>> <name>hbase.master.port</name>
>>>>> <value>60000</value>
>>>>> <description>The host and port that the HBase master runs at.</description>
>>>>> </property>-->
>>>>> <property>
>>>>> <name>hbase.cluster.distributed</name>
>>>>> <value>true</value>
>>>>> </property>
>>>>> <property>
>>>>> <name>hbase.tmp.dir</name>
>>>>> <value>/home/shashwat/Hadoop/hbase-0.90.4/temp</value>
>>>>> </property>
>>>>> <property>
>>>>> <name>hbase.zookeeper.quorum</name>
>>>>> <value>master</value>
>>>>> </property>
>>>>> <property>
>>>>> <name>dfs.replication</name>
>>>>> <value>1</value>
>>>>> </property>
>>>>> <property>
>>>>> <name>hbase.zookeeper.property.clientPort</name>
>>>>> <value>2181</value>
>>>>> <description>Property from ZooKeeper's config zoo.cfg.
>>>>> The port at which the clients will connect.
>>>>> </description>
>>>>> </property>
>>>>> <property>
>>>>> <name>hbase.zookeeper.property.dataDir</name>
>>>>> <value>/home/shashwat/zookeeper</value>
>>>>> <description>Property from ZooKeeper's config zoo.cfg.
>>>>> The directory where the snapshot is stored.
>>>>> </description>
>>>>> </property>
>>>>>
>>>>> </configuration>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> Check this out, and also stop hbase, If its not stopping kill all the
>>>>> processes, and after putting your  hdfs-site.xml, mapred-site.xml and
>>>>> core-site.sml to hbase conf directory try to restart, and also delete the
>>>>> folders created by hbase ,,, like temp directory or other then try to start.
>>>>>
>>>>> Regards
>>>>> ∞
>>>>> Shashwat Shriparv
>>>>>
>>>>>
>>>>> -----Original Message-----
>>>>> From: Dalia Sobhy [mailto:dalia.mohsobhy@hotmail.com]
>>>>> Sent: 12 May 2012 22:48
>>>>> To: user@hbase.apache.org
>>>>> Subject: RE: Important "Undefined Error"
>>>>>
>>>>>
>>>>> Hi Shashwat,
>>>>> I want to tell you about my configurations:
>>>>> I am using 4 nodesOne "Master": Namenode, SecondaryNamenode, Job Tracker,
>>>>> Zookeeper, HMasterThree "Slaves": datanodes, tasktrackers, regionservers In
>>>>> both master and slaves, all the hadoop daemons are working well, but as for
>>>>> the hbase master service it is not working..
>>>>> As for region server here is the error:12/05/12 14:42:13 INFO
>>>>> util.ServerCommandLine: vmName=Java HotSpot(TM) 64-Bit Server VM,
>>>>> vmVendor=Sun Microsystems Inc., vmVersion=20.1-b0212/05/12 14:42:13 INFO
>>>>> util.ServerCommandLine: vmInputArguments=[-Xmx1000m, -ea,
>>>>> -XX:+UseConcMarkSweepGC, -XX:+CMSIncrementalMode,
>>>>> -Dhbase.log.dir=/usr/lib/hbase/bin/../logs, -Dhbase.log.file=hbase.log,
>>>>> -Dhbase.home.dir=/usr/lib/hbase/bin/.., -Dhbase.id.str=,
>>>>> -Dhbase.root.logger=INFO,console,
>>>>> -Djava.library.path=/usr/lib/hadoop-0.20/lib/native/Linux-amd64-64:/usr/lib/
>>>>> hbase/bin/../lib/native/Linux-amd64-64]12/05/12 14:42:13 INFO
>>>>> ipc.HBaseRpcMetrics: Initializing RPC Metrics with hostName=HRegionServer,
>>>>> port=6002012/05/12 14:42:14 FATAL zookeeper.ZKConfig: The server in zoo.cfg
>>>>> cannot be set to localhost in a fully-distributed setup because it won't be
>>>>> reachable. See "Getting Started" for more information.12/05/12 14:42:14 WARN
>>>>> zookeeper.ZKConfig: Cannot read zoo.cfg, loading from XML
>>>>> filesjava.io.IOException:
>>>>>  The server in zoo.cfg cannot be set to localhost in a fully-distributed
>>>>> setup because it won't be reachable. See "Getting Started" for more
>>>>> information.        at
>>>>> org.apache.hadoop.hbase.zookeeper.ZKConfig.parseZooCfg(ZKConfig.java:172)
>>>>> at org.apache.hadoop.hbase.zookeeper.ZKConfig.makeZKProps(ZKConfig.java:68)
>>>>> at
>>>>> org.apache.hadoop.hbase.zookeeper.ZKConfig.getZKQuorumServersString(ZKConfig
>>>>> .java:249)  at
>>>>> org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher.<init>(ZooKeeperWatcher.j
>>>>> ava:117)    at
>>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.initializeZooKeeper(HRegi
>>>>> onServer.java:489)  at
>>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.initialize(HRegionServer.
>>>>> java:465)   at
>>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:56
>>>>> 4)  at java.lang.Thread.run(Thread.java:662)12/05/12 14:42:14 INFO
>>>>> zookeeper.ZooKeeper: Client environment:zookeeper.version=3.3.5-cdh3u4--1,
>>>>> built on 05/07/2012 21:12 GMT12/05/12 14:42:14 INFO zookeeper.ZooKeeper:
>>>>> Client environment:host.name=data
>>>>>  node212/05/12 14:42:14 INFO zookeeper.ZooKeeper: Client
>>>>> environment:java.version=1.6.0_2612/05/12 14:42:14 INFO zookeeper.ZooKeeper:
>>>>> Client environment:java.vendor=Sun Microsystems Inc.12/05/12 14:42:14 INFO
>>>>> zookeeper.ZooKeeper: Client environment:java.home=/usr/lib/jvm/java-6-sun-1.
>>>>> 6.0.26/jre12/05/12 14:42:14 INFO zookeeper.ZooKeeper: Client
>>>>> environment:java.class.path=/usr/lib/hbase/bin/../conf:/usr/lib/jvm/java-6-s
>>>>> un/lib/tools.jar:/usr/lib/hbase/bin/..:/usr/lib/hbase/bin/../hbase-0.90.6-cd
>>>>> h3u4.jar:/usr/lib/hbase/bin/../hbase-0.90.6-cdh3u4-tests.jar:/usr/lib/hbase/
>>>>> bin/../lib/activation-1.1.jar:/usr/lib/hbase/bin/../lib/asm-3.1.jar:/usr/lib
>>>>> /hbase/bin/../lib/avro-1.5.4.jar:/usr/lib/hbase/bin/../lib/avro-ipc-1.5.4.ja
>>>>> r:/usr/lib/hbase/bin/../lib/commons-cli-1.2.jar:/usr/lib/hbase/bin/../lib/co
>>>>> mmons-codec-1.4.jar:/usr/lib/hbase/bin/../lib/commons-el-1.0.jar:/usr/lib/hb
>>>>> ase/bin/../lib/commons-httpclient-3.1.jar:/usr/lib/hbase/bin/../lib/commons-
>>>>> lang-2.5.jar:/usr/lib/hbase/bin/../lib/commo
>>>>>
>>>>> ns-logging-1.1.1.jar:/usr/lib/hbase/bin/../lib/commons-net-1.4.1.jar:/usr/li
>>>>> b/hbase/bin/../lib/core-3.1.1.jar:/usr/lib/hbase/bin/../lib/guava-r06.jar:/u
>>>>> sr/lib/hbase/bin/../lib/guava-r09-jarjar.jar:/usr/lib/hbase/bin/../lib/hadoo
>>>>> p-core.jar:/usr/lib/hbase/bin/../lib/jackson-core-asl-1.5.2.jar:/usr/lib/hba
>>>>> se/bin/../lib/jackson-jaxrs-1.5.5.jar:/usr/lib/hbase/bin/../lib/jackson-mapp
>>>>> er-asl-1.5.2.jar:/usr/lib/hbase/bin/../lib/jackson-xc-1.5.5.jar:/usr/lib/hba
>>>>> se/bin/../lib/jamon-runtime-2.3.1.jar:/usr/lib/hbase/bin/../lib/jasper-compi
>>>>> ler-5.5.23.jar:/usr/lib/hbase/bin/../lib/jasper-runtime-5.5.23.jar:/usr/lib/
>>>>> hbase/bin/../lib/jaxb-api-2.1.jar:/usr/lib/hbase/bin/../lib/jaxb-impl-2.1.12
>>>>> .jar:/usr/lib/hbase/bin/../lib/jersey-core-1.4.jar:/usr/lib/hbase/bin/../lib
>>>>> /jersey-json-1.4.jar:/usr/lib/hbase/bin/../lib/jersey-server-1.4.jar:/usr/li
>>>>> b/hbase/bin/../lib/jettison-1.1.jar:/usr/lib/hbase/bin/../lib/jetty-6.1.26.j
>>>>> ar:/usr/lib/hbase/bin/../lib/jetty-util-6.1.26.jar:/usr/lib/hbase/bin/../lib
>>>>> /jruby-co
>>>>>
>>>>> mplete-1.6.0.jar:/usr/lib/hbase/bin/../lib/jsp-2.1-6.1.14.jar:/usr/lib/hbase
>>>>> /bin/../lib/jsp-api-2.1-6.1.14.jar:/usr/lib/hbase/bin/../lib/jsp-api-2.1.jar
>>>>> :/usr/lib/hbase/bin/../lib/jsr311-api-1.1.1.jar:/usr/lib/hbase/bin/../lib/lo
>>>>> g4j-1.2.16.jar:/usr/lib/hbase/bin/../lib/netty-3.2.4.Final.jar:/usr/lib/hbas
>>>>> e/bin/../lib/protobuf-java-2.3.0.jar:/usr/lib/hbase/bin/../lib/servlet-api-2
>>>>> .5-6.1.14.jar:/usr/lib/hbase/bin/../lib/servlet-api-2.5.jar:/usr/lib/hbase/b
>>>>> in/../lib/slf4j-api-1.5.8.jar:/usr/lib/hbase/bin/../lib/slf4j-log4j12-1.5.8.
>>>>> jar:/usr/lib/hbase/bin/../lib/snappy-java-1.0.3.2.jar:/usr/lib/hbase/bin/../
>>>>> lib/stax-api-1.0.1.jar:/usr/lib/hbase/bin/../lib/thrift-0.2.0.jar:/usr/lib/h
>>>>> base/bin/../lib/velocity-1.5.jar:/usr/lib/hbase/bin/../lib/xmlenc-0.52.jar:/
>>>>> usr/lib/hbase/bin/../lib/zookeeper.jar:/etc/zookeeper:/etc/hadoop-0.20/conf:
>>>>> /usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u4-ant.jar:/usr/lib/hadoop-0.20/hadoo
>>>>> p-0.20.2-cdh3u4-tools.jar:/usr/lib/hadoop-0.20/hadoop-tools.jar:/usr/lib/had
>>>>> oop-0.20/
>>>>>
>>>>> hadoop-examples-0.20.2-cdh3u4.jar:/usr/lib/hadoop-0.20/hadoop-ant-0.20.2-cdh
>>>>> 3u4.jar:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u4-examples.jar:/usr/lib/hado
>>>>> op-0.20/hadoop-ant.jar:/usr/lib/hadoop-0.20/hadoop-core-0.20.2-cdh3u4.jar:/u
>>>>> sr/lib/hadoop-0.20/hadoop-core.jar:/usr/lib/hadoop-0.20/hadoop-tools-0.20.2-
>>>>> cdh3u4.jar:/usr/lib/hadoop-0.20/hadoop-examples.jar:/usr/lib/hadoop-0.20/had
>>>>> oop-0.20.2-cdh3u4-core.jar:/usr/lib/hadoop-0.20/hadoop-test-0.20.2-cdh3u4.ja
>>>>> r:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u4-test.jar:/usr/lib/hadoop-0.20/ha
>>>>> doop-test.jar:/usr/lib/hadoop-0.20/lib/jetty-6.1.26.cloudera.1.jar:/usr/lib/
>>>>> hadoop-0.20/lib/jets3t-0.6.1.jar:/usr/lib/hadoop-0.20/lib/junit-4.5.jar:/usr
>>>>> /lib/hadoop-0.20/lib/commons-daemon-1.0.1.jar:/usr/lib/hadoop-0.20/lib/slf4j
>>>>> -api-1.4.3.jar:/usr/lib/hadoop-0.20/lib/commons-codec-1.4.jar:/usr/lib/hadoo
>>>>> p-0.20/lib/log4j-1.2.15.jar:/usr/lib/hadoop-0.20/lib/jasper-compiler-5.5.12.
>>>>> jar:/usr/lib/hadoop-0.20/lib/guava-r09-jarjar.jar:/usr/lib/hadoop-0.20/lib/j
>>>>> ackson-ma
>>>>>
>>>>> pper-asl-1.5.2.jar:/usr/lib/hadoop-0.20/lib/commons-logging-1.0.4.jar:/usr/l
>>>>> ib/hadoop-0.20/lib/ant-contrib-1.0b3.jar:/usr/lib/hadoop-0.20/lib/commons-la
>>>>> ng-2.4.jar:/usr/lib/hadoop-0.20/lib/kfs-0.2.2.jar:/usr/lib/hadoop-0.20/lib/j
>>>>> etty-util-6.1.26.cloudera.1.jar:/usr/lib/hadoop-0.20/lib/commons-httpclient-
>>>>> 3.1.jar:/usr/lib/hadoop-0.20/lib/jsch-0.1.42.jar:/usr/lib/hadoop-0.20/lib/sl
>>>>> f4j-log4j12-1.4.3.jar:/usr/lib/hadoop-0.20/lib/servlet-api-2.5-20081211.jar:
>>>>> /usr/lib/hadoop-0.20/lib/jasper-runtime-5.5.12.jar:/usr/lib/hadoop-0.20/lib/
>>>>> servlet-api-2.5-6.1.14.jar:/usr/lib/hadoop-0.20/lib/jetty-servlet-tester-6.1
>>>>> .26.cloudera.1.jar:/usr/lib/hadoop-0.20/lib/aspectjtools-1.6.5.jar:/usr/lib/
>>>>> hadoop-0.20/lib/xmlenc-0.52.jar:/usr/lib/hadoop-0.20/lib/hadoop-fairschedule
>>>>> r-0.20.2-cdh3u4.jar:/usr/lib/hadoop-0.20/lib/jackson-core-asl-1.5.2.jar:/usr
>>>>> /lib/hadoop-0.20/lib/mockito-all-1.8.2.jar:/usr/lib/hadoop-0.20/lib/commons-
>>>>> el-1.0.jar:/usr/lib/hadoop-0.20/lib/commons-logging-api-1.0.4.jar:/usr/lib/h
>>>>> adoop-0.2
>>>>>
>>>>> 0/lib/commons-net-3.1.jar:/usr/lib/hadoop-0.20/lib/commons-cli-1.2.jar:/usr/
>>>>> lib/hadoop-0.20/lib/aspectjrt-1.6.5.jar:/usr/lib/hadoop-0.20/lib/core-3.1.1.
>>>>> jar:/usr/lib/hadoop-0.20/lib/hsqldb-1.8.0.10.jar:/usr/lib/hadoop-0.20/lib/or
>>>>> o-2.0.8.jar:/usr/lib/zookeeper/zookeeper.jar:/usr/lib/zookeeper/zookeeper-3.
>>>>> 3.5-cdh3u4.jar:/usr/lib/zookeeper/lib/log4j-1.2.15.jar:/usr/lib/zookeeper/li
>>>>> b/jline-0.9.94.jar::/usr/lib/hadoop-0.20/conf:/usr/lib/hadoop-0.20/hadoop-co
>>>>> re-0.20.2-cdh3u4.jar:/usr/lib/hadoop-0.20/lib/ant-contrib-1.0b3.jar:/usr/lib
>>>>> /hadoop-0.20/lib/aspectjrt-1.6.5.jar:/usr/lib/hadoop-0.20/lib/aspectjtools-1
>>>>> .6.5.jar:/usr/lib/hadoop-0.20/lib/commons-cli-1.2.jar:/usr/lib/hadoop-0.20/l
>>>>> ib/commons-codec-1.4.jar:/usr/lib/hadoop-0.20/lib/commons-daemon-1.0.1.jar:/
>>>>> usr/lib/hadoop-0.20/lib/commons-el-1.0.jar:/usr/lib/hadoop-0.20/lib/commons-
>>>>> httpclient-3.1.jar:/usr/lib/hadoop-0.20/lib/commons-lang-2.4.jar:/usr/lib/ha
>>>>> doop-0.20/lib/commons-logging-1.0.4.jar:/usr/lib/hadoop-0.20/lib/commons-log
>>>>> ging-api-
>>>>>
>>>>> 1.0.4.jar:/usr/lib/hadoop-0.20/lib/commons-net-3.1.jar:/usr/lib/hadoop-0.20/
>>>>> lib/core-3.1.1.jar:/usr/lib/hadoop-0.20/lib/guava-r09-jarjar.jar:/usr/lib/ha
>>>>> doop-0.20/lib/hadoop-fairscheduler-0.20.2-cdh3u4.jar:/usr/lib/hadoop-0.20/li
>>>>> b/hsqldb-1.8.0.10.jar:/usr/lib/hadoop-0.20/lib/jackson-core-asl-1.5.2.jar:/u
>>>>> sr/lib/hadoop-0.20/lib/jackson-mapper-asl-1.5.2.jar:/usr/lib/hadoop-0.20/lib
>>>>> /jasper-compiler-5.5.12.jar:/usr/lib/hadoop-0.20/lib/jasper-runtime-5.5.12.j
>>>>> ar:/usr/lib/hadoop-0.20/lib/jets3t-0.6.1.jar:/usr/lib/hadoop-0.20/lib/jetty-
>>>>> 6.1.26.cloudera.1.jar:/usr/lib/hadoop-0.20/lib/jetty-servlet-tester-6.1.26.c
>>>>> loudera.1.jar:/usr/lib/hadoop-0.20/lib/jetty-util-6.1.26.cloudera.1.jar:/usr
>>>>> /lib/hadoop-0.20/lib/jsch-0.1.42.jar:/usr/lib/hadoop-0.20/lib/junit-4.5.jar:
>>>>> /usr/lib/hadoop-0.20/lib/kfs-0.2.2.jar:/usr/lib/hadoop-0.20/lib/log4j-1.2.15
>>>>> .jar:/usr/lib/hadoop-0.20/lib/mockito-all-1.8.2.jar:/usr/lib/hadoop-0.20/lib
>>>>> /oro-2.0.8.jar:/usr/lib/hadoop-0.20/lib/servlet-api-2.5-20081211.jar:/usr/li
>>>>> b/hadoop-
>>>>>  0.20/lib/servlet-api-2.5-6.1.14.jar:/usr/lib/hadoop-0.20/lib/slf4j-api-1.4.
>>>>> 3.jar:/usr/lib/hadoop-0.20/lib/slf4j-log4j12-1.4.3.jar:/usr/lib/hadoop-0.20/
>>>>> lib/xmlenc-0.52.jar12/05/12 14:42:14 INFO zookeeper.ZooKeeper: Client
>>>>> environment:java.library.path=/usr/lib/hadoop-0.20/lib/native/Linux-amd64-64
>>>>> :/usr/lib/hbase/bin/../lib/native/Linux-amd64-6412/05/12 14:42:14 INFO
>>>>> zookeeper.ZooKeeper: Client environment:java.io.tmpdir=/tmp12/05/12 14:42:14
>>>>> INFO zookeeper.ZooKeeper: Client environment:java.compiler=<NA>12/05/12
>>>>> 14:42:14 INFO zookeeper.ZooKeeper: Client environment:os.name=Linux12/05/12
>>>>> 14:42:14 INFO zookeeper.ZooKeeper: Client environment:os.arch=amd6412/05/12
>>>>> 14:42:14 INFO zookeeper.ZooKeeper: Client
>>>>> environment:os.version=2.6.35-22-server12/05/12 14:42:14 INFO
>>>>> zookeeper.ZooKeeper: Client environment:user.name=dalia12/05/12 14:42:14
>>>>> INFO zookeeper.ZooKeeper: Client environment:user.home=/home/dalia12/05/12
>>>>> 14:42:14 INFO zookeeper.ZooKeeper: Client environment:user.dir=/home/dalia12
>>>>>  /05/12 14:42:14 INFO zookeeper.ZooKeeper: Initiating client connection,
>>>>> connectString=localhost:2181 sessionTimeout=180000
>>>>> watcher=regionserver:6002012/05/12 14:42:14 INFO zookeeper.ClientCnxn:
>>>>> Opening socket connection to server localhost/0:0:0:0:0:0:0:1:218112/05/12
>>>>> 14:42:14 WARN zookeeper.ClientCnxn: Session 0x0 for server null, unexpected
>>>>> error, closing socket connection and attempting
>>>>> reconnectjava.net.ConnectException: Connection refused      at
>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)    at
>>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)      at
>>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
>>>>> 14:42:14 INFO zookeeper.ClientCnxn: Opening socket connection to server
>>>>> localhost/127.0.0.1:218112/05/12 14:42:14 WARN zookeeper.ClientCnxn: Session
>>>>> 0x0 for server null, unexpected error, closing socket connection and
>>>>> attempting reconnectjava.net.ConnectException: Connection refused   at
>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native
>>>>>   Method)   at
>>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)      at
>>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
>>>>> 14:42:15 INFO ipc.Client: Retrying connect to server:
>>>>> namenode/10.0.2.3:8020. Already tried 0 time(s).12/05/12 14:42:16 INFO
>>>>> zookeeper.ClientCnxn: Opening socket connection to server localhost/0:0:0:0:
>>>>> 0:0:0:1:218112/05/12 14:42:16 WARN zookeeper.ClientCnxn: Session 0x0 for
>>>>> server null, unexpected error, closing socket connection and attempting
>>>>> reconnectjava.net.ConnectException: Connection refused      at
>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)    at
>>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)      at
>>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
>>>>> 14:42:16 INFO ipc.Client: Retrying connect to server:
>>>>> namenode/10.0.2.3:8020. Already tried 1 time(s).12/05/12 14:42:16 INFO
>>>>> zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.
>>>>> 1:218112/05/12 14:
>>>>>  42:16 WARN zookeeper.ClientCnxn: Session 0x0 for server null, unexpected
>>>>> error, closing socket connection and attempting
>>>>> reconnectjava.net.ConnectException: Connection refused      at
>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)    at
>>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)      at
>>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
>>>>> 14:42:17 INFO ipc.Client: Retrying connect to server:
>>>>> namenode/10.0.2.3:8020. Already tried 2 time(s).12/05/12 14:42:18 INFO
>>>>> ipc.Client: Retrying connect to server: namenode/10.0.2.3:8020. Already
>>>>> tried 3 time(s).12/05/12 14:42:18 INFO zookeeper.ClientCnxn: Opening socket
>>>>> connection to server localhost/0:0:0:0:0:0:0:1:218112/05/12 14:42:18 WARN
>>>>> zookeeper.ClientCnxn: Session 0x0 for server null, unexpected error, closing
>>>>> socket connection and attempting reconnectjava.net.ConnectException:
>>>>> Connection refused  at sun.nio.ch.SocketChannelImpl.checkConnect(Native
>>>>> Method)     at sun.nio.ch.SocketChannelImpl.fin
>>>>>  ishConnect(SocketChannelImpl.java:567)     at
>>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
>>>>> 14:42:18 INFO zookeeper.ClientCnxn: Opening socket connection to server
>>>>> localhost/127.0.0.1:218112/05/12 14:42:18 WARN zookeeper.ClientCnxn: Session
>>>>> 0x0 for server null, unexpected error, closing socket connection and
>>>>> attempting reconnectjava.net.ConnectException: Connection refused   at
>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)    at
>>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)      at
>>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
>>>>> 14:42:19 INFO ipc.Client: Retrying connect to server:
>>>>> namenode/10.0.2.3:8020. Already tried 4 time(s).12/05/12 14:42:19 INFO
>>>>> zookeeper.ClientCnxn: Opening socket connection to server localhost/0:0:0:0:
>>>>> 0:0:0:1:218112/05/12 14:42:19 WARN zookeeper.ClientCnxn: Session 0x0 for
>>>>> server null, unexpected error, closing socket connection and attempting
>>>>> reconnectjava.net.ConnectException
>>>>>  : Connection refused       at sun.nio.ch.SocketChannelImpl.checkConnect(Native
>>>>> Method)     at
>>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)      at
>>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
>>>>> 14:42:20 INFO ipc.Client: Retrying connect to server:
>>>>> namenode/10.0.2.3:8020. Already tried 5 time(s).12/05/12 14:42:20 INFO
>>>>> zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.
>>>>> 1:218112/05/12 14:42:20 WARN zookeeper.ClientCnxn: Session 0x0 for server
>>>>> null, unexpected error, closing socket connection and attempting
>>>>> reconnectjava.net.ConnectException: Connection refused      at
>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)    at
>>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)      at
>>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
>>>>> 14:42:21 INFO ipc.Client: Retrying connect to server:
>>>>> namenode/10.0.2.3:8020. Already tried 6 time(s).12/05/12 14:42:22 INFO
>>>>> zookeeper.ClientCnxn: Openin
>>>>>  g socket connection to server localhost/0:0:0:0:0:0:0:1:218112/05/12 14:42:
>>>>> 22 WARN zookeeper.ClientCnxn: Session 0x0 for server null, unexpected error,
>>>>> closing socket connection and attempting reconnectjava.net.ConnectException:
>>>>> Connection refused  at sun.nio.ch.SocketChannelImpl.checkConnect(Native
>>>>> Method)     at
>>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)      at
>>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
>>>>> 14:42:22 INFO ipc.Client: Retrying connect to server:
>>>>> namenode/10.0.2.3:8020. Already tried 7 time(s).12/05/12 14:42:22 INFO
>>>>> zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.
>>>>> 1:218112/05/12 14:42:22 WARN zookeeper.ClientCnxn: Session 0x0 for server
>>>>> null, unexpected error, closing socket connection and attempting
>>>>> reconnectjava.net.ConnectException: Connection refused      at
>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)    at
>>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)      at
>>>>> org
>>>>>  .apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
>>>>> 14:42:23 INFO ipc.Client: Retrying connect to server:
>>>>> namenode/10.0.2.3:8020. Already tried 8 time(s).12/05/12 14:42:24 INFO
>>>>> ipc.Client: Retrying connect to server: namenode/10.0.2.3:8020. Already
>>>>> tried 9 time(s).Exception in thread "main" java.net.ConnectException: Call
>>>>> to namenode/10.0.2.3:8020 failed on connection exception:
>>>>> java.net.ConnectException: Connection refused       at
>>>>> org.apache.hadoop.ipc.Client.wrapException(Client.java:1134)        at
>>>>> org.apache.hadoop.ipc.Client.call(Client.java:1110) at
>>>>> org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:226)      at
>>>>> $Proxy5.getProtocolVersion(Unknown Source)  at
>>>>> org.apache.hadoop.ipc.RPC.getProxy(RPC.java:398)    at
>>>>> org.apache.hadoop.ipc.RPC.getProxy(RPC.java:384)    at
>>>>> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:129)      at
>>>>> org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:255) at
>>>>> org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:217) at
>>>>> org.apache.hadoop
>>>>>  .hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89)      at
>>>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1563)      at
>>>>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:67)      at
>>>>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:1597)     at
>>>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1579)     at
>>>>> org.apache.hadoop.fs.FileSystem.get(FileSystem.java:228)    at
>>>>> org.apache.hadoop.fs.FileSystem.get(FileSystem.java:111)    at
>>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.startRegionServer(HRegion
>>>>> Server.java:2785)   at
>>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.startRegionServer(HRegion
>>>>> Server.java:2768)   at
>>>>> org.apache.hadoop.hbase.regionserver.HRegionServerCommandLine.start(HRegionS
>>>>> erverCommandLine.java:61)   at
>>>>> org.apache.hadoop.hbase.regionserver.HRegionServerCommandLine.run(HRegionSer
>>>>> verCommandLine.java:75)     at org.apache.hadoop.util.ToolRunner.run(ToolRunner.
>>>>> java:65)    at
>>>>> org.apache.hadoop.hbase.util.ServerCommandLine.doMain(Serve
>>>>>  rCommandLine.java:76)      at
>>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.main(HRegionServer.java:2
>>>>> 829)Caused by: java.net.ConnectException: Connection refused        at
>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)    at
>>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)      at
>>>>> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:2
>>>>> 06) at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:429)    at
>>>>> org.apache.hadoop.net.NetUtils.connect(NetUtils.java:394)   at
>>>>> org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:425)
>>>>> at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:532)
>>>>> at org.apache.hadoop.ipc.Client$Connection.access$2300(Client.java:210)     at
>>>>> org.apache.hadoop.ipc.Client.getConnection(Client.java:1247)        at
>>>>> org.apache.hadoop.ipc.Client.call(Client.java:1078) ... 21 more12/05/12
>>>>> 14:42:24 INFO zookeeper.ClientCnxn: Opening socket connection to server
>>>>> localhost/0:0:0:0:0:0:0:1:218112/05/12 14:42:24 WARN zookeeper
>>>>>  .ClientCnxn: Session 0x0 for server null, unexpected error, closing socket
>>>>> connection and attempting reconnectjava.net.ConnectException: Connection
>>>>> refused     at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)     at
>>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)      at
>>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12
>>>>> 14:42:25 INFO zookeeper.ClientCnxn: Opening socket connection to server
>>>>> localhost/127.0.0.1:218112/05/12 14:42:25 INFO zookeeper.ZooKeeper: Session:
>>>>> 0x0 closed12/05/12 14:42:25 INFO zookeeper.ClientCnxn: EventThread shut
>>>>> down12/05/12 14:42:25 INFO ipc.HBaseServer: Stopping server on 6002012/05/12
>>>>> 14:42:25 FATAL regionserver.HRegionServer: ABORTING region server
>>>>> serverName=datanode2,60020,1336826533870, load=(requests=0, regions=0,
>>>>> usedHeap=0, maxHeap=0): Initialization of RS failed.  Hence aborting RS.org.
>>>>> apache.hadoop.hbase.ZooKeeperConnectionException: HBase is able to connect
>>>>> to ZooKeeper but the connection closes im
>>>>>  mediately. This could be a sign that the server has too many connections
>>>>> (30 is the default). Consider inspecting your ZK server logs for that error
>>>>> and then make sure you are reusing HBaseConfiguration as often as you can.
>>>>> See HTable's javadoc for more information.  at
>>>>> org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher.<init>(ZooKeeperWatcher.j
>>>>> ava:160)    at
>>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.initializeZooKeeper(HRegi
>>>>> onServer.java:489)  at
>>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.initialize(HRegionServer.
>>>>> java:465)   at
>>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:56
>>>>> 4)  at java.lang.Thread.run(Thread.java:662)Caused by:
>>>>> org.apache.zookeeper.KeeperException$ConnectionLossException:
>>>>> KeeperErrorCode = ConnectionLoss for /hbase at
>>>>> org.apache.zookeeper.KeeperException.create(KeeperException.java:90)        at
>>>>> org.apache.zookeeper.KeeperException.create(KeeperException.java:42)        at
>>>>> org.apache.zookeeper.ZooKeeper.exists(ZooKeeper.java:815)   at org.apac
>>>>>  he.zookeeper.ZooKeeper.exists(ZooKeeper.java:843)  at
>>>>> org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndFailSilent(ZKUtil.java:930
>>>>> )   at
>>>>> org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher.<init>(ZooKeeperWatcher.j
>>>>> ava:138)    ... 4 more12/05/12 14:42:25 INFO regionserver.HRegionServer:
>>>>> STOPPED: Initialization of RS failed.  Hence aborting RS.Exception in thread
>>>>> "regionserver60020" java.lang.NullPointerException  at
>>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:63
>>>>> 3)  at java.lang.Thread.run(Thread.java:662)
>>>>> So any help???
>>>>>> Date: Sat, 12 May 2012 20:22:03 +0530
>>>>>> Subject: Re: Important "Undefined Error"
>>>>>> From: dwivedishashwat@gmail.com
>>>>>> To: user@hbase.apache.org
>>>>>>
>>>>>> you can turn off hadoop safe mode uisng *hadoop dfsadmin -safemode
>>>>>> leave*
>>>>>>
>>>>>> On Sat, May 12, 2012 at 8:15 PM, shashwat shriparv <
>>>>>> dwivedishashwat@gmail.com> wrote:
>>>>>>
>>>>>>> First thing copy core-site.xml, dfs xml from hadoop conf directory
>>>>>>> to hbase conf dirctory. and turn of hadoop save mode and then try...
>>>>>>>
>>>>>>>
>>>>>>> On Sat, May 12, 2012 at 6:27 PM, Harsh J <harsh@cloudera.com> wrote:
>>>>>>>
>>>>>>>> Dalia,
>>>>>>>>
>>>>>>>> Is your NameNode running fine? The issue is that HBase Master has
>>>>>>>> been asked to talk to HDFS, but it can't connect to the HDFS
>>>>>>>> NameNode. Does "hadoop dfs -touchz foobar" pass or fail with similar
>>>>> retry issues?
>>>>>>>>
>>>>>>>> What's your fs.default.name's value in Hadoop's core-site.xml? And
>>>>>>>> whats the output of that fixed host command I'd posted before?
>>>>>>>>
>>>>>>>> On Sat, May 12, 2012 at 6:06 PM, Dalia Sobhy
>>>>>>>> <dalia.mohsobhy@hotmail.com>
>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Dear Harsh
>>>>>>>>> When I run $hbase master start
>>>>>>>>> I found the following errors:12/05/12 08:32:42 INFO
>>>>>>>> ipc.HBaseRpcMetrics: Initializing RPC Metrics with
>>>>>>>> hostName=HMaster,
>>>>>>>> port=6000012/05/12 08:32:42 INFO security.UserGroupInformation:
>>>>>>>> JAAS Configuration already set up for Hadoop, not
>>>>>>>> re-installing.12/05/12
>>>>>>>> 08:32:42 INFO ipc.HBaseServer: IPC Server Responder:
>>>>>>>> starting12/05/12
>>>>>>>> 08:32:42 INFO ipc.HBaseServer: IPC Server listener on 60000:
>>>>>>>> starting12/05/12 08:32:42 INFO ipc.HBaseServer: IPC Server handler
>>>>>>>> 0 on
>>>>>>>> 60000: starting12/05/12 08:32:42 INFO ipc.HBaseServer: IPC Server
>>>>>>>> handler 1 on 60000: starting12/05/12 08:32:42 INFO ipc.HBaseServer:
>>>>>>>> IPC Server handler 2 on 60000: starting12/05/12 08:32:42 INFO
>>>>>>>> ipc.HBaseServer: IPC Server handler 3 on 60000: starting12/05/12 08:32:
>>>>> 42 INFO ipc.HBaseServer:
>>>>>>>> IPC Server handler 5 on 60000: starting12/05/12 08:32:42 INFO
>>>>>>>> ipc.HBaseServer: IPC Server handler 4 on 60000: starting12/05/12
>>>>>>>> 08:32:42 INFO ipc.HBaseServer: IPC Server handler 7 on 60000:
>>>>>>>> starting12/05/12
>>>>>>>> 08:32:42 INFO ipc.HBaseServer: IPC Serv
>>>>>>>>>  er handler 6 on 60000: starting12/05/12 08:32:42 INFO
>>>>> ipc.HBaseServer:
>>>>>>>> IPC Server handler 8 on 60000: starting12/05/12 08:32:42 INFO
>>>>>>>> ipc.HBaseServer: IPC Server handler 9 on 60000: starting12/05/12
>>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client
>>>>>>>> environment:zookeeper.version=3.3.5-cdh3u4--1, built on 05/07/2012
>>>>>>>> 21:12
>>>>>>>> GMT12/05/12 08:32:42 INFO zookeeper.ZooKeeper: Client environment:
>>>>>>>> host.name=namenode12/05/12 08:32:42 INFO zookeeper.ZooKeeper:
>>>>>>>> Client
>>>>>>>> environment:java.version=1.6.0_3012/05/12 08:32:42 INFO
>>>>>>>> zookeeper.ZooKeeper: Client environment:java.vendor=Sun
>>>>>>>> Microsystems
>>>>>>>> Inc.12/05/12 08:32:42 INFO zookeeper.ZooKeeper: Client
>>>>>>>> environment:java.home=/usr/lib/jvm/java-6-sun-1.6.0.30/jre12/05/12
>>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client
>>>>>>>> environment:java.class.path=/usr/lib/hbase/bin/../conf:/usr/lib/jvm
>>>>>>>> /java-6-sun/lib/tools.jar:/usr/lib/hbase/bin/..:/usr/lib/hbase/bin/
>>>>>>>> ../hbase-0.90.4-cdh3u3.jar:/usr/lib/hbase/bin/../hbase-0.90.4-cdh3u
>>>>>>>> 3-tests.jar:/usr/lib/hbase/bin/../lib/activation-1.1.jar:/u
>>>>>>>>>
>>>>>>>>
>>>>>>>> sr/lib/hbase/bin/../lib/asm-3.1.jar:/usr/lib/hbase/bin/../lib/avro-
>>>>>>>> 1.5.4.jar:/usr/lib/hbase/bin/../lib/avro-ipc-1.5.4.jar:/usr/lib/hba
>>>>>>>> se/bin/../lib/commons-cli-1.2.jar:/usr/lib/hbase/bin/../lib/commons
>>>>>>>> -codec-1.4.jar:/usr/lib/hbase/bin/../lib/commons-el-1.0.jar:/usr/li
>>>>>>>> b/hbase/bin/../lib/commons-httpclient-3.1.jar:/usr/lib/hbase/bin/..
>>>>>>>> /lib/commons-lang-2.5.jar:/usr/lib/hbase/bin/../lib/commons-logging
>>>>>>>> -1.1.1.jar:/usr/lib/hbase/bin/../lib/commons-net-1.4.1.jar:/usr/lib
>>>>>>>> /hbase/bin/../lib/core-3.1.1.jar:/usr/lib/hbase/bin/../lib/guava-r0
>>>>>>>> 6.jar:/usr/lib/hbase/bin/../lib/guava-r09-jarjar.jar:/usr/lib/hbase
>>>>>>>> /bin/../lib/hadoop-core.jar:/usr/lib/hbase/bin/../lib/jackson-core-
>>>>>>>> asl-1.5.2.jar:/usr/lib/hbase/bin/../lib/jackson-jaxrs-1.5.5.jar:/us
>>>>>>>> r/lib/hbase/bin/../lib/jackson-mapper-asl-1.5.2.jar:/usr/lib/hbase/
>>>>>>>> bin/../lib/jackson-xc-1.5.5.jar:/usr/lib/hbase/bin/../lib/jamon-run
>>>>>>>> time-2.3.1.jar:/usr/lib/hbase/bin/../lib/jasper-compiler-5.5.23.jar
>>>>>>>> :/usr/lib/hbase/bin/../lib/jasper-runtime-5.5.23.jar:/
>>>>>  usr/l
>>>>>>>>>
>>>>>>>>
>>>>>>>> ib/hbase/bin/../lib/jaxb-api-2.1.jar:/usr/lib/hbase/bin/../lib/jaxb
>>>>>>>> -impl-2.1.12.jar:/usr/lib/hbase/bin/../lib/jersey-core-1.4.jar:/usr
>>>>>>>> /lib/hbase/bin/../lib/jersey-json-1.4.jar:/usr/lib/hbase/bin/../lib
>>>>>>>> /jersey-server-1.4.jar:/usr/lib/hbase/bin/../lib/jettison-1.1.jar:/
>>>>>>>> usr/lib/hbase/bin/../lib/jetty-6.1.26.jar:/usr/lib/hbase/bin/../lib
>>>>>>>> /jetty-util-6.1.26.jar:/usr/lib/hbase/bin/../lib/jruby-complete-1.6
>>>>>>>> .0.jar:/usr/lib/hbase/bin/../lib/jsp-2.1-6.1.14.jar:/usr/lib/hbase/
>>>>>>>> bin/../lib/jsp-api-2.1-6.1.14.jar:/usr/lib/hbase/bin/../lib/jsp-api
>>>>>>>> -2.1.jar:/usr/lib/hbase/bin/../lib/jsr311-api-1.1.1.jar:/usr/lib/hb
>>>>>>>> ase/bin/../lib/log4j-1.2.16.jar:/usr/lib/hbase/bin/../lib/netty-3.2
>>>>>>>> .4.Final.jar:/usr/lib/hbase/bin/../lib/protobuf-java-2.3.0.jar:/usr
>>>>>>>> /lib/hbase/bin/../lib/servlet-api-2.5-6.1.14.jar:/usr/lib/hbase/bin
>>>>>>>> /../lib/servlet-api-2.5.jar:/usr/lib/hbase/bin/../lib/slf4j-api-1.5
>>>>>>>> .8.jar:/usr/lib/hbase/bin/../lib/slf4j-log4j12-1.5.8.jar:/usr/lib/h
>>>>>>>> base/bin/../lib/snappy-java-1.0.3.2.jar:/usr/lib/hbase
>>>>>  /bin/
>>>>>>>>>
>>>>>>>>
>>>>>>>> ../lib/stax-api-1.0.1.jar:/usr/lib/hbase/bin/../lib/thrift-0.2.0.ja
>>>>>>>> r:/usr/lib/hbase/bin/../lib/velocity-1.5.jar:/usr/lib/hbase/bin/../
>>>>>>>> lib/xmlenc-0.52.jar:/usr/lib/hbase/bin/../lib/zookeeper.jar:/etc/zo
>>>>>>>> okeeper:/etc/hadoop-0.20/conf:/usr/lib/hadoop-0.20/hadoop-examples.
>>>>>>>> jar:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u3-core.jar:/usr/lib/had
>>>>>>>> oop-0.20/hadoop-0.20.2-cdh3u3-ant.jar:/usr/lib/hadoop-0.20/hadoop-c
>>>>>>>> ore-0.20.2-cdh3u3.jar:/usr/lib/hadoop-0.20/hadoop-test.jar:/usr/lib
>>>>>>>> /hadoop-0.20/hadoop-ant-0.20.2-cdh3u3.jar:/usr/lib/hadoop-0.20/hado
>>>>>>>> op-tools.jar:/usr/lib/hadoop-0.20/hadoop-tools-0.20.2-cdh3u3.jar:/u
>>>>>>>> sr/lib/hadoop-0.20/hadoop-test-0.20.2-cdh3u3.jar:/usr/lib/hadoop-0.
>>>>>>>> 20/hadoop-core.jar:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u3-exampl
>>>>>>>> es.jar:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u3-test.jar:/usr/lib/
>>>>>>>> hadoop-0.20/hadoop-ant.jar:/usr/lib/hadoop-0.20/hadoop-examples-0.2
>>>>>>>> 0.2-cdh3u3.jar:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u3-tools.jar:
>>>>>>>> /usr/lib/hadoop-0.20/lib/jasper-runtime-5.5.12.jar:/us
>>>>>  r/lib
>>>>>>>>>
>>>>>>>>
>>>>>>>> /hadoop-0.20/lib/hsqldb-1.8.0.10.jar:/usr/lib/hadoop-0.20/lib/jacks
>>>>>>>> on-mapper-asl-1.5.2.jar:/usr/lib/hadoop-0.20/lib/jets3t-0.6.1.jar:/
>>>>>>>> usr/lib/hadoop-0.20/lib/jetty-servlet-tester-6.1.26.cloudera.1.jar:
>>>>>>>> /usr/lib/hadoop-0.20/lib/jackson-core-asl-1.5.2.jar:/usr/lib/hadoop
>>>>>>>> -0.20/lib/oro-2.0.8.jar:/usr/lib/hadoop-0.20/lib/ant-contrib-1.0b3.
>>>>>>>> jar:/usr/lib/hadoop-0.20/lib/commons-daemon-1.0.1.jar:/usr/lib/hado
>>>>>>>> op-0.20/lib/mockito-all-1.8.2.jar:/usr/lib/hadoop-0.20/lib/aspectjr
>>>>>>>> t-1.6.5.jar:/usr/lib/hadoop-0.20/lib/commons-lang-2.4.jar:/usr/lib/
>>>>>>>> hadoop-0.20/lib/junit-4.5.jar:/usr/lib/hadoop-0.20/lib/commons-code
>>>>>>>> c-1.4.jar:/usr/lib/hadoop-0.20/lib/servlet-api-2.5-6.1.14.jar:/usr/
>>>>>>>> lib/hadoop-0.20/lib/log4j-1.2.15.jar:/usr/lib/hadoop-0.20/lib/jsch-
>>>>>>>> 0.1.42.jar:/usr/lib/hadoop-0.20/lib/core-3.1.1.jar:/usr/lib/hadoop-
>>>>>>>> 0.20/lib/jetty-6.1.26.cloudera.1.jar:/usr/lib/hadoop-0.20/lib/commo
>>>>>>>> ns-logging-1.0.4.jar:/usr/lib/hadoop-0.20/lib/jetty-util-6.1.26.clo
>>>>>>>> udera.1.jar:/usr/lib/hadoop-0.20/lib/servlet-api-2.5-2
>>>>>  00812
>>>>>>>>>
>>>>>>>>
>>>>>>>> 11.jar:/usr/lib/hadoop-0.20/lib/jasper-compiler-5.5.12.jar:/usr/lib
>>>>>>>> /hadoop-0.20/lib/kfs-0.2.2.jar:/usr/lib/hadoop-0.20/lib/commons-cli
>>>>>>>> -1.2.jar:/usr/lib/hadoop-0.20/lib/commons-net-1.4.1.jar:/usr/lib/ha
>>>>>>>> doop-0.20/lib/commons-httpclient-3.1.jar:/usr/lib/hadoop-0.20/lib/s
>>>>>>>> lf4j-api-1.4.3.jar:/usr/lib/hadoop-0.20/lib/xmlenc-0.52.jar:/usr/li
>>>>>>>> b/hadoop-0.20/lib/commons-logging-api-1.0.4.jar:/usr/lib/hadoop-0.2
>>>>>>>> 0/lib/commons-el-1.0.jar:/usr/lib/hadoop-0.20/lib/slf4j-log4j12-1.4
>>>>>>>> .3.jar:/usr/lib/hadoop-0.20/lib/aspectjtools-1.6.5.jar:/usr/lib/had
>>>>>>>> oop-0.20/lib/guava-r09-jarjar.jar:/usr/lib/hadoop-0.20/lib/hadoop-f
>>>>>>>> airscheduler-0.20.2-cdh3u3.jar:/usr/lib/zookeeper/zookeeper.jar:/us
>>>>>>>> r/lib/zookeeper/zookeeper-3.3.5-cdh3u4.jar:/usr/lib/zookeeper/lib/l
>>>>>>>> og4j-1.2.15.jar:/usr/lib/zookeeper/lib/jline-0.9.94.jar::/usr/lib/h
>>>>>>>> adoop-0.20/conf:/usr/lib/hadoop-0.20/hadoop-core-0.20.2-cdh3u3.jar:
>>>>>>>> /usr/lib/hadoop-0.20/lib/ant-contrib-1.0b3.jar:/usr/lib/hadoop-0.20
>>>>>>>> /lib/aspectjrt-1.6.5.jar:/usr/lib/hadoop-0.20/lib/aspe
>>>>>  ctjto
>>>>>>>>>
>>>>>>>>
>>>>>>>> ols-1.6.5.jar:/usr/lib/hadoop-0.20/lib/commons-cli-1.2.jar:/usr/lib
>>>>>>>> /hadoop-0.20/lib/commons-codec-1.4.jar:/usr/lib/hadoop-0.20/lib/com
>>>>>>>> mons-daemon-1.0.1.jar:/usr/lib/hadoop-0.20/lib/commons-el-1.0.jar:/
>>>>>>>> usr/lib/hadoop-0.20/lib/commons-httpclient-3.1.jar:/usr/lib/hadoop-
>>>>>>>> 0.20/lib/commons-lang-2.4.jar:/usr/lib/hadoop-0.20/lib/commons-logg
>>>>>>>> ing-1.0.4.jar:/usr/lib/hadoop-0.20/lib/commons-logging-api-1.0.4.ja
>>>>>>>> r:/usr/lib/hadoop-0.20/lib/commons-net-1.4.1.jar:/usr/lib/hadoop-0.
>>>>>>>> 20/lib/core-3.1.1.jar:/usr/lib/hadoop-0.20/lib/guava-r09-jarjar.jar
>>>>>>>> :/usr/lib/hadoop-0.20/lib/hadoop-fairscheduler-0.20.2-cdh3u3.jar:/u
>>>>>>>> sr/lib/hadoop-0.20/lib/hsqldb-1.8.0.10.jar:/usr/lib/hadoop-0.20/lib
>>>>>>>> /jackson-core-asl-1.5.2.jar:/usr/lib/hadoop-0.20/lib/jackson-mapper
>>>>>>>> -asl-1.5.2.jar:/usr/lib/hadoop-0.20/lib/jasper-compiler-5.5.12.jar:
>>>>>>>> /usr/lib/hadoop-0.20/lib/jasper-runtime-5.5.12.jar:/usr/lib/hadoop-
>>>>>>>> 0.20/lib/jets3t-0.6.1.jar:/usr/lib/hadoop-0.20/lib/jetty-6.1.26.clo
>>>>>>>> udera.1.jar:/usr/lib/hadoop-0.20/lib/jetty-servlet-tes
>>>>>  ter-6
>>>>>>>>>
>>>>>>>>
>>>>>>>> .1.26.cloudera.1.jar:/usr/lib/hadoop-0.20/lib/jetty-util-6.1.26.clo
>>>>>>>> udera.1.jar:/usr/lib/hadoop-0.20/lib/jsch-0.1.42.jar:/usr/lib/hadoo
>>>>>>>> p-0.20/lib/junit-4.5.jar:/usr/lib/hadoop-0.20/lib/kfs-0.2.2.jar:/us
>>>>>>>> r/lib/hadoop-0.20/lib/log4j-1.2.15.jar:/usr/lib/hadoop-0.20/lib/moc
>>>>>>>> kito-all-1.8.2.jar:/usr/lib/hadoop-0.20/lib/oro-2.0.8.jar:/usr/lib/
>>>>>>>> hadoop-0.20/lib/servlet-api-2.5-20081211.jar:/usr/lib/hadoop-0.20/l
>>>>>>>> ib/servlet-api-2.5-6.1.14.jar:/usr/lib/hadoop-0.20/lib/slf4j-api-1.
>>>>>>>> 4.3.jar:/usr/lib/hadoop-0.20/lib/slf4j-log4j12-1.4.3.jar:/usr/lib/h
>>>>>>>> adoop-0.20/lib/xmlenc-0.52.jar12/05/12
>>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client
>>>>>>>> environment:java.library.path=/usr/lib/hadoop-0.20/lib/native/Linux
>>>>>>>> -amd64-64:/usr/lib/hbase/bin/../lib/native/Linux-amd64-6412/05/12
>>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client
>>>>>>>> environment:java.io.tmpdir=/tmp12/05/12 08:32:42 INFO
>>>>> zookeeper.ZooKeeper:
>>>>>>>> Client environment:java.compiler=<NA>12/05/12 08:32:42 INFO
>>>>>>>> zookeeper.ZooKeeper: Client environment:os.name=Linux12/05/12
>>>>>>>> 08:32:42
>>>>>>>>>  INFO zookeeper.ZooKeeper: Client
>>>>>>>>> environment:os.arch=amd6412/05/12
>>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client
>>>>>>>> environment:os.version=2.6.35-22-server12/05/12 08:32:42 INFO
>>>>>>>> zookeeper.ZooKeeper: Client environment:user.name=dalia12/05/12
>>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client
>>>>>>>> environment:user.home=/home/dalia12/05/12
>>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client
>>>>>>>> environment:user.dir=/home/dalia12/05/12 08:32:42 INFO
>>>>> zookeeper.ZooKeeper:
>>>>>>>> Initiating client connection, connectString=namenode:2181
>>>>>>>> sessionTimeout=180000 watcher=master:6000012/05/12 08:32:42 INFO
>>>>>>>> zookeeper.ClientCnxn: Opening socket connection to server namenode/
>>>>>>>> 10.0.2.3:218112/05/12 08:32:42 INFO zookeeper.ClientCnxn: Socket
>>>>>>>> connection established to namenode/10.0.2.3:2181, initiating
>>>>>>>> session12/05/12 08:32:42 INFO zookeeper.ClientCnxn: Session
>>>>>>>> establishment complete on server namenode/10.0.2.3:2181, sessionid
>>>>>>>> = 0x13740bc4f70000c, negotiated timeout = 4000012/05/12 08:32:42
>>>>>>>> INFO
>>>>>>>> jvm.JvmMetrics: Initializing JVM Metrics with
>>>>>>>>>  processName=Master, sessionId=namenode:6000012/05/12 08:32:42
>>>>>>>>> INFO
>>>>>>>> hbase.metrics: MetricsString added: revision12/05/12 08:32:42 INFO
>>>>>>>> hbase.metrics: MetricsString added: hdfsUser12/05/12 08:32:42 INFO
>>>>>>>> hbase.metrics: MetricsString added: hdfsDate12/05/12 08:32:42 INFO
>>>>>>>> hbase.metrics: MetricsString added: hdfsUrl12/05/12 08:32:42 INFO
>>>>>>>> hbase.metrics: MetricsString added: date12/05/12 08:32:42 INFO
>>>>>>>> hbase.metrics: MetricsString added: hdfsRevision12/05/12 08:32:42
>>>>>>>> INFO
>>>>>>>> hbase.metrics: MetricsString added: user12/05/12 08:32:42 INFO
>>>>>>>> hbase.metrics: MetricsString added: hdfsVersion12/05/12 08:32:42
>>>>>>>> INFO
>>>>>>>> hbase.metrics: MetricsString added: url12/05/12 08:32:42 INFO
>>>>>>>> hbase.metrics: MetricsString added: version12/05/12 08:32:42 INFO
>>>>>>>> hbase.metrics: new MBeanInfo12/05/12 08:32:42 INFO hbase.metrics:
>>>>>>>> new
>>>>>>>> MBeanInfo12/05/12 08:32:42 INFO metrics.MasterMetrics:
>>>>>>>> Initialized12/05/12
>>>>>>>> 08:32:42 INFO master.ActiveMasterManager:
>>>>>>>> Master=namenode:6000012/05/12
>>>>>>>> 08:32:44 INFO ipc.Client: Retrying connect to serve
>>>>>>>>>  r: namenode/10.0.2.3:8020. Already tried 0 time(s).12/05/12
>>>>>>>>> 08:32:45
>>>>>>>> INFO ipc.Client: Retrying connect to server: namenode/10.0.2.3:8020.
>>>>>>>> Already tried 1 time(s).12/05/12 08:32:46 INFO ipc.Client: Retrying
>>>>>>>> connect to server: namenode/10.0.2.3:8020. Already tried 2
>>>>>>>> time(s).12/05/12
>>>>>>>> 08:32:47 INFO ipc.Client: Retrying connect to server: namenode/
>>>>>>>> 10.0.2.3:8020. Already tried 3 time(s).12/05/12 08:32:48 INFO
>>>>>>>> ipc.Client: Retrying connect to server: namenode/10.0.2.3:8020.
>>>>>>>> Already tried 4 time(s).12/05/12 08:32:49 INFO ipc.Client: Retrying
>>>>>>>> connect to
>>>>>>>> server: namenode/10.0.2.3:8020. Already tried 5 time(s).12/05/12
>>>>>>>> 08:32:50 INFO ipc.Client: Retrying connect to server: namenode/
>>>>>>>> 10.0.2.3:8020. Already tried 6 time(s).12/05/12 08:32:51 INFO
>>>>>>>> ipc.Client: Retrying connect to server: namenode/10.0.2.3:8020.
>>>>>>>> Already tried 7 time(s).12/05/12 08:32:52 INFO ipc.Client: Retrying
>>>>>>>> connect to
>>>>>>>> server: namenode/10.0.2.3:8020. Already tried 8 time(s).12/05/12
>>>>>>>> 08:32:53 INFO ipc.Client: Retrying connect to ser
>>>>>>>>>  ver: namenode/10.0.2.3:8020. Already tried 9 time(s).12/05/12
>>>>>>>> 08:32:53 FATAL master.HMaster: Unhandled exception. Starting
>>>>>>>> shutdown.java.net.ConnectException: Call to
>>>>> namenode/10.0.2.3:8020failed on connection exception:
>>>>> java.net.ConnectException: Connection
>>>>>>>> refused    at
>>>>> org.apache.hadoop.ipc.Client.wrapException(Client.java:1134)
>>>>>>>> at org.apache.hadoop.ipc.Client.call(Client.java:1110)  at
>>>>>>>> org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:226)       at
>>>>>>>> $Proxy6.getProtocolVersion(Unknown Source)   at
>>>>>>>> org.apache.hadoop.ipc.RPC.getProxy(RPC.java:398)     at
>>>>>>>> org.apache.hadoop.ipc.RPC.getProxy(RPC.java:384)     at
>>>>>>>> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:123)
>>>>>>>>   at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:246)
>>>>>>>> at
>>>>>>>> org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:208)  at
>>>>>>>>
>>>>> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSyste
>>>>> m.java:89)
>>>>>>>>       at
>>>>>>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1563)
>>>>>>>>   at org.apache.hadoop.fs.FileSystem.acc
>>>>>>>>>  ess$200(FileSystem.java:67)    at
>>>>>>>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:1597)
>>>>>>>>  at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1579)
>>>>> at
>>>>>>>> org.apache.hadoop.fs.FileSystem.get(FileSystem.java:228)     at
>>>>>>>> org.apache.hadoop.fs.Path.getFileSystem(Path.java:183)       at
>>>>>>>> org.apache.hadoop.hbase.util.FSUtils.getRootDir(FSUtils.java:364)    at
>>>>>>>>
>>>>> org.apache.hadoop.hbase.master.MasterFileSystem.<init>(MasterFileSystem.java
>>>>> :86)
>>>>>>>>     at
>>>>>>>>
>>>>> org.apache.hadoop.hbase.master.HMaster.finishInitialization(HMaster.java:360
>>>>> )
>>>>>>>>        at
>>>>>>>> org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:293)Caused by:
>>>>>>>> java.net.ConnectException: Connection refused at
>>>>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)     at
>>>>>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
>>>>>>>>   at
>>>>>>>>
>>>>> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:2
>>>>> 06)
>>>>>>>>      at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:408)    at
>>>>>>>>
>>>>> org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:425)
>>>>>>>>   a
>>>>>>>>>  t
>>>>>>>> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:
>>>>>>>> 532)  at
>>>>>>>> org.apache.hadoop.ipc.Client$Connection.access$2300(Client.java:210
>>>>>>>> ) at
>>>>>>>> org.apache.hadoop.ipc.Client.getConnection(Client.java:1247) at
>>>>>>>> org.apache.hadoop.ipc.Client.call(Client.java:1078)  ... 18
>>>>>>>> more12/05/12
>>>>>>>> 08:32:53 INFO master.HMaster: Aborting12/05/12 08:32:53 DEBUG
>>>>>>>> master.HMaster: Stopping service threads12/05/12 08:32:53 INFO
>>>>>>>> ipc.HBaseServer: Stopping server on 6000012/05/12 08:32:53 INFO
>>>>>>>> ipc.HBaseServer: IPC Server handler 5 on 60000: exiting12/05/12
>>>>>>>> 08:32:53 INFO ipc.HBaseServer: Stopping IPC Server listener on
>>>>>>>> 6000012/05/12
>>>>>>>> 08:32:53 INFO ipc.HBaseServer: IPC Server handler 1 on 60000:
>>>>>>>> exiting12/05/12 08:32:53 INFO ipc.HBaseServer: IPC Server handler 0
>>>>>>>> on
>>>>>>>> 60000: exiting12/05/12 08:32:53 INFO ipc.HBaseServer: IPC Server
>>>>>>>> handler 3 on 60000: exiting12/05/12 08:32:53 INFO ipc.HBaseServer:
>>>>>>>> IPC Server handler
>>>>>>>> 7 on 60000: exiting12/05/12 08:32:53 INFO ipc.HBaseServer: IPC
>>>>>>>> Server handler 9 on 60000: exiting1
>>>>>>>>>  2/05/12 08:32:53 INFO ipc.HBaseServer: IPC Server handler 6 on
>>>>> 60000:
>>>>>>>> exiting12/05/12 08:32:53 INFO ipc.HBaseServer: IPC Server handler 4
>>>>>>>> on
>>>>>>>> 60000: exiting12/05/12 08:32:53 INFO ipc.HBaseServer: IPC Server
>>>>>>>> handler 2 on 60000: exiting12/05/12 08:32:53 INFO ipc.HBaseServer:
>>>>>>>> IPC Server handler
>>>>>>>> 8 on 60000: exiting12/05/12 08:32:53 INFO ipc.HBaseServer: Stopping
>>>>>>>> IPC Server Responder12/05/12 08:32:53 INFO zookeeper.ZooKeeper:
>>>>> Session:
>>>>>>>> 0x13740bc4f70000c closed12/05/12 08:32:53 INFO zookeeper.ClientCnxn:
>>>>>>>> EventThread shut down12/05/12 08:32:53 INFO master.HMaster: HMaster
>>>>>>>> main thread exiting> From: harsh@cloudera.com
>>>>>>>>>> Date: Sat, 12 May 2012 17:28:29 +0530
>>>>>>>>>> Subject: Re: Important "Undefined Error"
>>>>>>>>>> To: user@hbase.apache.org
>>>>>>>>>>
>>>>>>>>>> Hi Dalia,
>>>>>>>>>>
>>>>>>>>>> On Sat, May 12, 2012 at 5:14 PM, Dalia Sobhy <
>>>>>>>> dalia.mohsobhy@hotmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>> Dear all,
>>>>>>>>>>> I have first a problem with Hbase I am trying to install it on
>>>>>>>>>>> a
>>>>>>>> distributed/multinode cluster..
>>>>>>>>>>> I am using the cloudera
>>>>>>>> https://ccp.cloudera.com/display/CDH4B2/HBase+Installation#HBaseIns
>>>>>>>> tallation-StartingtheHBaseMaster
>>>>>>>>>>> But when I write this command
>>>>>>>>>>> Creating the /hbase Directory in HDFS $sudo -u hdfs hadoop fs
>>>>>>>>>>> -mkdir
>>>>>>>> /hbase
>>>>>>>>>>> I get the following error:12/05/12 07:20:42 INFO
>>>>>>>> security.UserGroupInformation: JAAS Configuration already set up
>>>>>>>> for Hadoop, not re-installing.
>>>>>>>>>>
>>>>>>>>>> This is not an error and you shouldn't be worried. It is rather
>>>>>>>>>> a noisy INFO log that should be fixed (as a DEBUG level instead)
>>>>>>>>>> in subsequent releases (Are you using CDH3 or CDH4? IIRC only
>>>>>>>>>> CDH3u3 printed these, not in anything above that.)
>>>>>>>>>>
>>>>>>>>>>> 2. Another Aspect is when I start the hbase master it closes
>>>>>>>> automatically after a while.
>>>>>>>>>>
>>>>>>>>>> Could you post us your HMaster start->crash log? You can use a
>>>>>>>>>> service like pastebin.com to send us the output.
>>>>>>>>>>
>>>>>>>>>>> 3. Also this command is not working$host -v -t A
>>>>> `namenode`namenode:
>>>>>>>> command not found
>>>>>>>>>>
>>>>>>>>>> The right command is perhaps just:
>>>>>>>>>>
>>>>>>>>>> $host -v -t A `hostname`
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Harsh J
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Harsh J
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>>
>>>>>>>
>>>>>>> ∞
>>>>>>> Shashwat Shriparv
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>>
>>>>>>
>>>>>> ∞
>>>>>> Shashwat Shriparv
>>>>>
>>>>>
>>>>
>>>

Mime
View raw message