hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Muhammad Mudassar <mudassa...@gmail.com>
Subject Re: hbase connection issue
Date Wed, 15 Jul 2009 09:14:28 GMT
*I have started hadoop on single node before starting hbase and hadoop site
configurations are *


<configuration>
<property>
  <name>hadoop.tmp.dir</name>
  <value>/home/hadoop/Desktop/hadoop-datastore/hadoop-$hadoop</value>
  <description>A base for other temporary directories.</description>
</property>

<property>
  <name>fs.default.name</name>
  <value>hdfs://127.0.0.1:54310</value>
  <description>The name of the default file system.  A URI whose
  scheme and authority determine the FileSystem implementation.  The
  uri's scheme determines the config property (fs.SCHEME.impl) naming
  the FileSystem implementation class.  The uri's authority is used to
  determine the host, port, etc. for a filesystem.</description>
</property>

<property>
  <name>mapred.job.tracker</name>
  <value>127.0.0.1:54311</value>
  <description>The host and port that the MapReduce job tracker runs
  at.  If "local", then jobs are run in-process as a single map
  and reduce task.
  </description>
</property>

<property>
  <name>dfs.replication</name>
  <value>1</value>
  <description>Default block replication.
  The actual number of replications can be specified when the file is
created.
  The default is used if replication is not specified in create time.
  </description>
</property>

</configuration>



*and hbase site configuration are
*
<property>
      <name>hbase.rootdir</name>
      <value>hdfs://127.0.0.1:9000/hbase</value>
      <description>The directory shared by region servers.
      Should be fully-qualified to include the filesystem to use.
      E.g: hdfs://NAMENODE_SERVER:PORT/HBASE_ROOTDIR
      </description>
    </property>
      <property>
      <name>hbase.master</name>
      <value>local</value>
      <description>The host and port that the HBase master runs at.
      A value of 'local' runs the master and a regionserver in
      a single process.
      </description>
    </property>








*but after this i am still facing same problem I also tried with follwing
site configurations *



       <name>hbase.rootdir</name>
       <value>file:///var/hbase</value>
        <description>The directory shared by region servers.
       Should be fully-qualified to include the filesystem to use.
       E.g: hdfs://NAMENODE_SERVER:PORT/HBASE_ROOTDIR
       </description>
</property>


*but still not solved!*


On Tue, Jul 14, 2009 at 9:00 PM, <mrhoberto@gmail.com> wrote:

> Tuy
> Sent via BlackBerry from T-Mobile
>
> -----Original Message-----
> From: Ravindra Jain <sendtoravi@gmail.com>
>
> Date: Tue, 14 Jul 2009 20:21:01
> To: <hbase-user@hadoop.apache.org>
> Subject: Re: hbase connection issue
>
>
> Hi Muhammad,
>
> I faced exact same issue yesterday while I was trying to setup my
> first Hadoop/HBase cluster (Single-machine mode worked fine earlier).
> The solution lies in configuration in /etc/hosts file - first few
> lines in that file should look like following:
>
> 127.0.0.1       localhost
> 192.168.xx.xxx  master-hostname
> 192.168.xx.yyy  master-hostname.domain  master
> 192.168.xx.zzz      slave-hostname.domain       slave1
> ....
> ....
>
> Change it and restart networking using following command:
> sudo /etc/init.d/networking restart
>
> (I'm new to Linux networking and hence not sure if restart is required or
> not)
>
> Hopefully this should resolve your problem.
>
> Regards,
> Ravi..
>
>
>
> On Tue, Jul 14, 2009 at 8:05 PM, Muhammad Mudassar<mudassark7@gmail.com>
> wrote:
> > here is logs of master
> >
> >
> >
> > Tue Jul 14 20:28:20 PKST 2009 Starting master on mudassar-desktop
> > ulimit -n 1024
> > 2009-07-14 20:28:20,458 INFO org.apache.hadoop.hbase.master.HMaster:
> > vmName=Java HotSpot(TM) Server VM, vmVendor=Sun Microsystems Inc.,
> > vmVersion=14.0-b16
> > 2009-07-14 20:28:20,459 INFO org.apache.hadoop.hbase.master.HMaster:
> > vmInputArguments=[-Xmx1000m, -XX:+HeapDumpOnOutOfMemoryError,
> > -Dhbase.log.dir=/home/hadoop/Desktop/hbase-0.19.2/bin/../logs,
> > -Dhbase.log.file=hbase-hadoop-master-mudassar-desktop.log,
> > -Dhbase.home.dir=/home/hadoop/Desktop/hbase-0.19.2/bin/..,
> > -Dhbase.id.str=hadoop, -Dhbase.root.logger=INFO,DRFA,
> >
> -Djava.library.path=/home/hadoop/Desktop/hbase-0.19.2/bin/../lib/native/Linux-i386-32]
> > 2009-07-14 20:28:21,729 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: /127.0.0.1:60000. Already tried 0 time(s).
> > 2009-07-14 20:28:22,729 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: /127.0.0.1:60000. Already tried 1 time(s).
> > 2009-07-14 20:28:23,729 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: /127.0.0.1:60000. Already tried 2 time(s).
> > 2009-07-14 20:28:24,730 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: /127.0.0.1:60000. Already tried 3 time(s).
> > 2009-07-14 20:28:25,730 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: /127.0.0.1:60000. Already tried 4 time(s).
> > 2009-07-14 20:28:26,731 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: /127.0.0.1:60000. Already tried 5 time(s).
> > 2009-07-14 20:28:27,731 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: /127.0.0.1:60000. Already tried 6 time(s).
> > 2009-07-14 20:28:28,731 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: /127.0.0.1:60000. Already tried 7 time(s).
> > 2009-07-14 20:28:29,732 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: /127.0.0.1:60000. Already tried 8 time(s).
> > 2009-07-14 20:28:30,732 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: /127.0.0.1:60000. Already tried 9 time(s).
> > 2009-07-14 20:28:30,734 ERROR org.apache.hadoop.hbase.master.HMaster: Can
> > not start master
> > java.net.ConnectException: Call to /127.0.0.1:60000 failed on connection
> > exception: java.net.ConnectException: Connection refused
> >    at org.apache.hadoop.ipc.Client.wrapException(Client.java:724)
> >    at org.apache.hadoop.ipc.Client.call(Client.java:700)
> >    at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216)
> >    at $Proxy0.getProtocolVersion(Unknown Source)
> >    at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:348)
> >    at
> > org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:104)
> >    at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:176)
> >    at
> >
> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:75)
> >    at
> > org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1367)
> >    at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:56)
> >    at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1379)
> >    at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:215)
> >    at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:120)
> >    at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:186)
> >    at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:156)
> >    at
> >
> org.apache.hadoop.hbase.LocalHBaseCluster.<init>(LocalHBaseCluster.java:96)
> >    at
> >
> org.apache.hadoop.hbase.LocalHBaseCluster.<init>(LocalHBaseCluster.java:78)
> >    at org.apache.hadoop.hbase.master.HMaster.doMain(HMaster.java:1013)
> >    at org.apache.hadoop.hbase.master.HMaster.main(HMaster.java:1057)
> > Caused by: java.net.ConnectException: Connection refused
> >    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
> >    at
> > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:574)
> >    at sun.nio.ch.SocketAdaptor.connect(SocketAdaptor.java:100)
> >    at
> > org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:300)
> >    at
> org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:177)
> >    at org.apache.hadoop.ipc.Client.getConnection(Client.java:801)
> >    at org.apache.hadoop.ipc.Client.call(Client.java:686)
> >    ... 17 more
> >
> >
> >
> >
> >
> >
> >
> > On Tue, Jul 14, 2009 at 8:28 PM, Vaibhav Puranik <vpuranik@gmail.com>
> wrote:
> >
> >> Mhuhammad,
> >>
> >> Looks like your hbase master didn't start properly. You should check
> your
> >> master log.
> >>
> >> The master log will be in the logs directory. It will have more specific
> >> exception that can help you to find the real problem. If you couldn't
> solve
> >> it, paste the exception in the log here so that we can help you.
> >>
> >> Regards,
> >> Vaibhav
> >>
> >> On Tue, Jul 14, 2009 at 6:47 AM, Muhammad Mudassar <
> mudassark7@gmail.com
> >> >wrote:
> >>
> >> > Hi,
> >> >
> >> > I am running hbase on single node and my hbase-site seetings are as
> >> > follows:
> >> >
> >> >
> >> > <configuration>
> >> >  <property>
> >> >    <name>hbase.rootdir</name>
> >> >    <value>hdfs://127.0.0.1:9000/hbase</value>
> >> >    <description>The directory shared by region servers.
> >> >    Should be fully-qualified to include the filesystem to use.
> >> >    E.g: hdfs://NAMENODE_SERVER:PORT/HBASE_ROOTDIR
> >> >    </description>
> >> >  </property>
> >> >    <property>
> >> >    <name>hbase.master</name>
> >> >    <value>local</value>
> >> >    <description>The host and port that the HBase master runs at.
> >> >    A value of 'local' runs the master and a regionserver in
> >> >    a single process.
> >> >    </description>
> >> >  </property>
> >> >
> >> > </configuration>
> >> >
> >> >
> >> > After this when I created table in hbase shell it is saying trying to
> >> > connect to the server like:
> >> >
> >> > 09/07/14 19:41:03 INFO ipc.HBaseClass: Retrying connect to server:
> >> > localhost/127.0.0.1:60000. Already tried 0 time(s).
> >> > 09/07/14 19:41:04 INFO ipc.HBaseClass: Retrying connect to server:
> >> > localhost/127.0.0.1:60000. Already tried 1 time(s).
> >> > 09/07/14 19:41:05 INFO ipc.HBaseClass: Retrying connect to server:
> >> > localhost/127.0.0.1:60000. Already tried 2 time(s).
> >> > NativeException: org.apache.hadoop.hbase.MasterNotRunningException:
> >> > localhost:60000
> >> >    from org/apache/hadoop/hbase/client/HConnectionManager.java:239:in
> >> > `getMaster'
> >> >    from org/apache/hadoop/hbase/client/HBaseAdmin.java:70:in `<init>'
> >> >    from sun/reflect/NativeConstructorAccessorImpl.java:-2:in
> >> `newInstance0'
> >> >    from sun/reflect/NativeConstructorAccessorImpl.java:39:in
> >> `newInstance'
> >> >    from sun/reflect/DelegatingConstructorAccessorImpl.java:27:in
> >> > `newInstance'
> >> >
> >> >
> >> >
> >> > I required help to solve out it!
> >> >
> >> > waiting
> >> >
> >> >
> >> >
> >> > Regards
> >> >
> >> > Muhammad Mudassar
> >> >
> >>
> >
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message