hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Dhaya007 <mgdha...@gmail.com>
Subject Re: Examples for storing data in to Multi node cluster
Date Fri, 21 Dec 2007 11:03:12 GMT

Thanks for your reply,
I am new to hadoop if any think wrong please correct me ....
I already started looking on those websites and i configured a single/multi
node cluster using the link
I have followed the link but i am not able to start the haoop in multi node
The problems i am facing are as Follows:
1.I have configured master and slave nodes with ssh less pharase if try to
run the start-dfs.sh it prompt the password for master:slave machines.(I
have copied the .ssh/id_rsa.pub key of master in to slaves autherized_key

2.After giving password datanode,namenode,jobtracker,tasktraker started
successfully in master but datanode is started in slave.

3.Some time step 2 works and some time it says that permission denied.

4.I have checked the log file in the slave for datanode it says that
incompatible node, then i have formated the slave, master and start the dfs
by start-dfs.sh still i am getting the error

The host entry in etc/hosts are both master/slave

The hadoop-site.xml  for both master/slave
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

  <description>A base for other temporary directories.</description>
  <description>The name of the default file system.  A URI whose
  scheme and authority determine the FileSystem implementation.  The
  uri's scheme determines the config property (fs.SCHEME.impl) naming
  the FileSystem implementation class.  The uri's authority is used to
  determine the host, port, etc. for a filesystem.</description>
  <description>The host and port that the MapReduce job tracker runs
  at.  If "local", then jobs are run in-process as a single map
  and reduce task.
  <description>Default block replication.
  The actual number of replications can be specified when the file is
  The default is used if replication is not specified in create time.

  <description>As a rule of thumb, use 10x the number of slaves (i.e.,
number of tasktrackers).

  <description>As a rule of thumb, use 2x the number of slave processors
(i.e., number of tasktrackers).

Please help me to reslove the same. Or else provide any other tutorial for
multi node cluster setup.I am egarly waiting for the tutorials.


Khalil Honsali wrote:
> I think you are referring to storage on Parallel and/or Distributed File
> Systems, Hadoop is build on top of a Google FS like file system called
> All hadoop related info are on : http://lucene.apache.org/hadoop/
> Please start with this first:
> http://wiki.apache.org/lucene-hadoop/ImportantConcepts
> On 21/12/2007, Dhaya007 <mgdhayal@gmail.com> wrote:
>> I am Doing R&D on hadoop
>> My requirement is to store huge datas and retrive data by search,I have
>> searched on the web that Hadoop is the best solution.
>> If any one have this kind of Document (some examples to store the datas
>> in
>> multinode environment )
>> please share the document and help me on this
>> --
>> View this message in context:
>> http://www.nabble.com/Examples-for-storing-data-in-to-Multi-node-cluster-tp14450333p14450333.html
>> Sent from the Hadoop Users mailing list archive at Nabble.com.

View this message in context: http://www.nabble.com/Examples-for-storing-data-in-to-Multi-node-cluster-tp14450333p14453550.html
Sent from the Hadoop Users mailing list archive at Nabble.com.

View raw message