hadoop-hdfs-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Irfan Sayed <irfu.sa...@gmail.com>
Subject Re: about replication
Date Thu, 22 Aug 2013 20:12:15 GMT
ok. thanks
now, i need to start with all windows setup first as our product will be
based on windows
so, now, please tell me how to resolve the issue

datanode is not starting . please suggest

regards,
irfan



On Thu, Aug 22, 2013 at 7:56 PM, Mohammad Tariq <dontariq@gmail.com> wrote:

> It is possible. Theoretically Hadoop doesn't stop you from doing that. But
> it is not a very wise setup.
>
> Warm Regards,
> Tariq
> cloudfront.blogspot.com
>
>
> On Thu, Aug 22, 2013 at 5:01 PM, Irfan Sayed <irfu.sayed@gmail.com> wrote:
>
>> please suggest
>>
>> regards
>> irfan
>>
>>
>>
>> On Thu, Aug 22, 2013 at 11:49 AM, Irfan Sayed <irfu.sayed@gmail.com>wrote:
>>
>>> thanks.
>>> can i have setup like this :
>>> namenode will be on linux (flavour may be RHEL, CentOS, UBuntu etc)
>>> and datanodes are the combination of any OS (windows , linux , unix etc )
>>>
>>> however, my doubt is,  as the file systems of  both the systems (win and
>>> linux ) are different ,  datanodes of these systems can not be part of
>>> single cluster . i have to make windows cluster separate and UNIX cluster
>>> separate ?
>>>
>>> regards
>>>
>>>
>>>
>>> On Thu, Aug 22, 2013 at 11:26 AM, Arpit Agarwal <
>>> aagarwal@hortonworks.com> wrote:
>>>
>>>> I just noticed you are on Cygwin. IIRC Windows PIDs are not the same as
>>>> Cygwin PIDs so that may be causing the discrepancy. I don't know how well
>>>> Hadoop works in Cygwin as I have never tried it. Work is in progress for
>>>> native Windows support however there are no official releases with Windows
>>>> support yet. It may be easier to get familiar with a release<https://www.apache.org/dyn/closer.cgi/hadoop/common/>on
Linux if you are new to it.
>>>>
>>>>
>>>>
>>>> On Wed, Aug 21, 2013 at 10:05 PM, Irfan Sayed <irfu.sayed@gmail.com>wrote:
>>>>
>>>>> thanks
>>>>> here is what i did .
>>>>> i stopped all the namenodes and datanodes using ./stop-dfs.sh command
>>>>> then deleted all pid files for namenodes and datanodes
>>>>>
>>>>> started dfs again with command : "./start-dfs.sh"
>>>>>
>>>>> when i ran the "Jps" command . it shows
>>>>>
>>>>> Administrator@DFS-DC /cygdrive/c/Java/jdk1.7.0_25/bin
>>>>> $ ./jps.exe
>>>>> 4536 Jps
>>>>> 2076 NameNode
>>>>>
>>>>> however, when i open the pid file for namenode then it is not showing
>>>>> pid as : 4560. on the contrary, it shud show : 2076
>>>>>
>>>>> please suggest
>>>>>
>>>>> regards
>>>>>
>>>>>
>>>>>
>>>>> On Thu, Aug 22, 2013 at 9:59 AM, Arpit Agarwal <
>>>>> aagarwal@hortonworks.com> wrote:
>>>>>
>>>>>> Most likely there is a stale pid file. Something like
>>>>>> \tmp\hadoop-*datanode.pid. You could try deleting it and then restarting
>>>>>> the datanode.
>>>>>>
>>>>>> I haven't read the entire thread so you may have looked at this
>>>>>> already.
>>>>>>
>>>>>> -Arpit
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Wed, Aug 21, 2013 at 9:22 PM, Irfan Sayed <irfu.sayed@gmail.com>wrote:
>>>>>>
>>>>>>> datanode is trying to connect to namenode continuously but fails
>>>>>>>
>>>>>>> when i try to run "jps" command it says :
>>>>>>> $ ./jps.exe
>>>>>>> 4584 NameNode
>>>>>>> 4016 Jps
>>>>>>>
>>>>>>> and when i ran the "./start-dfs.sh" then it says :
>>>>>>>
>>>>>>> $ ./start-dfs.sh
>>>>>>> namenode running as process 3544. Stop it first.
>>>>>>> DFS-1: datanode running as process 4076. Stop it first.
>>>>>>> localhost: secondarynamenode running as process 4792. Stop it
first.
>>>>>>>
>>>>>>> both these logs are contradictory
>>>>>>> please find the attached logs
>>>>>>>
>>>>>>> should i attach the conf files as well ?
>>>>>>>
>>>>>>> regards
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Wed, Aug 21, 2013 at 5:28 PM, Mohammad Tariq <dontariq@gmail.com>wrote:
>>>>>>>
>>>>>>>> Your DN is still not running. Showing me the logs would be
helpful.
>>>>>>>>
>>>>>>>> Warm Regards,
>>>>>>>> Tariq
>>>>>>>> cloudfront.blogspot.com
>>>>>>>>
>>>>>>>>
>>>>>>>> On Wed, Aug 21, 2013 at 5:11 PM, Irfan Sayed <irfu.sayed@gmail.com>wrote:
>>>>>>>>
>>>>>>>>> i followed the url and did the steps mention in that.
i have
>>>>>>>>> deployed on the windows platform
>>>>>>>>>
>>>>>>>>> Now, i am able to browse url : http://localhost:50070
(name node )
>>>>>>>>> however, not able to browse url : http://localhost:50030
>>>>>>>>>
>>>>>>>>> please refer below
>>>>>>>>>
>>>>>>>>> [image: Inline image 1]
>>>>>>>>>
>>>>>>>>> i have modified all the config files as mentioned and
formatted
>>>>>>>>> the hdfs file system as well
>>>>>>>>> please suggest
>>>>>>>>>
>>>>>>>>> regards
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Tue, Aug 20, 2013 at 4:14 PM, Irfan Sayed <irfu.sayed@gmail.com
>>>>>>>>> > wrote:
>>>>>>>>>
>>>>>>>>>> thanks. i followed this url :
>>>>>>>>>> http://blog.sqltrainer.com/2012/01/installing-and-configuring-apache.html
>>>>>>>>>> let me follow the url which you gave for pseudo distributed
setup
>>>>>>>>>> and then will switch to distributed mode
>>>>>>>>>>
>>>>>>>>>> regards
>>>>>>>>>> irfan
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Tue, Aug 20, 2013 at 3:23 PM, Mohammad Tariq <
>>>>>>>>>> dontariq@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> You are welcome. Which link have you followed
for the
>>>>>>>>>>> configuration?Your *core-site.xml* is empty.
Remove the
>>>>>>>>>>> property *fs.default.name *from *hdfs-site.xml*
and add it to *
>>>>>>>>>>> core-site.xml*. Remove *mapred.job.tracker* as
well. It is
>>>>>>>>>>> required in *mapred-site.xml*.
>>>>>>>>>>>
>>>>>>>>>>> I would suggest you to do a pseudo distributed
setup first in
>>>>>>>>>>> order to get yourself familiar with the process
and then proceed to the
>>>>>>>>>>> distributed mode. You can visit this link<http://cloudfront.blogspot.in/2012/07/how-to-configure-hadoop.html#.UhM8d2T0-4I>if
you need some help. Let me know if you face any issue.
>>>>>>>>>>>
>>>>>>>>>>> HTH
>>>>>>>>>>>
>>>>>>>>>>> Warm Regards,
>>>>>>>>>>> Tariq
>>>>>>>>>>> cloudfront.blogspot.com
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Tue, Aug 20, 2013 at 2:56 PM, Irfan Sayed
<
>>>>>>>>>>> irfu.sayed@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> thanks tariq for response.
>>>>>>>>>>>> as discussed last time, i have sent you all
the config files in
>>>>>>>>>>>> my setup .
>>>>>>>>>>>> can you please go through that ?
>>>>>>>>>>>>
>>>>>>>>>>>> please let me know
>>>>>>>>>>>>
>>>>>>>>>>>> regards
>>>>>>>>>>>> irfan
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Tue, Aug 20, 2013 at 1:22 PM, Mohammad
Tariq <
>>>>>>>>>>>> dontariq@gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> I'm sorry for being unresponsive. Was
out of touch for
>>>>>>>>>>>>> sometime because of ramzan and eid. Resuming
work today.
>>>>>>>>>>>>>
>>>>>>>>>>>>> What's the current status?
>>>>>>>>>>>>>
>>>>>>>>>>>>> Warm Regards,
>>>>>>>>>>>>> Tariq
>>>>>>>>>>>>> cloudfront.blogspot.com
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Mon, Aug 19, 2013 at 7:18 PM, manish
dunani <
>>>>>>>>>>>>> manishd207@gmail.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> First of all read the concepts ..I
hope you will like it..
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> https://www.frcrc.org/sites/default/files/HadoopTutorialPart1.pdf
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Mon, Aug 19, 2013 at 9:45 AM,
Irfan Sayed <
>>>>>>>>>>>>>> irfu.sayed@gmail.com> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> please suggest
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> regards
>>>>>>>>>>>>>>> irfan
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Tue, Aug 13, 2013 at 12:56
PM, Irfan Sayed <
>>>>>>>>>>>>>>> irfu.sayed@gmail.com> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> hey Tariq,
>>>>>>>>>>>>>>>> i am still stuck ..
>>>>>>>>>>>>>>>> can you please suggest
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> regards
>>>>>>>>>>>>>>>> irfan
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Thu, Aug 8, 2013 at 5:56
AM, Irfan Sayed <
>>>>>>>>>>>>>>>> irfu.sayed@gmail.com>
wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> please suggest
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> regards
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Wed, Aug 7, 2013 at
9:49 AM, Irfan Sayed <
>>>>>>>>>>>>>>>>> irfu.sayed@gmail.com>
wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> attachment got quarantined
>>>>>>>>>>>>>>>>>> resending in txt
format. please rename it to conf.rar
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> regards
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On Wed, Aug 7, 2013
at 9:41 AM, Irfan Sayed <
>>>>>>>>>>>>>>>>>> irfu.sayed@gmail.com>
wrote:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> thanks.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> if i run the
jps command on namenode :
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Administrator@DFS-DC
/cygdrive/c/Java/jdk1.7.0_25/bin
>>>>>>>>>>>>>>>>>>> $ ./jps.exe
>>>>>>>>>>>>>>>>>>> 3164 NameNode
>>>>>>>>>>>>>>>>>>> 1892 Jps
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> same command
on datanode :
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Administrator@DFS-1
/cygdrive/c/Java/jdk1.7.0_25/bin
>>>>>>>>>>>>>>>>>>> $ ./jps.exe
>>>>>>>>>>>>>>>>>>> 3848 Jps
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> jps does not
list any process for datanode. however, on
>>>>>>>>>>>>>>>>>>> web browser i
can see one live data node
>>>>>>>>>>>>>>>>>>> please find the
attached conf rar file of namenode
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> regards
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> On Wed, Aug 7,
2013 at 1:52 AM, Mohammad Tariq <
>>>>>>>>>>>>>>>>>>> dontariq@gmail.com>
wrote:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> OK. we'll
start fresh. Could you plz show me your
>>>>>>>>>>>>>>>>>>>> latest config
files?
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> BTW, are
your daemons running fine?Use JPS to verify
>>>>>>>>>>>>>>>>>>>> that.
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Warm Regards,
>>>>>>>>>>>>>>>>>>>> Tariq
>>>>>>>>>>>>>>>>>>>> cloudfront.blogspot.com
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> On Tue, Aug
6, 2013 at 10:59 PM, Irfan Sayed <
>>>>>>>>>>>>>>>>>>>> irfu.sayed@gmail.com>
wrote:
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> i have
created these dir "wksp_data" and "wksp_name"
>>>>>>>>>>>>>>>>>>>>> on both
datanode and namenode
>>>>>>>>>>>>>>>>>>>>> made
the respective changes in "hdfs-site.xml" file
>>>>>>>>>>>>>>>>>>>>> formatted
the namenode
>>>>>>>>>>>>>>>>>>>>> started
the dfs
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> but still,
not able to browse the file system through
>>>>>>>>>>>>>>>>>>>>> web browser
>>>>>>>>>>>>>>>>>>>>> please
refer below
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> anything
still missing ?
>>>>>>>>>>>>>>>>>>>>> please
suggest
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> [image:
Inline image 1]
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> On Tue,
Aug 6, 2013 at 10:35 PM, Irfan Sayed <
>>>>>>>>>>>>>>>>>>>>> irfu.sayed@gmail.com>
wrote:
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> these
dir needs to be created on all datanodes and
>>>>>>>>>>>>>>>>>>>>>> namenodes
?
>>>>>>>>>>>>>>>>>>>>>> further,
 hdfs-site.xml needs to be updated on both
>>>>>>>>>>>>>>>>>>>>>> datanodes
and namenodes for these new dir?
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> regards
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> On
Tue, Aug 6, 2013 at 5:30 PM, Mohammad Tariq <
>>>>>>>>>>>>>>>>>>>>>> dontariq@gmail.com>
wrote:
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
Create 2 directories manually corresponding to the
>>>>>>>>>>>>>>>>>>>>>>>
values of dfs.name.dir and dfs.data.dir properties and change the
>>>>>>>>>>>>>>>>>>>>>>>
permissions of these directories to 755. When you start pushing data into
>>>>>>>>>>>>>>>>>>>>>>>
your HDFS, data will start going inside the directory specified by
>>>>>>>>>>>>>>>>>>>>>>>
dfs.data.dir and the associated metadata will go inside dfs.name.dir.
>>>>>>>>>>>>>>>>>>>>>>>
Remember, you store data in HDFS, but it eventually gets stored in your
>>>>>>>>>>>>>>>>>>>>>>>
local/native FS. But you cannot see this data directly on your local/native
>>>>>>>>>>>>>>>>>>>>>>>
FS.
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
Warm Regards,
>>>>>>>>>>>>>>>>>>>>>>>
Tariq
>>>>>>>>>>>>>>>>>>>>>>>
cloudfront.blogspot.com
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
On Tue, Aug 6, 2013 at 5:26 PM, Irfan Sayed <
>>>>>>>>>>>>>>>>>>>>>>>
irfu.sayed@gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
thanks.
>>>>>>>>>>>>>>>>>>>>>>>>
however, i need this to be working on windows
>>>>>>>>>>>>>>>>>>>>>>>>
environment as project requirement.
>>>>>>>>>>>>>>>>>>>>>>>>
i will add/work on Linux later
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
so, now , at this stage , c:\\wksp is the HDFS file
>>>>>>>>>>>>>>>>>>>>>>>>
system OR do i need to create it from command line ?
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
please suggest
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
regards,
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
On Tue, Aug 6, 2013 at 5:19 PM, Mohammad Tariq <
>>>>>>>>>>>>>>>>>>>>>>>>
dontariq@gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
Hello Irfan,
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
Sorry for being unresponsive. Got stuck with some
>>>>>>>>>>>>>>>>>>>>>>>>>
imp work.
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
HDFS webUI doesn't provide us the ability to
>>>>>>>>>>>>>>>>>>>>>>>>>
create file or directory. You can browse HDFS, view files, download files
>>>>>>>>>>>>>>>>>>>>>>>>>
etc. But operation like create, move, copy etc are not supported.
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
These values look fine to me.
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
One suggestion though. Try getting a Linux
>>>>>>>>>>>>>>>>>>>>>>>>>
machine(if possible). Or at least use a VM. I personally feel that using
>>>>>>>>>>>>>>>>>>>>>>>>>
Hadoop on windows is always messy.
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
Warm Regards,
>>>>>>>>>>>>>>>>>>>>>>>>>
Tariq
>>>>>>>>>>>>>>>>>>>>>>>>>
cloudfront.blogspot.com
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
On Tue, Aug 6, 2013 at 5:09 PM, Irfan Sayed <
>>>>>>>>>>>>>>>>>>>>>>>>>
irfu.sayed@gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
thanks.
>>>>>>>>>>>>>>>>>>>>>>>>>>
when i browse the file system , i am getting
>>>>>>>>>>>>>>>>>>>>>>>>>>
following :
>>>>>>>>>>>>>>>>>>>>>>>>>>
i haven't seen any make directory option there
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
i need to create it from command line ?
>>>>>>>>>>>>>>>>>>>>>>>>>>
further, in the hdfs-site.xml file , i have given
>>>>>>>>>>>>>>>>>>>>>>>>>>
following entries. are they correct ?
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
<property>
>>>>>>>>>>>>>>>>>>>>>>>>>>
  <name>dfs.data.dir</name>
>>>>>>>>>>>>>>>>>>>>>>>>>>
  <value>c:\\wksp</value>
>>>>>>>>>>>>>>>>>>>>>>>>>>
  </property>
>>>>>>>>>>>>>>>>>>>>>>>>>>
<property>
>>>>>>>>>>>>>>>>>>>>>>>>>>
  <name>dfs.name.dir</name>
>>>>>>>>>>>>>>>>>>>>>>>>>>
  <value>c:\\wksp</value>
>>>>>>>>>>>>>>>>>>>>>>>>>>
  </property>
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
please suggest
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
[image: Inline image 1]
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
On Tue, Aug 6, 2013 at 12:40 PM, manish dunani <
>>>>>>>>>>>>>>>>>>>>>>>>>>
manishd207@gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
*You are wrong at this:*
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
Administrator@DFS-DC/cygdrive/c/hadoop-1.1.2/hadoop-1.1.2/bin
>>>>>>>>>>>>>>>>>>>>>>>>>>>
$ ./hadoop dfs -copyFromLocal
>>>>>>>>>>>>>>>>>>>>>>>>>>>
/cygdrive/c/Users/Administrator/Desktop/hadoop-1.1.2.tar /wksp
>>>>>>>>>>>>>>>>>>>>>>>>>>>
copyFromLocal: File
>>>>>>>>>>>>>>>>>>>>>>>>>>>
/cygdrive/c/Users/Administrator/Desktop/hadoop-1.1.2.tar does not exist.
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
Administrator@DFS-DC/cygdrive/c/hadoop-1.1.2/hadoop-1.1.2/bin
>>>>>>>>>>>>>>>>>>>>>>>>>>>
$ ./hadoop dfs -copyFromLocal
>>>>>>>>>>>>>>>>>>>>>>>>>>>
/cygdrive/c/Users/Administrator/Desktop/hadoop-1.1.2.tar.gz /wksp
>>>>>>>>>>>>>>>>>>>>>>>>>>>
copyFromLocal: File
>>>>>>>>>>>>>>>>>>>>>>>>>>>
/cygdrive/c/Users/Administrator/Desktop/hadoop-1.1.2.tar.gz does not exist.
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
Because,You had wrote both the paths local and
>>>>>>>>>>>>>>>>>>>>>>>>>>>
You need not to copy hadoop into hdfs...Hadoop is already working..
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
Just check out in browser by after starting ur
>>>>>>>>>>>>>>>>>>>>>>>>>>>
single node cluster :
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
localhost:50070
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
then go for browse the filesystem link in it..
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
If there is no directory then make directory
>>>>>>>>>>>>>>>>>>>>>>>>>>>
there.
>>>>>>>>>>>>>>>>>>>>>>>>>>>
That is your hdfs directory.
>>>>>>>>>>>>>>>>>>>>>>>>>>>
Then copy any text file there(no need to copy
>>>>>>>>>>>>>>>>>>>>>>>>>>>
hadoop there).beacause u are going to do processing on that data in text
>>>>>>>>>>>>>>>>>>>>>>>>>>>
file.That's why hadoop is used for ,first u need to make it clear in ur
>>>>>>>>>>>>>>>>>>>>>>>>>>>
mind.Then and then u will do it...otherwise not possible..
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
*Try this: *
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
Administrator@DFS-DC/cygdrive/c/hadoop-1.1.2/hadoop-1.1.2
>>>>>>>>>>>>>>>>>>>>>>>>>>>
$ .bin/hadoop dfs -copyFromLocal
>>>>>>>>>>>>>>>>>>>>>>>>>>>
/full/local/path/to/ur/file /hdfs/directory/path
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
On Tue, Aug 6, 2013 at 11:49 AM, Irfan Sayed <
>>>>>>>>>>>>>>>>>>>>>>>>>>>
irfu.sayed@gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
thanks. yes , i am newbie.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
however, i need windows setup.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
let me surely refer the doc and link which u
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
sent but i need this to be working ...
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
can you please help
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
regards
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
--
>>>>>>>>>>>>>>>>>>>>>>>>>>>
MANISH DUNANI
>>>>>>>>>>>>>>>>>>>>>>>>>>>
-THANX
>>>>>>>>>>>>>>>>>>>>>>>>>>>
+91 9426881954,+91 8460656443
>>>>>>>>>>>>>>>>>>>>>>>>>>>
manishd207@gmail.com
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> --
>>>>>>>>>>>>>> Regards
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> *Manish Dunani*
>>>>>>>>>>>>>> *Contact No* : +91 9408329137
>>>>>>>>>>>>>> *skype id* : manish.dunani*
>>>>>>>>>>>>>> *
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>> CONFIDENTIALITY NOTICE
>>>>>> NOTICE: This message is intended for the use of the individual or
>>>>>> entity to which it is addressed and may contain information that
is
>>>>>> confidential, privileged and exempt from disclosure under applicable
law.
>>>>>> If the reader of this message is not the intended recipient, you
are hereby
>>>>>> notified that any printing, copying, dissemination, distribution,
>>>>>> disclosure or forwarding of this communication is strictly prohibited.
If
>>>>>> you have received this communication in error, please contact the
sender
>>>>>> immediately and delete it from your system. Thank You.
>>>>>
>>>>>
>>>>>
>>>>
>>>> CONFIDENTIALITY NOTICE
>>>> NOTICE: This message is intended for the use of the individual or
>>>> entity to which it is addressed and may contain information that is
>>>> confidential, privileged and exempt from disclosure under applicable law.
>>>> If the reader of this message is not the intended recipient, you are hereby
>>>> notified that any printing, copying, dissemination, distribution,
>>>> disclosure or forwarding of this communication is strictly prohibited. If
>>>> you have received this communication in error, please contact the sender
>>>> immediately and delete it from your system. Thank You.
>>>>
>>>
>>>
>>
>

Mime
View raw message