hadoop-mapreduce-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Mohammad Alkahtani <m.alkaht...@gmail.com>
Subject Re: Hadoop Debian Package
Date Sun, 17 Mar 2013 18:46:39 GMT
I tried echo $HADOOP_HOME

and I got blank

Mohammad Alkahtani
P.O.Box 102275
Riyadh 11675
Saudi Arabia
mobile: 00966 555 33 1717


On Sun, Mar 17, 2013 at 9:37 PM, Mohammad Alkahtani
<m.alkahtani@gmail.com>wrote:

> Thank you Mohammad
> I still get the same error with this msg
>
> localhost: Warning: $HADOOP_HOME is deprecated.
> I searched ~/.bashrc but only what I wrote is there.
>
>
> Mohammad Alkahtani
> P.O.Box 102275
> Riyadh 11675
> Saudi Arabia
> mobile: 00966 555 33 1717
>
>
> On Sun, Mar 17, 2013 at 9:21 PM, Mohammad Tariq <dontariq@gmail.com>wrote:
>
>> you can do that using these command :
>>
>> sudo gedit ~/.bashrc
>>
>> then go to the end of the file and add this line :
>> export HADOOP_HOME=/YOUR_FULL_HADOOP_PATH
>>
>> after that use it to freeze the changes :
>> source ~/.bashrc
>>
>> to check it :
>> echo $HADOOP_HOME
>>
>> This will permanently set your HADOOP_HOME.
>>
>> HTH
>>
>>
>> Warm Regards,
>> Tariq
>> https://mtariq.jux.com/
>> cloudfront.blogspot.com
>>
>>
>> On Sun, Mar 17, 2013 at 11:46 PM, Mohammad Alkahtani <
>> m.alkahtani@gmail.com> wrote:
>>
>>> Hi Tariq, Could you please tell me how to set HADOOP_HOME because I
>>> don't find it in the hadoop-env.sh
>>>
>>> Thank you Shashwat
>>> this is the output and it is already configured but hadoop don't read
>>> the configuration from here.
>>>
>>> /usr/share/maven-repo/org/apache
>>> /commons/commons-parent/22/commons-parent-22-site.xml
>>> /usr/share/maven-repo/org/apache/commons/commons-parent/debian
>>> /commons-parent-debian-site.xml
>>> /usr/share/maven-repo/org/apache/apache/10/apache-10-site.xml
>>> /usr/share/maven-repo/org/apache/apache/debian/apache-debian-site.xml
>>> /usr/share/compiz/composite.xml
>>> /usr/share/hadoop/templates/conf/mapred-site.xml
>>> /usr/share/hadoop/templates/conf/core-site.xml
>>> /usr/share/hadoop/templates/conf/hdfs-site.xml
>>>
>>> Mohammad Alkahtani
>>> P.O.Box 102275
>>> Riyadh 11675
>>> Saudi Arabia
>>> mobile: 00966 555 33 1717
>>>
>>>
>>> On Sun, Mar 17, 2013 at 9:07 PM, shashwat shriparv <
>>> dwivedishashwat@gmail.com> wrote:
>>>
>>>> try
>>>> find / -type f -iname "*site.xml"
>>>> it will show you where ever those files are..
>>>>
>>>>
>>>>
>>>> ∞
>>>> Shashwat Shriparv
>>>>
>>>>
>>>>
>>>> On Sun, Mar 17, 2013 at 11:34 PM, Mohammad Alkahtani <
>>>> m.alkahtani@gmail.com> wrote:
>>>>
>>>>> The problem is I tried I read the configuration file by changing
>>>>> export HADOOP_CONF_DIR=${HADOOP_CONF_
>>>>> DIR:-"/usr/shar/hadoop/templates/conf"}
>>>>> but I think Hadoop dosen't get the configration from this dir, I trid
>>>>> and searched the system for conf dir the only dir is this one which I
>>>>> changed.
>>>>>
>>>>> Mohammad Alkahtani
>>>>> P.O.Box 102275
>>>>> Riyadh 11675
>>>>> Saudi Arabia
>>>>> mobile: 00966 555 33 1717
>>>>>
>>>>>
>>>>> On Sun, Mar 17, 2013 at 8:57 PM, shashwat shriparv <
>>>>> dwivedishashwat@gmail.com> wrote:
>>>>>
>>>>>> Ye its is asking for file:/// instead of hdfs:// just check if it
is
>>>>>> taking setting configuration from other location...
>>>>>>
>>>>>>
>>>>>>
>>>>>> ∞
>>>>>> Shashwat Shriparv
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Sun, Mar 17, 2013 at 11:07 PM, Luangsay Sourygna <
>>>>>> luangsay@gmail.com> wrote:
>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> What is the version of Hadoop you use?
>>>>>>>
>>>>>>> Try using fs.defaultFS instead of fs.default.name (see the list
of
>>>>>>> all
>>>>>>> the deprecated properties here:
>>>>>>>
>>>>>>> http://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/DeprecatedProperties.html
>>>>>>> ).
>>>>>>> I remember I once had a similar error message and it was due
to the
>>>>>>> change in properties names.
>>>>>>>
>>>>>>> Regards,
>>>>>>>
>>>>>>> Sourygna
>>>>>>>
>>>>>>> On Sun, Mar 17, 2013 at 2:32 PM, Mohammad Alkahtani
>>>>>>> <m.alkahtani@gmail.com> wrote:
>>>>>>> > Hi to all users of Hadoop,
>>>>>>> >
>>>>>>> > I installed Hadoop the .deb file on Ubuntu 12.04 but I might
could
>>>>>>> not
>>>>>>> > configure it right. The conf dir is under templates in
>>>>>>> /usr/shar/hadoop. I
>>>>>>> > edit the core-site.xml, mapred-site.xml files to give
>>>>>>> > <property>
>>>>>>> > <name>fs.default.name</name>
>>>>>>> > <value>hdfs://localhost:9000</value>
>>>>>>> > </property>
>>>>>>> > and for mapred
>>>>>>> > <property>
>>>>>>> > <name>mapred.job.tracker</name>
>>>>>>> > <value>localhost:9001</value>
>>>>>>> > </property>
>>>>>>> >
>>>>>>> > but i get these errors, I assume that there is problem,
Hadoop
>>>>>>> cannot read
>>>>>>> > the configuration file.
>>>>>>> > I chaned the hadoop-env.sh to
>>>>>>> > export
>>>>>>> HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/usr/shar/hadoop/templates/conf"}
>>>>>>> > but dosen't solve the problem.
>>>>>>> >
>>>>>>> > ERROR org.apache.hadoop.hdfs.server.datanode.DataNode:
>>>>>>> > java.lang.IllegalArgumentException: Does not contain a valid
>>>>>>> host:port
>>>>>>> > authority: file:/// at
>>>>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164)
>>>>>>> at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201)
>>>>>>> > at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231)
>>>>>>> > at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:225)
>>>>>>> > at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:347)
>>>>>>> > at
>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:309)
at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1651)
>>>>>>> > at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1590)
>>>>>>> > at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1608)
>>>>>>> > at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1734)
>>>>>>> > at
>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1751)
>>>>>>> >
>>>>>>> > ________________________________
>>>>>>> >
>>>>>>> > FATAL org.apache.hadoop.mapred.JobTracker:
>>>>>>> > java.lang.IllegalArgumentException: Does not contain a valid
>>>>>>> host:port
>>>>>>> > authority: local at
>>>>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164)
>>>>>>> at
>>>>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130)
>>>>>>> at
>>>>>>> >
>>>>>>> org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2312)
at
>>>>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:2070)
at
>>>>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1889)
at
>>>>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1883)
at
>>>>>>> >
>>>>>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:312)
at
>>>>>>> >
>>>>>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:303)
at
>>>>>>> >
>>>>>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:298)
>>>>>>> > at org.apache.hadoop.mapred.JobTracker.main(JobTracker.java:4791)
>>>>>>> >
>>>>>>> > ________________________________
>>>>>>> >
>>>>>>> > ERROR org.apache.hadoop.hdfs.server.namenode.NameNode:
>>>>>>> > java.lang.IllegalArgumentException: Does not contain a valid
>>>>>>> host:port
>>>>>>> > authority: file:/// at
>>>>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164)
>>>>>>> at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201)
>>>>>>> > at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231)
>>>>>>> > at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:265)
>>>>>>> > at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:536)
at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1410)
>>>>>>> > at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1419)
>>>>>>> >
>>>>>>> > ________________________________
>>>>>>> >
>>>>>>> > Exception in thread "main" java.lang.IllegalArgumentException:
>>>>>>> Does not
>>>>>>> > contain a valid host:port authority: file:/// at
>>>>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164)
>>>>>>> at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201)
>>>>>>> > at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231)
>>>>>>> > at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:225)
>>>>>>> > at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.initialize(SecondaryNameNode.java:167)
>>>>>>> > at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.(SecondaryNameNode.java:135)
>>>>>>> > at
>>>>>>> >
>>>>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:650)
>>>>>>> >
>>>>>>> > ________________________________
>>>>>>> >
>>>>>>> > ERROR org.apache.hadoop.mapred.TaskTracker: Can not start
task
>>>>>>> tracker
>>>>>>> > because java.lang.IllegalArgumentException: Does not contain
a
>>>>>>> valid
>>>>>>> > host:port authority: local at
>>>>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164)
>>>>>>> at
>>>>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130)
>>>>>>> at
>>>>>>> >
>>>>>>> org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2312)
at
>>>>>>> > org.apache.hadoop.mapred.TaskTracker.(TaskTracker.java:1532)
at
>>>>>>> > org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:3906)
>>>>>>> >
>>>>>>> >
>>>>>>> > Regards,
>>>>>>> > Mohammad Alkahtani
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Mime
View raw message