hadoop-hdfs-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Mohammad Alkahtani <m.alkaht...@gmail.com>
Subject Re: Hadoop Debian Package
Date Sun, 17 Mar 2013 18:16:12 GMT
Hi Tariq, Could you please tell me how to set HADOOP_HOME because I don't
find it in the hadoop-env.sh

Thank you Shashwat
this is the output and it is already configured but hadoop don't read the
configuration from here.

/usr/share/maven-repo/org/apache
/commons/commons-parent/22/commons-parent-22-site.xml
/usr/share/maven-repo/org/apache/commons/commons-parent/debian
/commons-parent-debian-site.xml
/usr/share/maven-repo/org/apache/apache/10/apache-10-site.xml
/usr/share/maven-repo/org/apache/apache/debian/apache-debian-site.xml
/usr/share/compiz/composite.xml
/usr/share/hadoop/templates/conf/mapred-site.xml
/usr/share/hadoop/templates/conf/core-site.xml
/usr/share/hadoop/templates/conf/hdfs-site.xml

Mohammad Alkahtani
P.O.Box 102275
Riyadh 11675
Saudi Arabia
mobile: 00966 555 33 1717


On Sun, Mar 17, 2013 at 9:07 PM, shashwat shriparv <
dwivedishashwat@gmail.com> wrote:

> try
> find / -type f -iname "*site.xml"
> it will show you where ever those files are..
>
>
>
> ∞
> Shashwat Shriparv
>
>
>
> On Sun, Mar 17, 2013 at 11:34 PM, Mohammad Alkahtani <
> m.alkahtani@gmail.com> wrote:
>
>> The problem is I tried I read the configuration file by changing
>> export HADOOP_CONF_DIR=${HADOOP_CONF_
>> DIR:-"/usr/shar/hadoop/templates/conf"}
>> but I think Hadoop dosen't get the configration from this dir, I trid and
>> searched the system for conf dir the only dir is this one which I changed.
>>
>> Mohammad Alkahtani
>> P.O.Box 102275
>> Riyadh 11675
>> Saudi Arabia
>> mobile: 00966 555 33 1717
>>
>>
>> On Sun, Mar 17, 2013 at 8:57 PM, shashwat shriparv <
>> dwivedishashwat@gmail.com> wrote:
>>
>>> Ye its is asking for file:/// instead of hdfs:// just check if it is
>>> taking setting configuration from other location...
>>>
>>>
>>>
>>> ∞
>>> Shashwat Shriparv
>>>
>>>
>>>
>>> On Sun, Mar 17, 2013 at 11:07 PM, Luangsay Sourygna <luangsay@gmail.com>wrote:
>>>
>>>> Hi,
>>>>
>>>> What is the version of Hadoop you use?
>>>>
>>>> Try using fs.defaultFS instead of fs.default.name (see the list of all
>>>> the deprecated properties here:
>>>>
>>>> http://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/DeprecatedProperties.html
>>>> ).
>>>> I remember I once had a similar error message and it was due to the
>>>> change in properties names.
>>>>
>>>> Regards,
>>>>
>>>> Sourygna
>>>>
>>>> On Sun, Mar 17, 2013 at 2:32 PM, Mohammad Alkahtani
>>>> <m.alkahtani@gmail.com> wrote:
>>>> > Hi to all users of Hadoop,
>>>> >
>>>> > I installed Hadoop the .deb file on Ubuntu 12.04 but I might could not
>>>> > configure it right. The conf dir is under templates in
>>>> /usr/shar/hadoop. I
>>>> > edit the core-site.xml, mapred-site.xml files to give
>>>> > <property>
>>>> > <name>fs.default.name</name>
>>>> > <value>hdfs://localhost:9000</value>
>>>> > </property>
>>>> > and for mapred
>>>> > <property>
>>>> > <name>mapred.job.tracker</name>
>>>> > <value>localhost:9001</value>
>>>> > </property>
>>>> >
>>>> > but i get these errors, I assume that there is problem, Hadoop cannot
>>>> read
>>>> > the configuration file.
>>>> > I chaned the hadoop-env.sh to
>>>> > export
>>>> HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/usr/shar/hadoop/templates/conf"}
>>>> > but dosen't solve the problem.
>>>> >
>>>> > ERROR org.apache.hadoop.hdfs.server.datanode.DataNode:
>>>> > java.lang.IllegalArgumentException: Does not contain a valid host:port
>>>> > authority: file:/// at
>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201)
>>>> > at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231)
>>>> > at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:225)
>>>> > at
>>>> >
>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:347)
>>>> > at
>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:309) at
>>>> >
>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1651)
>>>> > at
>>>> >
>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1590)
>>>> > at
>>>> >
>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1608)
>>>> > at
>>>> >
>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1734)
>>>> > at
>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1751)
>>>> >
>>>> > ________________________________
>>>> >
>>>> > FATAL org.apache.hadoop.mapred.JobTracker:
>>>> > java.lang.IllegalArgumentException: Does not contain a valid host:port
>>>> > authority: local at
>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at
>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130) at
>>>> > org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2312)
>>>> at
>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:2070) at
>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1889) at
>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1883) at
>>>> > org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:312)
>>>> at
>>>> > org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:303)
>>>> at
>>>> > org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:298)
>>>> > at org.apache.hadoop.mapred.JobTracker.main(JobTracker.java:4791)
>>>> >
>>>> > ________________________________
>>>> >
>>>> > ERROR org.apache.hadoop.hdfs.server.namenode.NameNode:
>>>> > java.lang.IllegalArgumentException: Does not contain a valid host:port
>>>> > authority: file:/// at
>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201)
>>>> > at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231)
>>>> > at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:265)
>>>> > at
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:536) at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1410)
>>>> > at
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1419)
>>>> >
>>>> > ________________________________
>>>> >
>>>> > Exception in thread "main" java.lang.IllegalArgumentException: Does
>>>> not
>>>> > contain a valid host:port authority: file:/// at
>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201)
>>>> > at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231)
>>>> > at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:225)
>>>> > at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.initialize(SecondaryNameNode.java:167)
>>>> > at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.(SecondaryNameNode.java:135)
>>>> > at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:650)
>>>> >
>>>> > ________________________________
>>>> >
>>>> > ERROR org.apache.hadoop.mapred.TaskTracker: Can not start task tracker
>>>> > because java.lang.IllegalArgumentException: Does not contain a valid
>>>> > host:port authority: local at
>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at
>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130) at
>>>> > org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2312)
>>>> at
>>>> > org.apache.hadoop.mapred.TaskTracker.(TaskTracker.java:1532) at
>>>> > org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:3906)
>>>> >
>>>> >
>>>> > Regards,
>>>> > Mohammad Alkahtani
>>>>
>>>
>>>
>>
>

Mime
View raw message