hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From siddharth raghuvanshi <track009.siddha...@gmail.com>
Subject Re: Unknown Host Exception
Date Sun, 10 Oct 2010 14:21:00 GMT
Hi Shi,

I am a beginner in Hadoop. I have given the following value in core-site.xml
<name>hadoop.tmp.dir</name>
 <value>/users/user/hadoop-datastore/hadoop</value>


<name>fs.default.name</name>
 <value>hdfs://localhost:54310</value>
How will we check whether the host machine is reachable or not?

Also, in mapred-site.xml, I have given
<name>mapred.job.tracker</name>
 <value>localhost:54311</value>


Please check whether these values are correct or not, if not correct what
should I do?

Waiting for your reply
Regards
Siddharth



On Sat, Oct 9, 2010 at 11:47 PM, Shi Yu <shiyu@uchicago.edu> wrote:

> I suggest you change the hadoop.tmp.dir  value in hadoop-site.xml (0.19.x)
> and reformat, restart it. Also double check the host machine in
> fs.default.name and mapred.job.tracker is reachable or not.
>
> Shi
>
>
> On 2010-10-9 12:57, siddharth raghuvanshi wrote:
>
>> Hi,
>> I am also getting the following error. Please tell me whether this error
>> is
>> related to the previous error which I asked an hour before or this is a
>> separate error...
>>
>> [user@cs-sy-249 hadoop]$ bin/hadoop dfs -copyFromLocal
>> /users/user/Desktop/test_data/
>> gutenberg
>>
>> 10/10/09 23:22:15 WARN hdfs.DFSClient: DataStreamer Exception:
>> org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
>> /user/user/gutenberg/pg4300.txt could only be replicated to 0 nodes,
>> instead
>> of 1
>>         at
>>
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1310)
>>
>>         at
>>
>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:469)
>>
>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>> Method)
>>         at
>>
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>
>>         at
>>
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>
>>         at
>> java.lang.reflect.Method.invoke(Method.java:597)
>>
>>         at
>> org.apache.hadoop.ipc.RPC$Server.call(RPC.java:512)
>>
>>         at
>> org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:968)
>>
>>         at
>> org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:964)
>>
>>         at java.security.AccessController.doPrivileged(Native
>> Method)
>>         at
>> javax.security.auth.Subject.doAs(Subject.java:396)
>>
>>         at
>> org.apache.hadoop.ipc.Server$Handler.run(Server.java:962)
>>
>>
>>         at org.apache.hadoop.ipc.Client.call(Client.java:817)
>>         at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:221)
>>         at $Proxy0.addBlock(Unknown Source)
>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>         at
>>
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>         at
>>
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>
>>         at
>> java.lang.reflect.Method.invoke(Method.java:597)
>>
>>         at
>>
>> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
>>
>>         at
>>
>> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
>>
>>         at $Proxy0.addBlock(Unknown
>> Source)
>>         at
>>
>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:3000)
>>
>>         at
>>
>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2881)
>>
>>         at
>>
>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$1900(DFSClient.java:2139)
>>         at
>>
>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2329)
>>
>>
>> 10/10/09 23:22:15 WARN hdfs.DFSClient: Error Recovery for block null bad
>> datanode[0] nodes ==
>> null
>>
>> 10/10/09 23:22:15 WARN hdfs.DFSClient: Could not get block locations.
>> Source
>> file "/user/user/gutenberg/pg4300.txt" -
>> Aborting...
>> copyFromLocal: java.io.IOException: File /user/user/gutenberg/pg4300.txt
>> could only be replicated to 0 nodes, instead of
>> 1
>> 10/10/09 23:22:15 ERROR hdfs.DFSClient: Exception closing file
>> /user/user/gutenberg/pg4300.txt : org.apache.hadoop.ipc.RemoteException:
>> java.io.IOException: File /user/user/gutenberg/pg4300.txt could only be
>> replicated to 0 nodes, instead of 1
>>         at
>>
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1310)
>>
>>         at
>>
>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:469)
>>
>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>> Method)
>>         at
>>
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>
>>         at
>>
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>
>>         at
>> java.lang.reflect.Method.invoke(Method.java:597)
>>
>>         at
>> org.apache.hadoop.ipc.RPC$Server.call(RPC.java:512)
>>
>>         at
>> org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:968)
>>
>>         at
>> org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:964)
>>
>>         at java.security.AccessController.doPrivileged(Native
>> Method)
>>         at
>> javax.security.auth.Subject.doAs(Subject.java:396)
>>
>>         at
>> org.apache.hadoop.ipc.Server$Handler.run(Server.java:962)
>>
>>
>> org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
>> /user/user/gutenberg/pg4300.txt could only be replicated to 0 nodes,
>> instead
>> of 1
>>         at
>>
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1310)
>>
>>         at
>>
>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:469)
>>
>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>> Method)
>>         at
>>
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>
>>         at
>>
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>         at java.lang.reflect.Method.invoke(Method.java:597)
>>         at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:512)
>>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:968)
>>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:964)
>>         at java.security.AccessController.doPrivileged(Native Method)
>>         at javax.security.auth.Subject.doAs(Subject.java:396)
>>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:962)
>>
>>         at org.apache.hadoop.ipc.Client.call(Client.java:817)
>>         at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:221)
>>         at $Proxy0.addBlock(Unknown Source)
>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>         at
>>
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>         at
>>
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>         at java.lang.reflect.Method.invoke(Method.java:597)
>>         at
>>
>> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
>>         at
>>
>> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
>>         at $Proxy0.addBlock(Unknown Source)
>>         at
>>
>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:3000)
>>         at
>>
>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2881)
>>         at
>>
>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$1900(DFSClient.java:2139)
>>         at
>>
>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2329)
>> [user@cs-sy-249 hadoop]$
>>
>>
>> Regards
>> Siddharth
>>
>>
>>
>>
>> On Sat, Oct 9, 2010 at 10:39 PM, siddharth raghuvanshi<
>> track009.siddharth@gmail.com>  wrote:
>>
>>
>>
>>> Hi,
>>>
>>> When I am running the following command in Mandriva Linux
>>>  hadoop namenode -format
>>>
>>> I am getting the following error:
>>>
>>> 10/10/09 22:32:07 INFO namenode.NameNode: STARTUP_MSG:
>>> /************************************************************
>>> STARTUP_MSG: Starting NameNode
>>> STARTUP_MSG:   host = java.net.UnknownHostException:
>>> cs-sy-249.cse.iitkgp.ernet.in: cs-sy-249.cse.iitkgp.ernet.in
>>> STARTUP_MSG:   args = [-format]
>>> STARTUP_MSG:   version = 0.20.2+320
>>> STARTUP_MSG:   build =  -r 9b72d268a0b590b4fd7d13aca17c1c453f8bc957;
>>> compiled by 'root' on Mon Jun 28 19:13:09 EDT 2010
>>> ************************************************************/
>>> Re-format filesystem in /users/user/hadoop-datastore/hadoop-user/dfs/name
>>> ?
>>> (Y or N) Y
>>> 10/10/09 22:32:11 INFO namenode.FSNamesystem: fsOwner=user,user
>>> 10/10/09 22:32:11 INFO namenode.FSNamesystem: supergroup=supergroup
>>> 10/10/09 22:32:11 INFO namenode.FSNamesystem: isPermissionEnabled=true
>>> 10/10/09 22:32:12 INFO metrics.MetricsUtil: Unable to obtain hostName
>>> java.net.UnknownHostException: cs-sy-249.cse.iitkgp.ernet.in:
>>> cs-sy-249.cse.iitkgp.ernet.in
>>>         at java.net.InetAddress.getLocalHost(InetAddress.java:1353)
>>>         at
>>> org.apache.hadoop.metrics.MetricsUtil.getHostName(MetricsUtil.java:91)
>>>         at
>>> org.apache.hadoop.metrics.MetricsUtil.createRecord(MetricsUtil.java:80)
>>>         at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.initialize(FSDirectory.java:78)
>>>         at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.<init>(FSDirectory.java:73)
>>>         at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:383)
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:904)
>>>         at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:998)
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1015)
>>> 10/10/09 22:32:12 INFO common.Storage: Image file of size 94 saved in 0
>>> seconds.
>>> 10/10/09 22:32:12 INFO common.Storage: Storage directory
>>> /users/user/hadoop-datastore/hadoop-user/dfs/name has been successfully
>>> formatted.
>>> 10/10/09 22:32:12 INFO namenode.NameNode: SHUTDOWN_MSG:
>>> /************************************************************
>>> SHUTDOWN_MSG: Shutting down NameNode at java.net.UnknownHostException:
>>> cs-sy-249.cse.iitkgp.ernet.in: cs-sy-249.cse.iitkgp.ernet.in
>>> ************************************************************/
>>>
>>> Please help me in solving this problem.
>>>
>>> Thanks
>>> Regards
>>> Siddharth
>>>
>>>
>>>
>>
>>
>
>
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message