hadoop-hdfs-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Björn-Elmar Macek <ma...@cs.uni-kassel.de>
Subject Re: Namenode and Jobtracker dont start
Date Fri, 20 Jul 2012 15:38:20 GMT
Hi Mohammad,

Thanks for your fast reply. Here they are:

\_____________hadoop-env.sh___
I added those 2 lines:

# The java implementation to use.  Required.
export JAVA_HOME=/opt/jdk1.6.0_01/
export JAVA_OPTS="-Djava.net.preferIPv4Stack=true $JAVA_OPTS"


\_____________core-site.xml_____
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

<configuration>
     <property>
         <name>fs.default.name</name>
         <value>hdfs://its-cs100:9005</value>
     </property>
</configuration>


\_____________hdfs-site.xml____
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- configure data paths for masters and slaves -->

<configuration>
     <property>
         <name>dfs.name.dir</name>
         <value>/home/work/bmacek/hadoop/master</value>
     </property>
     <!-- maybe one cannot config masters and slaves on with the same 
file -->
     <property>
         <name>dfs.data.dir</name>
<value>/home/work/bmacek/hadoop/hdfs/slave</value>
     </property>
     <property>
         <name>hadoop.tmp.dir</name>
<value>/home/work/bmacek/hadoop/hdfs/tmp</value>
     </property>

     <property>
         <name>dfs.replication</name>
         <value>1</value>
     </property>
</configuration>


\_______mapred-site.xml____
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

<configuration>
     <!-- master -->
     <property>
         <name>mapred.job.tracker</name>
         <value>its-cs100:9004</value>
     </property>
     <!-- datanode -->
     <property>
         <name>dfs.hosts</name>
<value>/home/fb16/bmacek/hadoop-1.0.2/conf/hosts</value>
     </property>

     <property>
         <name>mapred.hosts</name>
<value>/home/fb16/bmacek/hadoop-1.0.2/conf/hosts</value>
     </property>
</configuration>

\_______masters____
its-cs101

\_______slaves______
its-cs102
its-cs103


Thats about it, i think. I hope i didnt forget anything.

Regards,
Björn-Elmar

Am 20.07.2012 16:58, schrieb Mohammad Tariq:
> Hello sir,
>
>        If possible, could you please paste your config files??
>
> Regards,
>      Mohammad Tariq
>
>
> On Fri, Jul 20, 2012 at 8:24 PM, Björn-Elmar Macek
> <macek@cs.uni-kassel.de> wrote:
>> Hi together,
>>
>> well just stumbled upon this post:
>> http://ankitasblogger.blogspot.de/2012/01/error-that-occured-in-hadoop-and-its.html
>>
>> And it says:
>> "Problem: Hadoop-datanode job failed or datanode not running:
>> java.io.IOException: File ../mapred/system/jobtracker.info could only be
>> replicated to 0 nodes, instead of 1.
>> ...
>> Cause: You may also get this message due to permissions. May be JobTracker
>> can not create jobtracker.info on startup."
>>
>> Since the file does not exist i think, this might be a probable reason for
>> my errors. But why should the JobTracker not be able to create that file. It
>> created several other directories on this node with easy via the slave.sh
>> script that i started with the very same user that calls start-all.sh.
>>
>> Any help would be really appreciated.
>>
>>
>> Am 20.07.2012 16:15, schrieb Björn-Elmar Macek:
>>
>> Hi Srinivas,
>>
>> thanks for your reply! I have been following your link and idea and been
>> playing around alot, but still got problems with the connection (though they
>> are different now):
>>
>> \_______ JAVA VERSION_________
>> "which java" tells me it is 1.6.0_01. If i got it right version 1.7 got
>> problems with ssh.
>>
>> \_______MY TESTS_____________
>> According to your suggestion to look for processes running on that port i
>> changed ports alot:
>> When i was posting the first post of this thread. i was using ports 999 for
>> namenode and 1000 for jobtracker.
>> Since due to some reasons commands like "lsof -i" etc dont give me any
>> output when usedin the cluster enviroment. So i started looking for ports
>> that are in general unused by programs.
>> When i changed the ports to 9004 and 9005 i got different errors which look
>> very much like the ones you posted in the beginning of this year in the
>> lucene section (
>> http://lucene.472066.n3.nabble.com/Unable-to-start-hadoop-0-20-2-but-able-to-start-hadoop-0-20-203-cluster-td2991350.html
>> ).
>>
>> It seems as if a DataNode can not communicate with the NameNode.
>>
>> The logs look like the following:
>>
>> \_______TEST RESULTS__________
>> ########## A DataNode #############
>> 2012-07-20 14:47:59,536 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
>> /************************************************************
>> STARTUP_MSG: Starting DataNode
>> STARTUP_MSG:   host = its-cs102.its.uni-kassel.de/141.51.205.12
>> STARTUP_MSG:   args = []
>> STARTUP_MSG:   version = 1.0.2
>> STARTUP_MSG:   build =
>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0.2 -r
>> 1304954; compiled by 'hortonfo' on Sat Mar 24 23:58:21 UTC 2012
>> ************************************************************/
>> 2012-07-20 14:47:59,824 INFO org.apache.hadoop.metrics2.impl.MetricsConfig:
>> loaded properties from hadoop-metrics2.properties
>> 2012-07-20 14:47:59,841 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>> MetricsSystem,sub=Stats registered.
>> 2012-07-20 14:47:59,843 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period
>> at 10 second(s).
>> 2012-07-20 14:47:59,844 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system
>> started
>> 2012-07-20 14:47:59,969 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi
>> registered.
>> 2012-07-20 14:48:26,792 INFO org.apache.hadoop.ipc.Client: Retrying connect
>> to server: its-cs100/141.51.205.10:9005. Already tried 0 time(s).
>> 2012-07-20 14:48:26,889 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Registered
>> FSDatasetStatusMBean
>> 2012-07-20 14:48:26,934 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Opened info server at 50010
>> 2012-07-20 14:48:26,949 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwith is
>> 1048576 bytes/s
>> 2012-07-20 14:48:27,014 INFO org.mortbay.log: Logging to
>> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
>> org.mortbay.log.Slf4jLog
>> 2012-07-20 14:48:27,147 INFO org.apache.hadoop.http.HttpServer: Added global
>> filtersafety (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
>> 2012-07-20 14:48:27,160 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: dfs.webhdfs.enabled = false
>> 2012-07-20 14:48:27,160 INFO org.apache.hadoop.http.HttpServer: Port
>> returned by webServer.getConnectors()[0].getLocalPort() before open() is -1.
>> Opening the listener on 50075
>> 2012-07-20 14:48:27,160 INFO org.apache.hadoop.http.HttpServer:
>> listener.getLocalPort() returned 50075
>> webServer.getConnectors()[0].getLocalPort() returned 50075
>> 2012-07-20 14:48:27,160 INFO org.apache.hadoop.http.HttpServer: Jetty bound
>> to port 50075
>> 2012-07-20 14:48:27,160 INFO org.mortbay.log: jetty-6.1.26
>> 2012-07-20 14:48:27,805 INFO org.mortbay.log: Started
>> SelectChannelConnector@0.0.0.0:50075
>> 2012-07-20 14:48:27,811 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source jvm
>> registered.
>> 2012-07-20 14:48:27,813 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>> DataNode registered.
>> 2012-07-20 14:48:28,484 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>> RpcDetailedActivityForPort50020 registered.
>> 2012-07-20 14:48:28,484 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>> RpcActivityForPort50020 registered.
>> 2012-07-20 14:48:28,487 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: dnRegistration =
>> DatanodeRegistration(its-cs102.its.uni-kassel.de:50010, storageID=,
>> infoPort=50075, ipcPort=50020)
>> 2012-07-20 14:48:28,489 INFO org.apache.hadoop.ipc.Server: Starting
>> SocketReader
>> 2012-07-20 14:48:38,706 ERROR
>> org.apache.hadoop.hdfs.server.datanode.DataNode:
>> org.apache.hadoop.ipc.RemoteException:
>> org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException: Datanode
>> denied communication with namenode: its-cs102.its.uni-kassel.de:50010
>>      at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.registerDatanode(FSNamesystem.java:2391)
>>      at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.register(NameNode.java:973)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>      at java.lang.reflect.Method.invoke(Method.java:597)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:396)
>>      at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>>
>>      at org.apache.hadoop.ipc.Client.call(Client.java:1066)
>>      at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
>>      at $Proxy5.register(Unknown Source)
>>      at
>> org.apache.hadoop.hdfs.server.datanode.DataNode.register(DataNode.java:673)
>>      at
>> org.apache.hadoop.hdfs.server.datanode.DataNode.runDatanodeDaemon(DataNode.java:1480)
>>      at
>> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1540)
>>      at
>> org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1665)
>>      at
>> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1682)
>>
>> 2012-07-20 14:48:38,712 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG:
>> /************************************************************
>> SHUTDOWN_MSG: Shutting down DataNode at
>> its-cs102.its.uni-kassel.de/141.51.205.12
>> ************************************************************/
>>
>>
>> ####### NameNode ##########################
>> CAUTION: Please recognize, that the file mentioned in the first error log
>> message (/home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info)
>> does not exist on the NameNode, when i checked for it.
>> The only path that has a simiar name is:
>> /home/work/bmacek/hadoop/hdfs/slave/tmp (containing no further subfolders or
>> files)
>>
>>
>>
>> 2012-07-20 14:47:58,033 INFO
>> org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG:
>> /************************************************************
>> STARTUP_MSG: Starting NameNode
>> STARTUP_MSG:   host = its-cs100.its.uni-kassel.de/141.51.205.10
>> STARTUP_MSG:   args = []
>> STARTUP_MSG:   version = 1.0.2
>> STARTUP_MSG:   build =
>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0.2 -r
>> 1304954; compiled by 'hortonfo' on Sat Mar 24 23:58:21 UTC 2012
>> ************************************************************/
>> 2012-07-20 14:47:58,985 INFO org.apache.hadoop.metrics2.impl.MetricsConfig:
>> loaded properties from hadoop-metrics2.properties
>> 2012-07-20 14:47:59,037 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>> MetricsSystem,sub=Stats registered.
>> 2012-07-20 14:47:59,052 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period
>> at 10 second(s).
>> 2012-07-20 14:47:59,052 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system
>> started
>> 2012-07-20 14:47:59,622 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi
>> registered.
>> 2012-07-20 14:47:59,685 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source jvm
>> registered.
>> 2012-07-20 14:47:59,703 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>> NameNode registered.
>> 2012-07-20 14:47:59,896 INFO org.apache.hadoop.hdfs.util.GSet: VM type
>> = 64-bit
>> 2012-07-20 14:47:59,897 INFO org.apache.hadoop.hdfs.util.GSet: 2% max memory
>> = 17.77875 MB
>> 2012-07-20 14:47:59,897 INFO org.apache.hadoop.hdfs.util.GSet: capacity
>> = 2^21 = 2097152 entries
>> 2012-07-20 14:47:59,897 INFO org.apache.hadoop.hdfs.util.GSet:
>> recommended=2097152, actual=2097152
>> 2012-07-20 14:48:00,067 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner=bmacek
>> 2012-07-20 14:48:00,067 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup
>> 2012-07-20 14:48:00,067 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>> isPermissionEnabled=true
>> 2012-07-20 14:48:00,083 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>> dfs.block.invalidate.limit=100
>> 2012-07-20 14:48:00,084 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>> isAccessTokenEnabled=false accessKeyUpdateInterval=0 min(s),
>> accessTokenLifetime=0 min(s)
>> 2012-07-20 14:48:01,573 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered
>> FSNamesystemStateMBean and NameNodeMXBean
>> 2012-07-20 14:48:01,643 INFO
>> org.apache.hadoop.hdfs.server.namenode.NameNode: Caching file names occuring
>> more than 10 times
>> 2012-07-20 14:48:01,686 INFO org.apache.hadoop.hdfs.server.common.Storage:
>> Number of files = 1
>> 2012-07-20 14:48:01,712 INFO org.apache.hadoop.hdfs.server.common.Storage:
>> Number of files under construction = 0
>> 2012-07-20 14:48:01,713 INFO org.apache.hadoop.hdfs.server.common.Storage:
>> Image file of size 112 loaded in 0 seconds.
>> 2012-07-20 14:48:01,796 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Invalid opcode, reached
>> end of edit log Number of transactions found 53
>> 2012-07-20 14:48:01,796 INFO org.apache.hadoop.hdfs.server.common.Storage:
>> Edits file /home/work/bmacek/hadoop/master/current/edits of size 1049092
>> edits # 53 loaded in 0 seconds.
>> 2012-07-20 14:48:01,797 INFO org.apache.hadoop.hdfs.server.common.Storage:
>> Image file of size 861 saved in 0 seconds.
>> 2012-07-20 14:48:02,003 INFO org.apache.hadoop.hdfs.server.common.Storage:
>> Image file of size 861 saved in 0 seconds.
>> 2012-07-20 14:48:02,159 INFO
>> org.apache.hadoop.hdfs.server.namenode.NameCache: initialized with 1 entries
>> 11 lookups
>> 2012-07-20 14:48:02,159 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished loading
>> FSImage in 2135 msecs
>> 2012-07-20 14:48:02,203 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Total number of blocks
>> = 0
>> 2012-07-20 14:48:02,204 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of invalid
>> blocks = 0
>> 2012-07-20 14:48:02,204 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
>> under-replicated blocks = 0
>> 2012-07-20 14:48:02,204 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
>> over-replicated blocks = 0
>> 2012-07-20 14:48:02,204 INFO org.apache.hadoop.hdfs.StateChange: STATE* Safe
>> mode termination scan for invalid, over- and under-replicated blocks
>> completed in 44 msec
>> 2012-07-20 14:48:02,204 INFO org.apache.hadoop.hdfs.StateChange: STATE*
>> Leaving safe mode after 2 secs.
>> 2012-07-20 14:48:02,204 INFO org.apache.hadoop.hdfs.StateChange: STATE*
>> Network topology has 0 racks and 0 datanodes
>> 2012-07-20 14:48:02,205 INFO org.apache.hadoop.hdfs.StateChange: STATE*
>> UnderReplicatedBlocks has 0 blocks
>> 2012-07-20 14:48:02,265 INFO org.apache.hadoop.util.HostsFileReader:
>> Refreshing hosts (include/exclude) list
>> 2012-07-20 14:48:02,275 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: ReplicateQueue
>> QueueProcessingStatistics: First cycle completed 0 blocks in 1 msec
>> 2012-07-20 14:48:02,277 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: ReplicateQueue
>> QueueProcessingStatistics: Queue flush completed 0 blocks in 1 msec
>> processing time, 1 msec clock time, 1 cycles
>> 2012-07-20 14:48:02,277 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: InvalidateQueue
>> QueueProcessingStatistics: First cycle completed 0 blocks in 0 msec
>> 2012-07-20 14:48:02,277 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: InvalidateQueue
>> QueueProcessingStatistics: Queue flush completed 0 blocks in 0 msec
>> processing time, 0 msec clock time, 1 cycles
>> 2012-07-20 14:48:02,281 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>> FSNamesystemMetrics registered.
>> 2012-07-20 14:48:02,336 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>> RpcDetailedActivityForPort9005 registered.
>> 2012-07-20 14:48:02,337 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>> RpcActivityForPort9005 registered.
>> 2012-07-20 14:48:02,341 INFO org.apache.hadoop.ipc.Server: Starting
>> SocketReader
>> 2012-07-20 14:48:02,356 INFO
>> org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode up at:
>> its-cs100.its.uni-kassel.de/141.51.205.10:9005
>> 2012-07-20 14:48:02,878 INFO org.mortbay.log: Logging to
>> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
>> org.mortbay.log.Slf4jLog
>> 2012-07-20 14:48:03,312 INFO org.apache.hadoop.http.HttpServer: Added global
>> filtersafety (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
>> 2012-07-20 14:48:03,426 INFO org.apache.hadoop.http.HttpServer:
>> dfs.webhdfs.enabled = false
>> 2012-07-20 14:48:03,465 INFO org.apache.hadoop.http.HttpServer: Port
>> returned by webServer.getConnectors()[0].getLocalPort() before open() is -1.
>> Opening the listener on 50070
>> 2012-07-20 14:48:03,511 INFO org.apache.hadoop.http.HttpServer:
>> listener.getLocalPort() returned 50070
>> webServer.getConnectors()[0].getLocalPort() returned 50070
>> 2012-07-20 14:48:03,511 INFO org.apache.hadoop.http.HttpServer: Jetty bound
>> to port 50070
>> 2012-07-20 14:48:03,511 INFO org.mortbay.log: jetty-6.1.26
>> 2012-07-20 14:48:06,528 INFO org.mortbay.log: Started
>> SelectChannelConnector@0.0.0.0:50070
>> 2012-07-20 14:48:06,528 INFO
>> org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at:
>> 0.0.0.0:50070
>> 2012-07-20 14:48:06,561 INFO org.apache.hadoop.ipc.Server: IPC Server
>> Responder: starting
>> 2012-07-20 14:48:06,593 INFO org.apache.hadoop.ipc.Server: IPC Server
>> listener on 9005: starting
>> 2012-07-20 14:48:06,656 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 0 on 9005: starting
>> 2012-07-20 14:48:06,685 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 1 on 9005: starting
>> 2012-07-20 14:48:06,731 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 2 on 9005: starting
>> 2012-07-20 14:48:06,759 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 3 on 9005: starting
>> 2012-07-20 14:48:06,791 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 4 on 9005: starting
>> 2012-07-20 14:48:06,849 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 5 on 9005: starting
>> 2012-07-20 14:48:06,874 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 6 on 9005: starting
>> 2012-07-20 14:48:06,898 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 7 on 9005: starting
>> 2012-07-20 14:48:06,921 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 8 on 9005: starting
>> 2012-07-20 14:48:06,974 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 9 on 9005: starting
>> 2012-07-20 14:48:27,222 ERROR
>> org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
>> as:bmacek cause:java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> 2012-07-20 14:48:27,224 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 5 on 9005, call
>> addBlock(/home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info,
>> DFSClient_-1997886712, null) from 141.51.205.10:56513: error:
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>>      at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
>>      at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>      at java.lang.reflect.Method.invoke(Method.java:597)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:396)
>>      at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>> 2012-07-20 14:48:38,701 ERROR
>> org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
>> as:bmacek
>> cause:org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException:
>> Datanode denied communication with namenode:
>> its-cs102.its.uni-kassel.de:50010
>> 2012-07-20 14:48:38,701 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 7 on 9005, call
>> register(DatanodeRegistration(its-cs102.its.uni-kassel.de:50010,
>> storageID=DS-1791721778-141.51.205.12-50010-1342788518692, infoPort=50075,
>> ipcPort=50020)) from 141.51.205.12:33789: error:
>> org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException: Datanode
>> denied communication with namenode: its-cs102.its.uni-kassel.de:50010
>> org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException: Datanode
>> denied communication with namenode: its-cs102.its.uni-kassel.de:50010
>>      at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.registerDatanode(FSNamesystem.java:2391)
>>      at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.register(NameNode.java:973)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>      at java.lang.reflect.Method.invoke(Method.java:597)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:396)
>>      at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>> 2012-07-20 14:48:54,331 ERROR
>> org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
>> as:bmacek cause:java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> 2012-07-20 14:48:54,331 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 2 on 9005, call
>> addBlock(/home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info,
>> DFSClient_-1997886712, null) from 141.51.205.10:56514: error:
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>>      at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
>>      at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>      at java.lang.reflect.Method.invoke(Method.java:597)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:396)
>>      at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>> 2012-07-20 14:49:18,079 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of transactions:
>> 13 Total time for transactions(ms): 22Number of transactions batched in
>> Syncs: 0 Number of syncs: 9 SyncTimes(ms): 111
>> 2012-07-20 14:49:18,151 ERROR
>> org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
>> as:bmacek cause:java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> 2012-07-20 14:49:18,151 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 9 on 9005, call
>> addBlock(/home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info,
>> DFSClient_-1997886712, null) from 141.51.205.10:56515: error:
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>>      at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
>>      at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>      at java.lang.reflect.Method.invoke(Method.java:597)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:396)
>>      at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>> 2012-07-20 14:49:41,419 ERROR
>> org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
>> as:bmacek cause:java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> 2012-07-20 14:49:41,419 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 5 on 9005, call
>> addBlock(/home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info,
>> DFSClient_-1997886712, null) from 141.51.205.10:56516: error:
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>>      at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
>>      at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>      at java.lang.reflect.Method.invoke(Method.java:597)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:396)
>>      at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>> 2012-07-20 14:50:04,474 ERROR
>> org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
>> as:bmacek cause:java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> 2012-07-20 14:50:04,474 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 2 on 9005, call
>> addBlock(/home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info,
>> DFSClient_-1997886712, null) from 141.51.205.10:56517: error:
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>>      at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
>>      at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>      at java.lang.reflect.Method.invoke(Method.java:597)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:396)
>>      at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>> 2012-07-20 14:50:26,299 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of transactions:
>> 25 Total time for transactions(ms): 22Number of transactions batched in
>> Syncs: 0 Number of syncs: 18 SyncTimes(ms): 170
>> 2012-07-20 14:50:26,359 ERROR
>> org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
>> as:bmacek cause:java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> 2012-07-20 14:50:26,359 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 9 on 9005, call
>> addBlock(/home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info,
>> DFSClient_-1997886712, null) from 141.51.205.10:56518: error:
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>>      at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
>>      at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>      at java.lang.reflect.Method.invoke(Method.java:597)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:396)
>>      at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>> 2012-07-20 14:50:47,243 ERROR
>> org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
>> as:bmacek cause:java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> 2012-07-20 14:50:47,243 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 7 on 9005, call
>> addBlock(/home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info,
>> DFSClient_-1997886712, null) from 141.51.205.10:56519: error:
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>>      at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
>>      at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>      at java.lang.reflect.Method.invoke(Method.java:597)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:396)
>>      at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>> 2012-07-20 14:51:06,865 ERROR
>> org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
>> as:bmacek
>> cause:org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException:
>> Datanode denied communication with namenode:
>> its-cs103.its.uni-kassel.de:50010
>> 2012-07-20 14:51:06,865 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 1 on 9005, call
>> register(DatanodeRegistration(its-cs103.its.uni-kassel.de:50010,
>> storageID=DS-1725464844-141.51.205.13-50010-1342788666863, infoPort=50075,
>> ipcPort=50020)) from 141.51.205.13:48227: error:
>> org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException: Datanode
>> denied communication with namenode: its-cs103.its.uni-kassel.de:50010
>> org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException: Datanode
>> denied communication with namenode: its-cs103.its.uni-kassel.de:50010
>>      at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.registerDatanode(FSNamesystem.java:2391)
>>      at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.register(NameNode.java:973)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>      at java.lang.reflect.Method.invoke(Method.java:597)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:396)
>>      at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>> 2012-07-20 14:51:08,305 ERROR
>> org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
>> as:bmacek cause:java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> 2012-07-20 14:51:08,305 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 5 on 9005, call
>> addBlock(/home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info,
>> DFSClient_-1997886712, null) from 141.51.205.10:56520: error:
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>>      at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
>>      at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>      at java.lang.reflect.Method.invoke(Method.java:597)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:396)
>>      at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>> 2012-07-20 14:51:34,855 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of transactions:
>> 37 Total time for transactions(ms): 22Number of transactions batched in
>> Syncs: 0 Number of syncs: 27 SyncTimes(ms): 256
>> 2012-07-20 14:51:34,932 ERROR
>> org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
>> as:bmacek cause:java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> 2012-07-20 14:51:34,932 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 0 on 9005, call
>> addBlock(/home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info,
>> DFSClient_-1997886712, null) from 141.51.205.10:56521: error:
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>>      at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
>>      at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>      at java.lang.reflect.Method.invoke(Method.java:597)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:396)
>>      at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>> 2012-07-20 14:51:57,128 ERROR
>> org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
>> as:bmacek cause:java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> 2012-07-20 14:51:57,128 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 8 on 9005, call
>> addBlock(/home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info,
>> DFSClient_-1997886712, null) from 141.51.205.10:56522: error:
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>>      at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
>>      at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>      at java.lang.reflect.Method.invoke(Method.java:597)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:396)
>>      at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>> 2012-07-20 14:52:21,974 ERROR
>> org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
>> as:bmacek cause:java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> 2012-07-20 14:52:21,976 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 4 on 9005, call
>> addBlock(/home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info,
>> DFSClient_-1997886712, null) from 141.51.205.10:56523: error:
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>>      at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
>>      at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>      at java.lang.reflect.Method.invoke(Method.java:597)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:396)
>>      at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>> 2012-07-20 14:52:43,473 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of transactions:
>> 49 Total time for transactions(ms): 22Number of transactions batched in
>> Syncs: 0 Number of syncs: 36 SyncTimes(ms): 341
>> 2012-07-20 14:52:43,570 ERROR
>> org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
>> as:bmacek cause:java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> 2012-07-20 14:52:43,570 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 0 on 9005, call
>> addBlock(/home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info,
>> DFSClient_-1997886712, null) from 141.51.205.10:56524: error:
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>> java.io.IOException: File
>> /home/work/bmacek/hadoop/hdfs/tmp/mapred/system/jobtracker.info could only
>> be replicated to 0 nodes, instead of 1
>>      at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
>>      at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>      at java.lang.reflect.Method.invoke(Method.java:597)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:396)
>>      at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>>
>>
>>
>>
>>
>> I am not
>>
>> Am 18.07.2012 19:47, schrieb Suresh Srinivas:
>>
>> Can you share information on the java version that you are using.
>> - Is it as obvious as some previous processes still running and new
>> processes cannot bind to the port?
>> - Another pointer -
>> http://stackoverflow.com/questions/8360913/weird-java-net-socketexception-permission-denied-connect-error-when-running-groo
>>
>> On Wed, Jul 18, 2012 at 7:29 AM, Björn-Elmar Macek <macek@cs.uni-kassel.de>
>> wrote:
>>> Hi,
>>>
>>> i have lately been running into problems since i started running hadoop on
>>> a cluster:
>>>
>>> The setup is the following:
>>> 1 Computer is NameNode and Jobtracker
>>> 1 Computer is SecondaryNameNode
>>> 2 Computers are TaskTracker and DataNode
>>>
>>> I ran into problems with running the wordcount example: NameNode and
>>> Jobtracker do not start properly both having connection problems of some
>>> kind.
>>> And this is although ssh is configured that way, that no prompt happens
>>> when i connect from any node in the cluster to any other.
>>>
>>> Is there any reason why this happens?
>>>
>>> The logs look like the following:
>>> \________ JOBTRACKER__________________________________________________
>>> 2012-07-18 16:08:05,808 INFO org.apache.hadoop.mapred.JobTracker:
>>> STARTUP_MSG:
>>> /************************************************************
>>> STARTUP_MSG: Starting JobTracker
>>> STARTUP_MSG:   host = its-cs100.its.uni-kassel.de/141.51.205.10
>>> STARTUP_MSG:   args = []
>>> STARTUP_MSG:   version = 1.0.2
>>> STARTUP_MSG:   build =
>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0.2 -r
>>> 1304954; compiled by 'hortonfo' on Sat Mar 24 23:58:21 UTC 2012
>>> ************************************************************/
>>> 2012-07-18 16:08:06,479 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from
>>> hadoop-metrics2.properties
>>> 2012-07-18 16:08:06,534 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>> MetricsSystem,sub=Stats registered.
>>> 2012-07-18 16:08:06,554 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period
>>> at 10 second(s).
>>> 2012-07-18 16:08:06,554 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: JobTracker metrics system
>>> started
>>> 2012-07-18 16:08:07,157 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>> QueueMetrics,q=default registered.
>>> 2012-07-18 16:08:10,395 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi
>>> registered.
>>> 2012-07-18 16:08:10,417 INFO
>>> org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager:
>>> Updating the current master key for generating delegation tokens
>>> 2012-07-18 16:08:10,436 INFO org.apache.hadoop.mapred.JobTracker:
>>> Scheduler configured with (memSizeForMapSlotOnJT, memSizeForReduceSlotOnJT,
>>> limitMaxMemForMapTasks, limitMaxMemForReduceTasks) (-1, -1, -1, -1)
>>> 2012-07-18 16:08:10,438 INFO org.apache.hadoop.util.HostsFileReader:
>>> Refreshing hosts (include/exclude) list
>>> 2012-07-18 16:08:10,440 INFO
>>> org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager:
>>> Starting expired delegation token remover thread,
>>> tokenRemoverScanInterval=60 min(s)
>>> 2012-07-18 16:08:10,465 INFO
>>> org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager:
>>> Updating the current master key for generating delegation tokens
>>> 2012-07-18 16:08:10,510 INFO org.apache.hadoop.mapred.JobTracker: Starting
>>> jobtracker with owner as bmacek
>>> 2012-07-18 16:08:10,620 WARN org.apache.hadoop.mapred.JobTracker: Error
>>> starting tracker: java.net.SocketException: Permission denied
>>>      at sun.nio.ch.Net.bind(Native Method)
>>>      at
>>> sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:119)
>>>      at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:59)
>>>      at org.apache.hadoop.ipc.Server.bind(Server.java:225)
>>>      at org.apache.hadoop.ipc.Server$Listener.<init>(Server.java:301)
>>>      at org.apache.hadoop.ipc.Server.<init>(Server.java:1483)
>>>      at org.apache.hadoop.ipc.RPC$Server.<init>(RPC.java:545)
>>>      at org.apache.hadoop.ipc.RPC.getServer(RPC.java:506)
>>>      at org.apache.hadoop.mapred.JobTracker.<init>(JobTracker.java:2306)
>>>      at org.apache.hadoop.mapred.JobTracker.<init>(JobTracker.java:2192)
>>>      at org.apache.hadoop.mapred.JobTracker.<init>(JobTracker.java:2186)
>>>      at
>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:300)
>>>      at
>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:291)
>>>      at org.apache.hadoop.mapred.JobTracker.main(JobTracker.java:4978)
>>>
>>> 2012-07-18 16:08:13,861 WARN
>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Source name
>>> QueueMetrics,q=default already exists!
>>> 2012-07-18 16:08:13,885 WARN
>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Source name ugi already
>>> exists!
>>> 2012-07-18 16:08:13,885 INFO
>>> org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager:
>>> Updating the current master key for generating delegation tokens
>>> 2012-07-18 16:08:13,910 INFO org.apache.hadoop.mapred.JobTracker:
>>> Scheduler configured with (memSizeForMapSlotOnJT, memSizeForReduceSlotOnJT,
>>> limitMaxMemForMapTasks, limitMaxMemForReduceTasks) (-1, -1, -1, -1)
>>> 2012-07-18 16:08:13,911 INFO org.apache.hadoop.util.HostsFileReader:
>>> Refreshing hosts (include/exclude) list
>>> 2012-07-18 16:08:13,911 INFO org.apache.hadoop.mapred.JobTracker: Starting
>>> jobtracker with owner as bmacek
>>> 2012-07-18 16:08:13,912 WARN org.apache.hadoop.mapred.JobTracker: Error
>>> starting tracker: java.net.SocketException: Permission denied
>>>      at sun.nio.ch.Net.bind(Native Method)
>>>      at
>>> sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:119)
>>>      at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:59)
>>>      at org.apache.hadoop.ipc.Server.bind(Server.java:225)
>>>      at org.apache.hadoop.ipc.Server$Listener.<init>(Server.java:301)
>>>      at org.apache.hadoop.ipc.Server.<init>(Server.java:1483)
>>>      at org.apache.hadoop.ipc.RPC$Server.<init>(RPC.java:545)
>>>      at org.apache.hadoop.ipc.RPC.getServer(RPC.java:506)
>>>      at org.apache.hadoop.mapred.JobTracker.<init>(JobTracker.java:2306)
>>>      at org.apache.hadoop.mapred.JobTracker.<init>(JobTracker.java:2192)
>>>      at org.apache.hadoop.mapred.JobTracker.<init>(JobTracker.java:2186)
>>>      at
>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:300)
>>>      at
>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:291)
>>>      at org.apache.hadoop.mapred.JobTracker.main(JobTracker.java:4978)
>>>
>>> 2012-07-18 16:08:13,912 INFO
>>> org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager:
>>> Starting expired delegation token remover thread,
>>> tokenRemoverScanInterval=60 min(s)
>>> 2012-07-18 16:08:13,913 INFO
>>> org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager:
>>> Updating the current master key for generating delegation tokens
>>> 2012-07-18 16:08:21,348 WARN
>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Source name
>>> QueueMetrics,q=default already exists!
>>> 2012-07-18 16:08:21,390 WARN
>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Source name ugi already
>>> exists!
>>> 2012-07-18 16:08:21,390 INFO
>>> org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager:
>>> Updating the current master key for generating delegation tokens
>>> 2012-07-18 16:08:21,426 INFO org.apache.hadoop.mapred.JobTracker:
>>> Scheduler configured with (memSizeForMapSlotOnJT, memSizeForReduceSlotOnJT,
>>> limitMaxMemForMapTasks, limitMaxMemForReduceTasks) (-1, -1, -1, -1)
>>> 2012-07-18 16:08:21,427 INFO org.apache.hadoop.util.HostsFileReader:
>>> Refreshing hosts (include/exclude) list
>>> 2012-07-18 16:08:21,427 INFO org.apache.hadoop.mapred.JobTracker: Starting
>>> jobtracker with owner as bmacek
>>> 2012-07-18 16:08:21,428 WARN org.apache.hadoop.mapred.JobTracker: Error
>>> starting tracker: java.net.SocketException: Permission denied
>>>      at sun.nio.ch.Net.bind(Native Method)
>>>      at
>>> sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:119)
>>>      at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:59)
>>>      at org.apache.hadoop.ipc.Server.bind(Server.java:225)
>>>      at org.apache.hadoop.ipc.Server$Listener.<init>(Server.java:301)
>>>      at org.apache.hadoop.ipc.Server.<init>(Server.java:1483)
>>>      at org.apache.hadoop.ipc.RPC$Server.<init>(RPC.java:545)
>>>      at org.apache.hadoop.ipc.RPC.getServer(RPC.java:506)
>>>      at org.apache.hadoop.mapred.JobTracker.<init>(JobTracker.java:2306)
>>>      at org.apache.hadoop.mapred.JobTracker.<init>(JobTracker.java:2192)
>>>      at org.apache.hadoop.mapred.JobTracker.<init>(JobTracker.java:2186)
>>>      at
>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:300)
>>>      at
>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:291)
>>>      at org.apache.hadoop.mapred.JobTracker.main(JobTracker.java:4978)
>>>
>>>
>>> \________ DATANODE__________________________________________________
>>> 2012-07-18 16:07:58,759 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG:
>>> /************************************************************
>>> STARTUP_MSG: Starting NameNode
>>> STARTUP_MSG:   host = its-cs100.its.uni-kassel.de/141.51.205.10
>>> STARTUP_MSG:   args = []
>>> STARTUP_MSG:   version = 1.0.2
>>> STARTUP_MSG:   build =
>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0.2 -r
>>> 1304954; compiled by 'hortonfo' on Sat Mar 24 23:58:21 UTC 2012
>>> ************************************************************/
>>> 2012-07-18 16:07:59,738 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from
>>> hadoop-metrics2.properties
>>> 2012-07-18 16:07:59,790 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>> MetricsSystem,sub=Stats registered.
>>> 2012-07-18 16:07:59,807 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period
>>> at 10 second(s).
>>> 2012-07-18 16:07:59,807 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system
>>> started
>>> 2012-07-18 16:08:00,382 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi
>>> registered.
>>> 2012-07-18 16:08:00,454 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source jvm
>>> registered.
>>> 2012-07-18 16:08:00,456 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>> NameNode registered.
>>> 2012-07-18 16:08:00,645 INFO org.apache.hadoop.hdfs.util.GSet: VM type
>>> = 64-bit
>>> 2012-07-18 16:08:00,645 INFO org.apache.hadoop.hdfs.util.GSet: 2% max
>>> memory = 17.77875 MB
>>> 2012-07-18 16:08:00,645 INFO org.apache.hadoop.hdfs.util.GSet: capacity
>>> = 2^21 = 2097152 entries
>>> 2012-07-18 16:08:00,645 INFO org.apache.hadoop.hdfs.util.GSet:
>>> recommended=2097152, actual=2097152
>>> 2012-07-18 16:08:00,812 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner=bmacek
>>> 2012-07-18 16:08:00,812 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup
>>> 2012-07-18 16:08:00,824 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>> isPermissionEnabled=true
>>> 2012-07-18 16:08:00,846 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>> dfs.block.invalidate.limit=100
>>> 2012-07-18 16:08:00,846 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>> isAccessTokenEnabled=false accessKeyUpdateInterval=0 min(s),
>>> accessTokenLifetime=0 min(s)
>>> 2012-07-18 16:08:02,746 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered
>>> FSNamesystemStateMBean and NameNodeMXBean
>>> 2012-07-18 16:08:02,868 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: Caching file names occuring
>>> more than 10 times
>>> 2012-07-18 16:08:02,932 INFO org.apache.hadoop.hdfs.server.common.Storage:
>>> Number of files = 1
>>> 2012-07-18 16:08:02,963 INFO org.apache.hadoop.hdfs.server.common.Storage:
>>> Number of files under construction = 0
>>> 2012-07-18 16:08:02,966 INFO org.apache.hadoop.hdfs.server.common.Storage:
>>> Image file of size 112 loaded in 0 seconds.
>>> 2012-07-18 16:08:02,975 INFO org.apache.hadoop.hdfs.server.common.Storage:
>>> Edits file /home/work/bmacek/hadoop/master/current/edits of size 4 edits # 0
>>> loaded in 0 seconds.
>>> 2012-07-18 16:08:02,977 INFO org.apache.hadoop.hdfs.server.common.Storage:
>>> Image file of size 112 saved in 0 seconds.
>>> 2012-07-18 16:08:03,191 INFO org.apache.hadoop.hdfs.server.common.Storage:
>>> Image file of size 112 saved in 0 seconds.
>>> 2012-07-18 16:08:03,334 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameCache: initialized with 0 entries
>>> 0 lookups
>>> 2012-07-18 16:08:03,334 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished loading
>>> FSImage in 2567 msecs
>>> 2012-07-18 16:08:03,401 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Total number of blocks
>>> = 0
>>> 2012-07-18 16:08:03,401 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of invalid
>>> blocks = 0
>>> 2012-07-18 16:08:03,401 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
>>> under-replicated blocks = 0
>>> 2012-07-18 16:08:03,401 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
>>> over-replicated blocks = 0
>>> 2012-07-18 16:08:03,401 INFO org.apache.hadoop.hdfs.StateChange: STATE*
>>> Safe mode termination scan for invalid, over- and under-replicated blocks
>>> completed in 61 msec
>>> 2012-07-18 16:08:03,402 INFO org.apache.hadoop.hdfs.StateChange: STATE*
>>> Leaving safe mode after 2 secs.
>>> 2012-07-18 16:08:03,412 INFO org.apache.hadoop.hdfs.StateChange: STATE*
>>> Network topology has 0 racks and 0 datanodes
>>> 2012-07-18 16:08:03,412 INFO org.apache.hadoop.hdfs.StateChange: STATE*
>>> UnderReplicatedBlocks has 0 blocks
>>> 2012-07-18 16:08:03,472 INFO org.apache.hadoop.util.HostsFileReader:
>>> Refreshing hosts (include/exclude) list
>>> 2012-07-18 16:08:03,488 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: ReplicateQueue
>>> QueueProcessingStatistics: First cycle completed 0 blocks in 1 msec
>>> 2012-07-18 16:08:03,490 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: ReplicateQueue
>>> QueueProcessingStatistics: Queue flush completed 0 blocks in 1 msec
>>> processing time, 1 msec clock time, 1 cycles
>>> 2012-07-18 16:08:03,490 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: InvalidateQueue
>>> QueueProcessingStatistics: First cycle completed 0 blocks in 0 msec
>>> 2012-07-18 16:08:03,490 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: InvalidateQueue
>>> QueueProcessingStatistics: Queue flush completed 0 blocks in 0 msec
>>> processing time, 0 msec clock time, 1 cycles
>>> 2012-07-18 16:08:03,495 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>> FSNamesystemMetrics registered.
>>> 2012-07-18 16:08:03,553 WARN
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: ReplicationMonitor
>>> thread received InterruptedException.java.lang.InterruptedException: sleep
>>> interrupted
>>> 2012-07-18 16:08:03,555 INFO
>>> org.apache.hadoop.hdfs.server.namenode.DecommissionManager: Interrupted
>>> Monitor
>>> java.lang.InterruptedException: sleep interrupted
>>>      at java.lang.Thread.sleep(Native Method)
>>>      at
>>> org.apache.hadoop.hdfs.server.namenode.DecommissionManager$Monitor.run(DecommissionManager.java:65)
>>>      at java.lang.Thread.run(Thread.java:619)
>>> 2012-07-18 16:08:03,556 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of transactions:
>>> 0 Total time for transactions(ms): 0Number of transactions batched in Syncs:
>>> 0 Number of syncs: 0 SyncTimes(ms): 0
>>> 2012-07-18 16:08:03,594 ERROR
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: java.net.SocketException:
>>> Permission denied
>>>      at sun.nio.ch.Net.bind(Native Method)
>>>      at
>>> sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:119)
>>>      at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:59)
>>>      at org.apache.hadoop.ipc.Server.bind(Server.java:225)
>>>      at org.apache.hadoop.ipc.Server$Listener.<init>(Server.java:301)
>>>      at org.apache.hadoop.ipc.Server.<init>(Server.java:1483)
>>>      at org.apache.hadoop.ipc.RPC$Server.<init>(RPC.java:545)
>>>      at org.apache.hadoop.ipc.RPC.getServer(RPC.java:506)
>>>      at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:294)
>>>      at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:496)
>>>      at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1279)
>>>      at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1288)
>>>
>>> 2012-07-18 16:08:03,627 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
>>> /************************************************************
>>> SHUTDOWN_MSG: Shutting down NameNode at
>>> its-cs100.its.uni-kassel.de/141.51.205.10
>>> ************************************************************/
>>
>>
>>
>> --
>> http://hortonworks.com/download/
>>
>>
>>
>>
>>



Mime
View raw message