hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From José Luis Larroque <larroques...@gmail.com>
Subject Re: Can't run hadoop examples with YARN Single node cluster
Date Mon, 07 Mar 2016 21:50:49 GMT
Hi again guys, i could, finally, find what the issue was!!!

This is my mapred-site.xml, here it's the problem :
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<configuration>
<property>
<name>mapred.job.tracker</name>
<!--
<value>local</local> Para debug
<value>hdnode01:54311</value> Para cosas posta
-->
<value>hdnode01:54311</value>
</property>

<property>
<name>mapred.tasktracker.map.tasks.maximum</name>
<value>4</value>
</property>

<property>
<name>mapreduce.job.maps</name>
<value>4</value>
</property>

<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>

<property>
<name>mapreduce.map.memory.mb</name>
<value>512</value>
</property>

<property>
<name>mapreduce.reduce.memory.mb</name>
<value>512</value>
</property>

<property>
<name>mapreduce.map.java.opts</name>
<value>256</value>
</property>

<property>
<name>mapreduce.reduce.java.opts</name>
<value>256</value>
</property>
<configuration>

If i suppress the last two properties ( mapreduce.map.java.opts ,
mapreduce.reduce.java.opts ), wordcount works!

I remember putting those last two properties for a memory issue of some
kind, but maybe for some reason they clash with the others two (
mapreduce.map.memory.mb, mapreduce.reduce.memory.mb) ?

It will be great if someone can give me a short explanation in order to
understand better the memory management of a YARN cluster.


PD: Thanks again Namikaze and Gaurav for their help!!

Bye!
Jose

2016-01-25 21:19 GMT-03:00 José Luis Larroque <larroquester@gmail.com>:

> Thanks Namikaze for keep trying, don't give up!! :D
>
> - I have these lines in *$HOME/.bashrc*
>
>
> export HADOOP_PREFIX=/usr/local/hadoop
>
> # Others variables
>
> export HADOOP_COMMON_HOME=${HADOOP_PREFIX}
>
> export HADOOP_MAPRED_HOME=${HADOOP_PREFIX}
>
> export HADOOP_HDFS_HOME=${HADOOP_PREFIX}
>
> export HADOOP_YARN_HOME=${HADOOP_PREFIX}
>
>
>   - in *hadoop-env.sh* i have:
>
> export HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/usr/local/hadoop/etc/hadoop"}
>
>
>   - I read that SO question and all answers to it. The only useful answer
> in my opinion was checking yarn classpath. I have three times the following
> line:
>
> /usr/local/hadoop/etc/hadoop:
>
>
> I put yarn.application.classpath on yarn-site.xml because i don't know any
> other way to fix it, with the value recomended for default in this
> <https://hadoop.apache.org/docs/r2.4.1/hadoop-yarn/hadoop-yarn-common/yarn-default.xml>
> (see for yarn.application.classpath):
>
>
> $HADOOP_CONF_DIR, $HADOOP_COMMON_HOME/share/hadoop/common/*,
> $HADOOP_COMMON_HOME/share/hadoop/common/lib/*, $HADOOP_HDFS_HOME/share/hadoop/hdfs/*,
> $HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*, $HADOOP_YARN_HOME/share/hadoop/yarn/*,
> $HADOOP_YARN_HOME/share/hadoop/yarn/lib/*
>
>
> But the classpath remains the same. And i can't find any other way to fix
> it. Maybe this is the problem?
>
>
>  - yarn.log-aggregation-enable was always set to true. I couldn't find
> nothing in *datanodes logs*, here they are:
>
> 2016-01-25 21:13:07,006 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
> /************************************************************
> STARTUP_MSG: Starting DataNode
> STARTUP_MSG:   host = jose-ubuntu/127.0.0.1
> STARTUP_MSG:   args = []
> STARTUP_MSG:   version = 2.4.0
> STARTUP_MSG:   classpath = /usr/local/hadoop/etc/hadoop:/usr/local/hadoop/share/hadoop/common/lib/jackson-jaxrs-1.8.8.jar:/usr/local/hadoop/share/hadoop/common/lib/jersey-json-1.9.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-compress-1.4.1.jar:/usr/local/hadoop/share/hadoop/common/lib/httpcore-4.2.5.jar:/usr/local/hadoop/share/hadoop/common/lib/httpclient-4.2.5.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-beanutils-1.7.0.jar:/usr/local/hadoop/share/hadoop/common/lib/jettison-1.1.jar:/usr/local/hadoop/share/hadoop/common/lib/jaxb-impl-2.2.3-1.jar:/usr/local/hadoop/share/hadoop/common/lib/activation-1.1.jar:/usr/local/hadoop/share/hadoop/common/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-net-3.1.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-cli-1.2.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-codec-1.4.jar:/usr/local/hadoop/share/hadoop/common/lib/junit-4.8.2.jar:/usr/local/hadoop/share/hadoop/common/lib/paranamer-2.3.jar:/usr/local/hadoop/share/hadoop/common/lib/jasper-compiler-5.5.23.jar:/usr/local/hadoop/share/hadoop/common/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-digester-1.8.jar:/usr/local/hadoop/share/hadoop/common/lib/jets3t-0.9.0.jar:/usr/local/hadoop/share/hadoop/common/lib/jackson-xc-1.8.8.jar:/usr/local/hadoop/share/hadoop/common/lib/jersey-server-1.9.jar:/usr/local/hadoop/share/hadoop/common/lib/jetty-util-6.1.26.jar:/usr/local/hadoop/share/hadoop/common/lib/mockito-all-1.8.5.jar:/usr/local/hadoop/share/hadoop/common/lib/xmlenc-0.52.jar:/usr/local/hadoop/share/hadoop/common/lib/slf4j-api-1.7.5.jar:/usr/local/hadoop/share/hadoop/common/lib/stax-api-1.0-2.jar:/usr/local/hadoop/share/hadoop/common/lib/avro-1.7.4.jar:/usr/local/hadoop/share/hadoop/common/lib/hadoop-annotations-2.4.0.jar:/usr/local/hadoop/share/hadoop/common/lib/hadoop-auth-2.4.0.jar:/usr/local/hadoop/share/hadoop/common/lib/netty-3.6.2.Final.jar:/usr/local/hadoop/share/hadoop/common/lib/jasper-runtime-5.5.23.jar:/usr/local/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-logging-1.1.3.jar:/usr/local/hadoop/share/hadoop/common/lib/guava-11.0.2.jar:/usr/local/hadoop/share/hadoop/common/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop/share/hadoop/common/lib/jsp-api-2.1.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-configuration-1.6.jar:/usr/local/hadoop/share/hadoop/common/lib/snappy-java-1.0.4.1.jar:/usr/local/hadoop/share/hadoop/common/lib/log4j-1.2.17.jar:/usr/local/hadoop/share/hadoop/common/lib/jsch-0.1.42.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-io-2.4.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-beanutils-core-1.8.0.jar:/usr/local/hadoop/share/hadoop/common/lib/jsr305-1.3.9.jar:/usr/local/hadoop/share/hadoop/common/lib/jersey-core-1.9.jar:/usr/local/hadoop/share/hadoop/common/lib/asm-3.2.jar:/usr/local/hadoop/share/hadoop/common/lib/servlet-api-2.5.jar:/usr/local/hadoop/share/hadoop/common/lib/jaxb-api-2.2.2.jar:/usr/local/hadoop/share/hadoop/common/lib/jetty-6.1.26.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-lang-2.6.jar:/usr/local/hadoop/share/hadoop/common/lib/java-xmlbuilder-0.4.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-collections-3.2.1.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-el-1.0.jar:/usr/local/hadoop/share/hadoop/common/lib/zookeeper-3.4.5.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-httpclient-3.1.jar:/usr/local/hadoop/share/hadoop/common/lib/xz-1.0.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-math3-3.1.1.jar:/usr/local/hadoop/share/hadoop/common/hadoop-common-2.4.0-tests.jar:/usr/local/hadoop/share/hadoop/common/hadoop-nfs-2.4.0.jar:/usr/local/hadoop/share/hadoop/common/hadoop-common-2.4.0.jar:/usr/local/hadoop/share/hadoop/hdfs:/usr/local/hadoop/share/hadoop/hdfs/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-cli-1.2.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-codec-1.4.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jersey-server-1.9.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jetty-util-6.1.26.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/xmlenc-0.52.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-daemon-1.0.13.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/netty-3.6.2.Final.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jasper-runtime-5.5.23.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-logging-1.1.3.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/guava-11.0.2.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jsp-api-2.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/log4j-1.2.17.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-io-2.4.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jsr305-1.3.9.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jersey-core-1.9.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/asm-3.2.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/servlet-api-2.5.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jetty-6.1.26.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-lang-2.6.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-el-1.0.jar:/usr/local/hadoop/share/hadoop/hdfs/hadoop-hdfs-2.4.0-tests.jar:/usr/local/hadoop/share/hadoop/hdfs/hadoop-hdfs-2.4.0.jar:/usr/local/hadoop/share/hadoop/hdfs/hadoop-hdfs-nfs-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jackson-jaxrs-1.8.8.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jersey-json-1.9.jar:/usr/local/hadoop/share/hadoop/yarn/lib/commons-compress-1.4.1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jersey-client-1.9.jar:/usr/local/hadoop/share/hadoop/yarn/lib/guice-servlet-3.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/giraph-examples-1.1.0-for-hadoop-2.4.0-jar-with-dependencies.jar:/usr/local/hadoop/share/hadoop/yarn/lib/hadoop-mapreduce-examples-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jettison-1.1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jaxb-impl-2.2.3-1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/activation-1.1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop/share/hadoop/yarn/lib/commons-cli-1.2.jar:/usr/local/hadoop/share/hadoop/yarn/lib/commons-codec-1.4.jar:/usr/local/hadoop/share/hadoop/yarn/lib/aopalliance-1.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/leveldbjni-all-1.8.jar:/usr/local/hadoop/share/hadoop/yarn/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jackson-xc-1.8.8.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jersey-server-1.9.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jetty-util-6.1.26.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jline-0.9.94.jar:/usr/local/hadoop/share/hadoop/yarn/lib/stax-api-1.0-2.jar:/usr/local/hadoop/share/hadoop/yarn/lib/guice-3.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/commons-logging-1.1.3.jar:/usr/local/hadoop/share/hadoop/yarn/lib/guava-11.0.2.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop/share/hadoop/yarn/lib/log4j-1.2.17.jar:/usr/local/hadoop/share/hadoop/yarn/lib/commons-io-2.4.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jsr305-1.3.9.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jersey-core-1.9.jar:/usr/local/hadoop/share/hadoop/yarn/lib/asm-3.2.jar:/usr/local/hadoop/share/hadoop/yarn/lib/servlet-api-2.5.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jaxb-api-2.2.2.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jetty-6.1.26.jar:/usr/local/hadoop/share/hadoop/yarn/lib/commons-lang-2.6.jar:/usr/local/hadoop/share/hadoop/yarn/lib/commons-collections-3.2.1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jersey-guice-1.9.jar:/usr/local/hadoop/share/hadoop/yarn/lib/zookeeper-3.4.5.jar:/usr/local/hadoop/share/hadoop/yarn/lib/commons-httpclient-3.1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/javax.inject-1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/xz-1.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-common-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-client-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-api-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-applications-unmanaged-am-launcher-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-nodemanager-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-tests-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-resourcemanager-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-web-proxy-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-applications-distributedshell-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-applicationhistoryservice-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-common-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/commons-compress-1.4.1.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/guice-servlet-3.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/hamcrest-core-1.1.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/paranamer-2.3.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/aopalliance-1.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/jersey-server-1.9.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/avro-1.7.4.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/hadoop-annotations-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/guice-3.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/junit-4.10.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/netty-3.6.2.Final.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/snappy-java-1.0.4.1.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/log4j-1.2.17.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/commons-io-2.4.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/jersey-core-1.9.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/asm-3.2.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/jersey-guice-1.9.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/javax.inject-1.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/xz-1.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/giraph-examples-1.1.0-for-hadoop-2.4.0-jar-with-dependencies.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/procesamiento_salida_grafo_caminos_navegacionales-0.0.1.jar:/usr/local/hadoop/share/hadoop/mapreduce/zookeeper-3.4.6.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-shuffle-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/procesamiento_grafo_wikiquote-0.0.1.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-app-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-common-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-core-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/generacion_grafo_wikiquote-0.0.1-SNAPSHOT.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.4.0-tests.jar:/usr/local/hadoop/contrib/capacity-scheduler/*.jar:/usr/local/hadoop/contrib/capacity-scheduler/*.jar:/usr/local/hadoop/contrib/capacity-scheduler/*.jar
> STARTUP_MSG:   build = http://svn.apache.org/repos/asf/hadoop/common -r 1583262; compiled by 'jenkins' on 2014-03-31T08:29Z
> STARTUP_MSG:   java = 1.7.0_79
> ************************************************************/
> 2016-01-25 21:13:07,015 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: registered UNIX signal handlers for [TERM, HUP, INT]
> 2016-01-25 21:13:07,188 WARN org.apache.hadoop.hdfs.server.common.Util: Path /usr/local/hadoop/dfs/name/data should be specified as a URI in configuration files. Please update hdfs configuration.
> 2016-01-25 21:13:07,356 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
> 2016-01-25 21:13:07,648 INFO org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from hadoop-metrics2.properties
> 2016-01-25 21:13:07,723 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s).
> 2016-01-25 21:13:07,723 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system started
> 2016-01-25 21:13:07,727 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Configured hostname is localhost
> 2016-01-25 21:13:07,728 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Starting DataNode with maxLockedMemory = 0
> 2016-01-25 21:13:07,757 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Opened streaming server at /0.0.0.0:50010
> 2016-01-25 21:13:07,760 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwith is 1048576 bytes/s
> 2016-01-25 21:13:07,839 INFO org.mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog
> 2016-01-25 21:13:07,843 INFO org.apache.hadoop.http.HttpRequestLog: Http request log for http.requests.datanode is not defined
> 2016-01-25 21:13:07,853 INFO org.apache.hadoop.http.HttpServer2: Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)
> 2016-01-25 21:13:07,856 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context datanode
> 2016-01-25 21:13:07,856 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs
> 2016-01-25 21:13:07,856 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context static
> 2016-01-25 21:13:07,872 INFO org.apache.hadoop.http.HttpServer2: addJerseyResourcePackage: packageName=org.apache.hadoop.hdfs.server.datanode.web.resources;org.apache.hadoop.hdfs.web.resources, pathSpec=/webhdfs/v1/*
> 2016-01-25 21:13:07,875 INFO org.apache.hadoop.http.HttpServer2: Jetty bound to port 50075
> 2016-01-25 21:13:07,875 INFO org.mortbay.log: jetty-6.1.26
> 2016-01-25 21:13:08,137 INFO org.mortbay.log: Started SelectChannelConnector@0.0.0.0:50075
> 2016-01-25 21:13:08,225 INFO org.apache.hadoop.ipc.CallQueueManager: Using callQueue class java.util.concurrent.LinkedBlockingQueue
> 2016-01-25 21:13:08,239 INFO org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 50020
> 2016-01-25 21:13:08,288 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Opened IPC server at /0.0.0.0:50020
> 2016-01-25 21:13:08,300 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Refresh request received for nameservices: null
> 2016-01-25 21:13:08,316 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Starting BPOfferServices for nameservices: <default>
> 2016-01-25 21:13:08,321 WARN org.apache.hadoop.hdfs.server.common.Util: Path /usr/local/hadoop/dfs/name/data should be specified as a URI in configuration files. Please update hdfs configuration.
> 2016-01-25 21:13:08,325 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool <registering> (Datanode Uuid unassigned) service to hdnode01/192.168.0.10:54310 starting to offer service
> 2016-01-25 21:13:08,326 INFO org.apache.hadoop.ipc.Server: IPC Server Responder: starting
> 2016-01-25 21:13:08,326 INFO org.apache.hadoop.ipc.Server: IPC Server listener on 50020: starting
> 2016-01-25 21:13:08,719 INFO org.apache.hadoop.hdfs.server.common.Storage: Data-node version: -55 and name-node layout version: -56
> 2016-01-25 21:13:08,828 INFO org.apache.hadoop.hdfs.server.common.Storage: Lock on /usr/local/hadoop/dfs/name/data/in_use.lock acquired by nodename 10365@jose-ubuntu
> 2016-01-25 21:13:08,833 INFO org.apache.hadoop.hdfs.server.common.Storage: Storage directory /usr/local/hadoop/dfs/name/data is not formatted
> 2016-01-25 21:13:08,833 INFO org.apache.hadoop.hdfs.server.common.Storage: Formatting ...
> 2016-01-25 21:13:09,017 INFO org.apache.hadoop.hdfs.server.common.Storage: Analyzing storage directories for bpid BP-216406264-127.0.0.1-1453767164845
> 2016-01-25 21:13:09,017 INFO org.apache.hadoop.hdfs.server.common.Storage: Locking is disabled
> 2016-01-25 21:13:09,018 INFO org.apache.hadoop.hdfs.server.common.Storage: Storage directory /usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845 is not formatted.
> 2016-01-25 21:13:09,018 INFO org.apache.hadoop.hdfs.server.common.Storage: Formatting ...
> 2016-01-25 21:13:09,018 INFO org.apache.hadoop.hdfs.server.common.Storage: Formatting block pool BP-216406264-127.0.0.1-1453767164845 directory /usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current
> 2016-01-25 21:13:09,072 INFO org.apache.hadoop.hdfs.server.common.Storage: Restored 0 block files from trash.
> 2016-01-25 21:13:09,198 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Setting up storage: nsid=1479061672;bpid=BP-216406264-127.0.0.1-1453767164845;lv=-55;nsInfo=lv=-56;cid=CID-8fa0e75b-6942-452a-a4e6-8cd0c24de652;nsid=1479061672;c=0;bpid=BP-216406264-127.0.0.1-1453767164845;dnuuid=null
> 2016-01-25 21:13:09,248 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Generated and persisted new Datanode UUID 6b4236c8-2183-49ba-84d7-a273298ba37a
> 2016-01-25 21:13:09,268 WARN org.apache.hadoop.hdfs.server.common.Util: Path /usr/local/hadoop/dfs/name/data should be specified as a URI in configuration files. Please update hdfs configuration.
> 2016-01-25 21:13:09,270 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Added volume - /usr/local/hadoop/dfs/name/data/current, StorageType: DISK
> 2016-01-25 21:13:09,279 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Registered FSDatasetState MBean
> 2016-01-25 21:13:09,282 INFO org.apache.hadoop.hdfs.server.datanode.DirectoryScanner: Periodic Directory Tree Verification scan starting at 1453784080282 with interval 21600000
> 2016-01-25 21:13:09,283 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Adding block pool BP-216406264-127.0.0.1-1453767164845
> 2016-01-25 21:13:09,284 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Scanning block pool BP-216406264-127.0.0.1-1453767164845 on volume /usr/local/hadoop/dfs/name/data/current...
> 2016-01-25 21:13:09,299 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Time taken to scan block pool BP-216406264-127.0.0.1-1453767164845 on /usr/local/hadoop/dfs/name/data/current: 15ms
> 2016-01-25 21:13:09,300 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Total time to scan all replicas for block pool BP-216406264-127.0.0.1-1453767164845: 17ms
> 2016-01-25 21:13:09,301 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Adding replicas to map for block pool BP-216406264-127.0.0.1-1453767164845 on volume /usr/local/hadoop/dfs/name/data/current...
> 2016-01-25 21:13:09,301 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Time to add replicas to map for block pool BP-216406264-127.0.0.1-1453767164845 on volume /usr/local/hadoop/dfs/name/data/current: 0ms
> 2016-01-25 21:13:09,301 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Total time to add all replicas to map: 1ms
> 2016-01-25 21:13:09,305 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool BP-216406264-127.0.0.1-1453767164845 (Datanode Uuid null) service to hdnode01/192.168.0.10:54310 beginning handshake with NN
> 2016-01-25 21:13:09,355 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool Block pool BP-216406264-127.0.0.1-1453767164845 (Datanode Uuid null) service to hdnode01/192.168.0.10:54310 successfully registered with NN
> 2016-01-25 21:13:09,356 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: For namenode hdnode01/192.168.0.10:54310 using DELETEREPORT_INTERVAL of 300000 msec  BLOCKREPORT_INTERVAL of 21600000msec CACHEREPORT_INTERVAL of 10000msec Initial delay: 0msec; heartBeatInterval=3000
> 2016-01-25 21:13:09,444 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Namenode Block pool BP-216406264-127.0.0.1-1453767164845 (Datanode Uuid 6b4236c8-2183-49ba-84d7-a273298ba37a) service to hdnode01/192.168.0.10:54310 trying to claim ACTIVE state with txid=1
> 2016-01-25 21:13:09,444 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Acknowledging ACTIVE Namenode Block pool BP-216406264-127.0.0.1-1453767164845 (Datanode Uuid 6b4236c8-2183-49ba-84d7-a273298ba37a) service to hdnode01/192.168.0.10:54310
> 2016-01-25 21:13:09,487 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Sent 1 blockreports 0 blocks total. Took 1 msec to generate and 42 msecs for RPC and NN processing.  Got back commands none
> 2016-01-25 21:13:09,492 INFO org.apache.hadoop.util.GSet: Computing capacity for map BlockMap
> 2016-01-25 21:13:09,493 INFO org.apache.hadoop.util.GSet: VM type       = 64-bit
> 2016-01-25 21:13:09,494 INFO org.apache.hadoop.util.GSet: 0.5% max memory 1.8 GB = 9.1 MB
> 2016-01-25 21:13:09,494 INFO org.apache.hadoop.util.GSet: capacity      = 2^20 = 1048576 entries
> 2016-01-25 21:13:09,495 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Periodic Block Verification Scanner initialized with interval 504 hours for block pool BP-216406264-127.0.0.1-1453767164845
> 2016-01-25 21:13:09,499 INFO org.apache.hadoop.hdfs.server.datanode.DataBlockScanner: Added bpid=BP-216406264-127.0.0.1-1453767164845 to blockPoolScannerMap, new size=1
> 2016-01-25 21:13:32,355 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-216406264-127.0.0.1-1453767164845:blk_1073741825_1001 src: /192.168.0.10:58649 dest: /192.168.0.10:50010
> 2016-01-25 21:13:32,482 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:58649, dest: /192.168.0.10:50010, bytes: 343691, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_538002429_1, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741825_1001, duration: 98632367
> 2016-01-25 21:13:32,482 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-216406264-127.0.0.1-1453767164845:blk_1073741825_1001, type=LAST_IN_PIPELINE, downstreams=0:[] terminating
> 2016-01-25 21:13:34,291 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification succeeded for BP-216406264-127.0.0.1-1453767164845:blk_1073741825_1001
> 2016-01-25 21:14:10,176 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-216406264-127.0.0.1-1453767164845:blk_1073741826_1002 src: /192.168.0.10:58663 dest: /192.168.0.10:50010
> 2016-01-25 21:14:10,220 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:58663, dest: /192.168.0.10:50010, bytes: 270263, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_342504113_1, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741826_1002, duration: 42378742
> 2016-01-25 21:14:10,221 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-216406264-127.0.0.1-1453767164845:blk_1073741826_1002, type=LAST_IN_PIPELINE, downstreams=0:[] terminating
> 2016-01-25 21:14:10,714 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-216406264-127.0.0.1-1453767164845:blk_1073741827_1003 src: /192.168.0.10:58664 dest: /192.168.0.10:50010
> 2016-01-25 21:14:10,721 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:58664, dest: /192.168.0.10:50010, bytes: 121, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_342504113_1, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741827_1003, duration: 2656758
> 2016-01-25 21:14:10,721 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-216406264-127.0.0.1-1453767164845:blk_1073741827_1003, type=LAST_IN_PIPELINE, downstreams=0:[] terminating
> 2016-01-25 21:14:10,853 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-216406264-127.0.0.1-1453767164845:blk_1073741828_1004 src: /192.168.0.10:58665 dest: /192.168.0.10:50010
> 2016-01-25 21:14:10,860 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:58665, dest: /192.168.0.10:50010, bytes: 26, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_342504113_1, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741828_1004, duration: 3257396
> 2016-01-25 21:14:10,861 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-216406264-127.0.0.1-1453767164845:blk_1073741828_1004, type=LAST_IN_PIPELINE, downstreams=0:[] terminating
> 2016-01-25 21:14:11,717 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-216406264-127.0.0.1-1453767164845:blk_1073741829_1005 src: /192.168.0.10:58666 dest: /192.168.0.10:50010
> 2016-01-25 21:14:11,726 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:58666, dest: /192.168.0.10:50010, bytes: 77957, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_342504113_1, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741829_1005, duration: 6180229
> 2016-01-25 21:14:11,727 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-216406264-127.0.0.1-1453767164845:blk_1073741829_1005, type=LAST_IN_PIPELINE, downstreams=0:[] terminating
> 2016-01-25 21:14:14,298 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification succeeded for BP-216406264-127.0.0.1-1453767164845:blk_1073741829_1005
> 2016-01-25 21:14:14,299 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification succeeded for BP-216406264-127.0.0.1-1453767164845:blk_1073741827_1003
> 2016-01-25 21:14:14,305 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification succeeded for BP-216406264-127.0.0.1-1453767164845:blk_1073741826_1002
> 2016-01-25 21:14:14,305 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification succeeded for BP-216406264-127.0.0.1-1453767164845:blk_1073741828_1004
> 2016-01-25 21:14:16,099 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:50010, dest: /192.168.0.10:58674, bytes: 272375, op: HDFS_READ, cliID: DFSClient_NONMAPREDUCE_76231625_102, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741826_1002, duration: 2878920
> 2016-01-25 21:14:16,253 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:50010, dest: /192.168.0.10:58674, bytes: 30, op: HDFS_READ, cliID: DFSClient_NONMAPREDUCE_76231625_102, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741828_1004, duration: 236423
> 2016-01-25 21:14:16,312 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:50010, dest: /192.168.0.10:58674, bytes: 125, op: HDFS_READ, cliID: DFSClient_NONMAPREDUCE_76231625_102, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741827_1003, duration: 909236
> 2016-01-25 21:14:16,364 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:50010, dest: /192.168.0.10:58674, bytes: 78569, op: HDFS_READ, cliID: DFSClient_NONMAPREDUCE_76231625_102, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741829_1005, duration: 1489437
> 2016-01-25 21:14:20,174 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:50010, dest: /192.168.0.10:58676, bytes: 30, op: HDFS_READ, cliID: DFSClient_NONMAPREDUCE_694066886_1, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741828_1004, duration: 899980
> 2016-01-25 21:14:22,692 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-216406264-127.0.0.1-1453767164845:blk_1073741830_1006 src: /192.168.0.10:58679 dest: /192.168.0.10:50010
> 2016-01-25 21:14:22,754 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:58679, dest: /192.168.0.10:50010, bytes: 92684, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_694066886_1, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741830_1006, duration: 60114851
> 2016-01-25 21:14:22,754 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-216406264-127.0.0.1-1453767164845:blk_1073741830_1006, type=LAST_IN_PIPELINE, downstreams=0:[] terminating
> 2016-01-25 21:14:24,319 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification succeeded for BP-216406264-127.0.0.1-1453767164845:blk_1073741830_1006
> 2016-01-25 21:14:25,808 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-216406264-127.0.0.1-1453767164845:blk_1073741831_1007 src: /192.168.0.10:58681 dest: /192.168.0.10:50010
> 2016-01-25 21:14:35,846 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:58681, dest: /192.168.0.10:50010, bytes: 21176, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_694066886_1, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741831_1007, duration: 9975409048
> 2016-01-25 21:14:35,846 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-216406264-127.0.0.1-1453767164845:blk_1073741831_1007, type=LAST_IN_PIPELINE, downstreams=0:[] terminating
> 2016-01-25 21:14:36,066 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-216406264-127.0.0.1-1453767164845:blk_1073741832_1008 src: /192.168.0.10:58682 dest: /192.168.0.10:50010
> 2016-01-25 21:14:36,075 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:58682, dest: /192.168.0.10:50010, bytes: 332, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_694066886_1, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741832_1008, duration: 4992595
> 2016-01-25 21:14:36,075 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-216406264-127.0.0.1-1453767164845:blk_1073741832_1008, type=LAST_IN_PIPELINE, downstreams=0:[] terminating
> 2016-01-25 21:14:36,548 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:50010, dest: /192.168.0.10:58683, bytes: 21344, op: HDFS_READ, cliID: DFSClient_NONMAPREDUCE_694066886_1, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741831_1007, duration: 497225
> 2016-01-25 21:14:36,564 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-216406264-127.0.0.1-1453767164845:blk_1073741833_1009 src: /192.168.0.10:58684 dest: /192.168.0.10:50010
> 2016-01-25 21:14:36,572 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:58684, dest: /192.168.0.10:50010, bytes: 21176, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_694066886_1, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741833_1009, duration: 2649337
> 2016-01-25 21:14:36,573 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-216406264-127.0.0.1-1453767164845:blk_1073741833_1009, type=LAST_IN_PIPELINE, downstreams=0:[] terminating
> 2016-01-25 21:14:36,622 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:50010, dest: /192.168.0.10:58683, bytes: 93412, op: HDFS_READ, cliID: DFSClient_NONMAPREDUCE_694066886_1, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741830_1006, duration: 379439
> 2016-01-25 21:14:36,638 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-216406264-127.0.0.1-1453767164845:blk_1073741834_1010 src: /192.168.0.10:58685 dest: /192.168.0.10:50010
> 2016-01-25 21:14:36,646 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:58685, dest: /192.168.0.10:50010, bytes: 92684, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_694066886_1, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741834_1010, duration: 3135698
> 2016-01-25 21:14:36,646 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-216406264-127.0.0.1-1453767164845:blk_1073741834_1010, type=LAST_IN_PIPELINE, downstreams=0:[] terminating
> 2016-01-25 21:14:39,335 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification succeeded for BP-216406264-127.0.0.1-1453767164845:blk_1073741833_1009
> 2016-01-25 21:14:39,336 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification succeeded for BP-216406264-127.0.0.1-1453767164845:blk_1073741831_1007
> 2016-01-25 21:14:39,337 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification succeeded for BP-216406264-127.0.0.1-1453767164845:blk_1073741832_1008
> 2016-01-25 21:14:39,338 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification succeeded for BP-216406264-127.0.0.1-1453767164845:blk_1073741834_1010
> 2016-01-25 21:14:39,376 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService: Scheduling blk_1073741826_1002 file /usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741826 for deletion
> 2016-01-25 21:14:39,379 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService: Scheduling blk_1073741827_1003 file /usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741827 for deletion
> 2016-01-25 21:14:39,379 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService: Deleted BP-216406264-127.0.0.1-1453767164845 blk_1073741826_1002 file /usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741826
> 2016-01-25 21:14:39,379 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService: Scheduling blk_1073741828_1004 file /usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741828 for deletion
> 2016-01-25 21:14:39,380 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService: Deleted BP-216406264-127.0.0.1-1453767164845 blk_1073741827_1003 file /usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741827
> 2016-01-25 21:14:39,380 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService: Scheduling blk_1073741829_1005 file /usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741829 for deletion
> 2016-01-25 21:14:39,380 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService: Deleted BP-216406264-127.0.0.1-1453767164845 blk_1073741828_1004 file /usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741828
> 2016-01-25 21:14:39,380 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService: Scheduling blk_1073741830_1006 file /usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741830 for deletion
> 2016-01-25 21:14:39,381 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService: Deleted BP-216406264-127.0.0.1-1453767164845 blk_1073741829_1005 file /usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741829
> 2016-01-25 21:14:39,381 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService: Scheduling blk_1073741831_1007 file /usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741831 for deletion
> 2016-01-25 21:14:39,382 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService: Deleted BP-216406264-127.0.0.1-1453767164845 blk_1073741830_1006 file /usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741830
> 2016-01-25 21:14:39,382 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService: Deleted BP-216406264-127.0.0.1-1453767164845 blk_1073741831_1007 file /usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741831
> 2016-01-25 21:14:44,797 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-216406264-127.0.0.1-1453767164845:blk_1073741835_1011 src: /192.168.0.10:58688 dest: /192.168.0.10:50010
> 2016-01-25 21:14:44,834 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:58688, dest: /192.168.0.10:50010, bytes: 57450, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_-433405715_88, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741835_1011, duration: 34522284
> 2016-01-25 21:14:44,834 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-216406264-127.0.0.1-1453767164845:blk_1073741835_1011, type=LAST_IN_PIPELINE, downstreams=0:[] terminating
> 2016-01-25 21:14:49,343 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification succeeded for BP-216406264-127.0.0.1-1453767164845:blk_1073741835_1011
> 2016-01-25 21:16:33,785 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /192.168.0.10:50010, dest: /192.168.0.10:58694, bytes: 336, op: HDFS_READ, cliID: DFSClient_NONMAPREDUCE_-1832227986_1, offset: 0, srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid: BP-216406264-127.0.0.1-1453767164845:blk_1073741832_1008, duration: 284719
> 2016-01-25 21:16:36,371 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService: Scheduling blk_1073741832_1008 file /usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741832 for deletion
> 2016-01-25 21:16:36,372 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService: Deleted BP-216406264-127.0.0.1-1453767164845 blk_1073741832_1008 file /usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741832
>
>
>
>
> 2016-01-21 18:52 GMT-03:00 Namikaze Minato <lloydsensei@gmail.com>:
>
>> It could be a classpath issue (see
>> http://stackoverflow.com/a/25090151/4486184), I'm strongly thinking
>> this is the case.
>>
>> You could drill down to the exact root cause by looking at the
>> datanode logs (see
>>
>> http://mail-archives.apache.org/mod_mbox/hadoop-user/201410.mbox/%3CCAEMetGubzq12LXbLRk6N4ejOoKse9dLEWMW8_WE6aRj=+RQtVw@mail.gmail.com%3E
>> )
>> But I'm not sure we would get another error than what we had...
>>
>> Check if your application has the correct values for the following
>> variables:
>> HADOOP_CONF_DIR
>> HADOOP_COMMON_HOME
>> HADOOP_HDFS_HOME
>> HADOOP_MAPRED_HOME
>> HADOOP_YARN_HOME
>>
>> I'm afraid I can't help you much more than this myself, sorry...
>>
>> LLoyd
>>
>> On 20 January 2016 at 02:08, José Luis Larroque <larroquester@gmail.com>
>> wrote:
>> > Hi guys, thanks for your answers.
>> >
>> > Wordcount logs:
>> >
>> > 16/01/19 21:58:32 INFO client.RMProxy: Connecting to ResourceManager at
>> > hdnode01/192.168.0.10:8050
>> > SLF4J: Class path contains multiple SLF4J bindings.
>> > SLF4J: Found binding in
>> >
>> [jar:file:/usr/local/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> > SLF4J: Found binding in
>> >
>> [jar:file:/usr/local/hadoop/share/hadoop/yarn/lib/giraph-examples-1.1.0-for-hadoop-2.4.0-jar-with-dependencies.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> > SLF4J: Found binding in
>> >
>> [jar:file:/usr/local/hadoop/share/hadoop/mapreduce/giraph-examples-1.1.0-for-hadoop-2.4.0-jar-with-dependencies.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>> > explanation.
>> > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
>> > 16/01/19 21:58:32 WARN util.NativeCodeLoader: Unable to load
>> native-hadoop
>> > library for your platform... using builtin-java classes where applicable
>> > hduser@jose-ubuntu:/usr/local/hadoop$ nano
>> >
>> /home/hduser/Desktop/Tesina/casos_de_prueba/resultados/resultado_cluster_modo_yarn_wordcount
>> > hduser@jose-ubuntu:/usr/local/hadoop$ nano
>> >
>> /home/hduser/Desktop/Tesina/casos_de_prueba/resultados/resultado_cluster_modo_yarn_wordcount
>> > hduser@jose-ubuntu:/usr/local/hadoop$ cat
>> >
>> /home/hduser/Desktop/Tesina/casos_de_prueba/resultados/resultado_cluster_modo_yarn_wordcount
>> >
>> >
>> > Container: container_1453244277886_0001_01_000002 on localhost_35711
>> > ======================================================================
>> > LogType: stderr
>> > LogLength: 45
>> > Log Contents:
>> > Error: Could not find or load main class 256
>> >
>> > LogType: stdout
>> > LogLength: 0
>> > Log Contents:
>> >
>> >
>> >
>> > Container: container_1453244277886_0001_01_000003 on localhost_35711
>> > ======================================================================
>> > LogType: stderr
>> > LogLength: 45
>> > Log Contents:
>> > Error: Could not find or load main class 256
>> >
>> > LogType: stdout
>> > LogLength: 0
>> > Log Contents:
>> >
>> >
>> >
>> > Container: container_1453244277886_0001_01_000004 on localhost_35711
>> > ======================================================================
>> > LogType: stderr
>> > LogLength: 45
>> > Log Contents:
>> > Error: Could not find or load main class 256
>> >
>> > LogType: stdout
>> > LogLength: 0
>> > Log Contents:
>> >
>> >
>> >
>> > Container: container_1453244277886_0001_01_000005 on localhost_35711
>> > ======================================================================
>> > LogType: stderr
>> > LogLength: 45
>> > Log Contents:
>> > Error: Could not find or load main class 256
>> >
>> > LogType: stdout
>> > LogLength: 0
>> > Log Contents:
>> >
>> >
>> >
>> > Container: container_1453244277886_0001_01_000001 on localhost_35711
>> > ======================================================================
>> > LogType: stderr
>> > LogLength: 929
>> > Log Contents:
>> > SLF4J: Class path contains multiple SLF4J bindings.
>> > SLF4J: Found binding in
>> >
>> [jar:file:/usr/local/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> > SLF4J: Found binding in
>> >
>> [jar:file:/usr/local/hadoop/share/hadoop/yarn/lib/giraph-examples-1.1.0-for-hadoop-2.4.0-jar-with-dependencies.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> > SLF4J: Found binding in
>> >
>> [jar:file:/usr/local/hadoop/share/hadoop/mapreduce/giraph-examples-1.1.0-for-hadoop-2.4.0-jar-with-dependencies.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>> > explanation.
>> > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
>> > log4j:WARN No appenders could be found for logger
>> > (org.apache.hadoop.ipc.Server).
>> > log4j:WARN Please initialize the log4j system properly.
>> > log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig
>> for
>> > more info.
>> >
>> > LogType: stdout
>> > LogLength: 0
>> > Log Contents:
>> >
>> > LogType: syslog
>> > LogLength: 56780
>> > Log Contents:
>> > 2016-01-19 20:04:11,329 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Created MRAppMaster for
>> > application appattempt_1453244277886_0001_000001
>> > 2016-01-19 20:04:11,657 WARN [main]
>> org.apache.hadoop.conf.Configuration:
>> > job.xml:an attempt to override final parameter:
>> > mapreduce.job.end-notification.max.retry.interval;  Ignoring.
>> > 2016-01-19 20:04:11,674 WARN [main]
>> org.apache.hadoop.conf.Configuration:
>> > job.xml:an attempt to override final parameter:
>> > mapreduce.job.end-notification.max.attempts;  Ignoring.
>> > 2016-01-19 20:04:11,765 WARN [main]
>> org.apache.hadoop.util.NativeCodeLoader:
>> > Unable to load native-hadoop library for your platform... using
>> builtin-java
>> > classes where applicable
>> > 2016-01-19 20:04:11,776 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Executing with tokens:
>> > 2016-01-19 20:04:11,776 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Kind: YARN_AM_RM_TOKEN,
>> > Service: , Ident:
>> > (org.apache.hadoop.yarn.security.AMRMTokenIdentifier@73e8f4b9)
>> > 2016-01-19 20:04:11,801 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: The specific max
>> attempts: 2
>> > for application: 1. Attempt num: 1 is last retry: false
>> > 2016-01-19 20:04:11,806 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Using mapred
>> > newApiCommitter.
>> > 2016-01-19 20:04:11,934 WARN [main]
>> org.apache.hadoop.conf.Configuration:
>> > job.xml:an attempt to override final parameter: dfs.datanode.data.dir;
>> > Ignoring.
>> > 2016-01-19 20:04:11,939 WARN [main]
>> org.apache.hadoop.conf.Configuration:
>> > job.xml:an attempt to override final parameter:
>> > mapreduce.job.end-notification.max.retry.interval;  Ignoring.
>> > 2016-01-19 20:04:11,948 WARN [main]
>> org.apache.hadoop.conf.Configuration:
>> > job.xml:an attempt to override final parameter: dfs.namenode.name.dir;
>> > Ignoring.
>> > 2016-01-19 20:04:11,953 WARN [main]
>> org.apache.hadoop.conf.Configuration:
>> > job.xml:an attempt to override final parameter:
>> > mapreduce.job.end-notification.max.attempts;  Ignoring.
>> > 2016-01-19 20:04:12,464 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter set in
>> > config null
>> > 2016-01-19 20:04:12,526 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter is
>> > org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
>> > 2016-01-19 20:04:12,548 INFO [main]
>> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
>> > org.apache.hadoop.mapreduce.jobhistory.EventType for class
>> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler
>> > 2016-01-19 20:04:12,549 INFO [main]
>> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
>> > org.apache.hadoop.mapreduce.v2.app.job.event.JobEventType for class
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobEventDispatcher
>> > 2016-01-19 20:04:12,550 INFO [main]
>> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
>> > org.apache.hadoop.mapreduce.v2.app.job.event.TaskEventType for class
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskEventDispatcher
>> > 2016-01-19 20:04:12,551 INFO [main]
>> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
>> > org.apache.hadoop.mapreduce.v2.app.job.event.TaskAttemptEventType for
>> class
>> >
>> org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskAttemptEventDispatcher
>> > 2016-01-19 20:04:12,552 INFO [main]
>> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
>> > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventType for class
>> > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler
>> > 2016-01-19 20:04:12,557 INFO [main]
>> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
>> > org.apache.hadoop.mapreduce.v2.app.speculate.Speculator$EventType for
>> class
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster$SpeculatorEventDispatcher
>> > 2016-01-19 20:04:12,558 INFO [main]
>> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
>> > org.apache.hadoop.mapreduce.v2.app.rm.ContainerAllocator$EventType for
>> class
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerAllocatorRouter
>> > 2016-01-19 20:04:12,559 INFO [main]
>> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncher$EventType
>> for
>> > class
>> org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerLauncherRouter
>> > 2016-01-19 20:04:12,615 INFO [main]
>> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Perms
>> after
>> > creating 488, Expected: 504
>> > 2016-01-19 20:04:12,615 INFO [main]
>> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler:
>> Explicitly
>> > setting permissions to : 504, rwxrwx---
>> > 2016-01-19 20:04:12,731 INFO [main]
>> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
>> > org.apache.hadoop.mapreduce.v2.app.job.event.JobFinishEvent$Type for
>> class
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobFinishEventHandler
>> > 2016-01-19 20:04:12,956 INFO [main]
>> > org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from
>> > hadoop-metrics2.properties
>> > 2016-01-19 20:04:13,018 INFO [main]
>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
>> period
>> > at 10 second(s).
>> > 2016-01-19 20:04:13,018 INFO [main]
>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: MRAppMaster metrics
>> > system started
>> > 2016-01-19 20:04:13,026 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Adding job token
>> for
>> > job_1453244277886_0001 to jobTokenSecretManager
>> > 2016-01-19 20:04:13,139 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Not uberizing
>> > job_1453244277886_0001 because: not enabled;
>> > 2016-01-19 20:04:13,154 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Input size for job
>> > job_1453244277886_0001 = 343691. Number of splits = 1
>> > 2016-01-19 20:04:13,156 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Number of reduces
>> for
>> > job job_1453244277886_0001 = 1
>> > 2016-01-19 20:04:13,156 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl:
>> > job_1453244277886_0001Job Transitioned from NEW to INITED
>> > 2016-01-19 20:04:13,157 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: MRAppMaster launching
>> > normal, non-uberized, multi-container job job_1453244277886_0001.
>> > 2016-01-19 20:04:13,186 INFO [main]
>> org.apache.hadoop.ipc.CallQueueManager:
>> > Using callQueue class java.util.concurrent.LinkedBlockingQueue
>> > 2016-01-19 20:04:13,195 INFO [Socket Reader #1 for port 56461]
>> > org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 56461
>> > 2016-01-19 20:04:13,237 INFO [main]
>> > org.apache.hadoop.yarn.factories.impl.pb.RpcServerFactoryPBImpl: Adding
>> > protocol org.apache.hadoop.mapreduce.v2.api.MRClientProtocolPB to the
>> server
>> > 2016-01-19 20:04:13,238 INFO [IPC Server Responder]
>> > org.apache.hadoop.ipc.Server: IPC Server Responder: starting
>> > 2016-01-19 20:04:13,238 INFO [IPC Server listener on 56461]
>> > org.apache.hadoop.ipc.Server: IPC Server listener on 56461: starting
>> > 2016-01-19 20:04:13,239 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.client.MRClientService: Instantiated
>> > MRClientService at jose-ubuntu/127.0.0.1:56461
>> > 2016-01-19 20:04:13,300 INFO [main] org.mortbay.log: Logging to
>> > org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
>> > org.mortbay.log.Slf4jLog
>> > 2016-01-19 20:04:13,304 INFO [main]
>> org.apache.hadoop.http.HttpRequestLog:
>> > Http request log for http.requests.mapreduce is not defined
>> > 2016-01-19 20:04:13,315 INFO [main] org.apache.hadoop.http.HttpServer2:
>> > Added global filter 'safety'
>> > (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)
>> > 2016-01-19 20:04:13,320 INFO [main] org.apache.hadoop.http.HttpServer2:
>> > Added filter AM_PROXY_FILTER
>> > (class=org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter) to
>> > context mapreduce
>> > 2016-01-19 20:04:13,321 INFO [main] org.apache.hadoop.http.HttpServer2:
>> > Added filter AM_PROXY_FILTER
>> > (class=org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter) to
>> > context static
>> > 2016-01-19 20:04:13,324 INFO [main] org.apache.hadoop.http.HttpServer2:
>> > adding path spec: /mapreduce/*
>> > 2016-01-19 20:04:13,324 INFO [main] org.apache.hadoop.http.HttpServer2:
>> > adding path spec: /ws/*
>> > 2016-01-19 20:04:13,335 INFO [main] org.apache.hadoop.http.HttpServer2:
>> > Jetty bound to port 44070
>> > 2016-01-19 20:04:13,335 INFO [main] org.mortbay.log: jetty-6.1.26
>> > 2016-01-19 20:04:13,370 INFO [main] org.mortbay.log: Extract
>> >
>> jar:file:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-common-2.4.0.jar!/webapps/mapreduce
>> > to /tmp/Jetty_0_0_0_0_44070_mapreduce____rdpvio/webapp
>> > 2016-01-19 20:04:13,647 INFO [main] org.mortbay.log: Started
>> > SelectChannelConnector@0.0.0.0:44070
>> > 2016-01-19 20:04:13,647 INFO [main]
>> org.apache.hadoop.yarn.webapp.WebApps:
>> > Web app /mapreduce started at 44070
>> > 2016-01-19 20:04:13,956 INFO [main]
>> org.apache.hadoop.yarn.webapp.WebApps:
>> > Registered webapp guice modules
>> > 2016-01-19 20:04:13,960 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.speculate.DefaultSpeculator:
>> JOB_CREATE
>> > job_1453244277886_0001
>> > 2016-01-19 20:04:13,961 INFO [main]
>> org.apache.hadoop.ipc.CallQueueManager:
>> > Using callQueue class java.util.concurrent.LinkedBlockingQueue
>> > 2016-01-19 20:04:13,961 INFO [Socket Reader #1 for port 45584]
>> > org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 45584
>> > 2016-01-19 20:04:13,966 INFO [IPC Server Responder]
>> > org.apache.hadoop.ipc.Server: IPC Server Responder: starting
>> > 2016-01-19 20:04:13,966 INFO [IPC Server listener on 45584]
>> > org.apache.hadoop.ipc.Server: IPC Server listener on 45584: starting
>> > 2016-01-19 20:04:13,987 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
>> > nodeBlacklistingEnabled:true
>> > 2016-01-19 20:04:13,987 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
>> > maxTaskFailuresPerNode is 3
>> > 2016-01-19 20:04:13,988 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
>> > blacklistDisablePercent is 33
>> > 2016-01-19 20:04:14,052 WARN [main]
>> org.apache.hadoop.conf.Configuration:
>> > job.xml:an attempt to override final parameter: dfs.datanode.data.dir;
>> > Ignoring.
>> > 2016-01-19 20:04:14,054 WARN [main]
>> org.apache.hadoop.conf.Configuration:
>> > job.xml:an attempt to override final parameter:
>> > mapreduce.job.end-notification.max.retry.interval;  Ignoring.
>> > 2016-01-19 20:04:14,057 WARN [main]
>> org.apache.hadoop.conf.Configuration:
>> > job.xml:an attempt to override final parameter: dfs.namenode.name.dir;
>> > Ignoring.
>> > 2016-01-19 20:04:14,059 WARN [main]
>> org.apache.hadoop.conf.Configuration:
>> > job.xml:an attempt to override final parameter:
>> > mapreduce.job.end-notification.max.attempts;  Ignoring.
>> > 2016-01-19 20:04:14,062 INFO [main]
>> org.apache.hadoop.yarn.client.RMProxy:
>> > Connecting to ResourceManager at hdnode01/192.168.0.10:8030
>> > 2016-01-19 20:04:14,158 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> > maxContainerCapability: 2000
>> > 2016-01-19 20:04:14,158 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: queue:
>> default
>> > 2016-01-19 20:04:14,162 INFO [main]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Upper
>> > limit on the thread pool size is 500
>> > 2016-01-19 20:04:14,164 INFO [main]
>> > org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy:
>> > yarn.client.max-nodemanagers-proxies : 500
>> > 2016-01-19 20:04:14,172 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl:
>> > job_1453244277886_0001Job Transitioned from INITED to SETUP
>> > 2016-01-19 20:04:14,174 INFO [CommitterEvent Processor #0]
>> > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler:
>> Processing
>> > the event EventType: JOB_SETUP
>> > 2016-01-19 20:04:14,210 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl:
>> > job_1453244277886_0001Job Transitioned from SETUP to RUNNING
>> > 2016-01-19 20:04:14,227 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.yarn.util.RackResolver: Resolved localhost to
>> > /default-rack
>> > 2016-01-19 20:04:14,230 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl:
>> > task_1453244277886_0001_m_000000 Task Transitioned from NEW to SCHEDULED
>> > 2016-01-19 20:04:14,230 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl:
>> > task_1453244277886_0001_r_000000 Task Transitioned from NEW to SCHEDULED
>> > 2016-01-19 20:04:14,232 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_0 TaskAttempt Transitioned from NEW
>> to
>> > UNASSIGNED
>> > 2016-01-19 20:04:14,232 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_r_000000_0 TaskAttempt Transitioned from NEW
>> to
>> > UNASSIGNED
>> > 2016-01-19 20:04:14,233 INFO [Thread-50]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> > mapResourceReqt:512
>> > 2016-01-19 20:04:14,245 INFO [Thread-50]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> > reduceResourceReqt:512
>> > 2016-01-19 20:04:14,324 INFO [eventHandlingThread]
>> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Event
>> Writer
>> > setup for JobId: job_1453244277886_0001, File:
>> >
>> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/hduser/.staging/job_1453244277886_0001/job_1453244277886_0001_1.jhist
>> > 2016-01-19 20:04:15,162 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before
>> > Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:0
>> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:0 ContRel:0
>> > HostLocal:0 RackLocal:0
>> > 2016-01-19 20:04:15,217 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
>> getResources()
>> > for application_1453244277886_0001: ask=3 release= 0 newContainers=0
>> > finishedContainers=0 resourcelimit=<memory:1280, vCores:0> knownNMs=1
>> > 2016-01-19 20:04:15,217 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> Recalculating
>> > schedule, headroom=1280
>> > 2016-01-19 20:04:15,217 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
>> > start threshold not met. completedMapsForReduceSlowstart 1
>> > 2016-01-19 20:04:16,240 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got
>> allocated
>> > containers 1
>> > 2016-01-19 20:04:16,241 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned
>> > container container_1453244277886_0001_01_000002 to
>> > attempt_1453244277886_0001_m_000000_0
>> > 2016-01-19 20:04:16,243 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> Recalculating
>> > schedule, headroom=768
>> > 2016-01-19 20:04:16,243 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
>> > start threshold not met. completedMapsForReduceSlowstart 1
>> > 2016-01-19 20:04:16,243 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
>> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1
>> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:1 ContRel:0
>> > HostLocal:1 RackLocal:0
>> > 2016-01-19 20:04:16,291 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.yarn.util.RackResolver: Resolved localhost to
>> > /default-rack
>> > 2016-01-19 20:04:16,316 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: The job-jar
>> > file on the remote FS is
>> >
>> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/hduser/.staging/job_1453244277886_0001/job.jar
>> > 2016-01-19 20:04:16,322 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: The
>> job-conf
>> > file on the remote FS is
>> > /tmp/hadoop-yarn/staging/hduser/.staging/job_1453244277886_0001/job.xml
>> > 2016-01-19 20:04:16,325 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Adding #0
>> > tokens and #1 secret keys for NM use for launching container
>> > 2016-01-19 20:04:16,325 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Size of
>> > containertokens_dob is 1
>> > 2016-01-19 20:04:16,325 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Putting
>> shuffle
>> > token in serviceData
>> > 2016-01-19 20:04:16,350 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_0 TaskAttempt Transitioned from
>> > UNASSIGNED to ASSIGNED
>> > 2016-01-19 20:04:16,354 INFO [ContainerLauncher #0]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> > Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container
>> > container_1453244277886_0001_01_000002 taskAttempt
>> > attempt_1453244277886_0001_m_000000_0
>> > 2016-01-19 20:04:16,356 INFO [ContainerLauncher #0]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> Launching
>> > attempt_1453244277886_0001_m_000000_0
>> > 2016-01-19 20:04:16,357 INFO [ContainerLauncher #0]
>> > org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy:
>> > Opening proxy : localhost:35711
>> > 2016-01-19 20:04:16,411 INFO [ContainerLauncher #0]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> Shuffle
>> > port returned by ContainerManager for
>> attempt_1453244277886_0001_m_000000_0
>> > : 13562
>> > 2016-01-19 20:04:16,413 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> TaskAttempt:
>> > [attempt_1453244277886_0001_m_000000_0] using containerId:
>> > [container_1453244277886_0001_01_000002 on NM: [localhost:35711]
>> > 2016-01-19 20:04:16,418 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_0 TaskAttempt Transitioned from
>> ASSIGNED
>> > to RUNNING
>> > 2016-01-19 20:04:16,419 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.speculate.DefaultSpeculator:
>> > ATTEMPT_START task_1453244277886_0001_m_000000
>> > 2016-01-19 20:04:16,419 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl:
>> > task_1453244277886_0001_m_000000 Task Transitioned from SCHEDULED to
>> RUNNING
>> > 2016-01-19 20:04:17,251 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
>> getResources()
>> > for application_1453244277886_0001: ask=3 release= 0 newContainers=0
>> > finishedContainers=0 resourcelimit=<memory:768, vCores:-1> knownNMs=1
>> > 2016-01-19 20:04:18,269 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received
>> > completed container container_1453244277886_0001_01_000002
>> > 2016-01-19 20:04:18,269 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> Recalculating
>> > schedule, headroom=768
>> > 2016-01-19 20:04:18,269 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
>> > start threshold not met. completedMapsForReduceSlowstart 1
>> > 2016-01-19 20:04:18,270 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
>> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0
>> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:1 ContRel:0
>> > HostLocal:1 RackLocal:0
>> > 2016-01-19 20:04:18,280 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_0 TaskAttempt Transitioned from
>> RUNNING
>> > to FAIL_CONTAINER_CLEANUP
>> > 2016-01-19 20:04:18,280 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics
>> > report from attempt_1453244277886_0001_m_000000_0: Exception from
>> > container-launch: org.apache.hadoop.util.Shell$ExitCodeException:
>> > org.apache.hadoop.util.Shell$ExitCodeException:
>> >     at org.apache.hadoop.util.Shell.runCommand(Shell.java:505)
>> >     at org.apache.hadoop.util.Shell.run(Shell.java:418)
>> >     at
>> >
>> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:650)
>> >     at
>> >
>> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195)
>> >     at
>> >
>> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:300)
>> >     at
>> >
>> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:81)
>> >     at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>> >     at
>> >
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>> >     at
>> >
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>> >     at java.lang.Thread.run(Thread.java:745)
>> >
>> >
>> > Container exited with a non-zero exit code 1
>> >
>> > 2016-01-19 20:04:18,281 INFO [ContainerLauncher #1]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> > Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container
>> > container_1453244277886_0001_01_000002 taskAttempt
>> > attempt_1453244277886_0001_m_000000_0
>> > 2016-01-19 20:04:18,281 INFO [ContainerLauncher #1]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> KILLING
>> > attempt_1453244277886_0001_m_000000_0
>> > 2016-01-19 20:04:18,299 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_0 TaskAttempt Transitioned from
>> > FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
>> > 2016-01-19 20:04:18,300 INFO [CommitterEvent Processor #1]
>> > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler:
>> Processing
>> > the event EventType: TASK_ABORT
>> > 2016-01-19 20:04:18,313 WARN [CommitterEvent Processor #1]
>> > org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not
>> delete
>> >
>> hdfs://hdnode01:54310/user/hduser/output/wordcount/_temporary/1/_temporary/attempt_1453244277886_0001_m_000000_0
>> > 2016-01-19 20:04:18,317 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_0 TaskAttempt Transitioned from
>> > FAIL_TASK_CLEANUP to FAILED
>> > 2016-01-19 20:04:18,326 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.yarn.util.RackResolver: Resolved localhost to
>> > /default-rack
>> > 2016-01-19 20:04:18,327 INFO [Thread-50]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 1 failures
>> on
>> > node localhost
>> > 2016-01-19 20:04:18,329 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_1 TaskAttempt Transitioned from NEW
>> to
>> > UNASSIGNED
>> > 2016-01-19 20:04:18,329 INFO [Thread-50]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added
>> > attempt_1453244277886_0001_m_000000_1 to list of failed maps
>> > 2016-01-19 20:04:19,270 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before
>> > Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:0
>> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:1 ContRel:0
>> > HostLocal:1 RackLocal:0
>> > 2016-01-19 20:04:19,277 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
>> getResources()
>> > for application_1453244277886_0001: ask=1 release= 0 newContainers=0
>> > finishedContainers=0 resourcelimit=<memory:768, vCores:-1> knownNMs=1
>> > 2016-01-19 20:04:19,277 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> Recalculating
>> > schedule, headroom=768
>> > 2016-01-19 20:04:19,278 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
>> > start threshold not met. completedMapsForReduceSlowstart 1
>> > 2016-01-19 20:04:20,285 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got
>> allocated
>> > containers 1
>> > 2016-01-19 20:04:20,285 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning
>> > container Container: [ContainerId:
>> container_1453244277886_0001_01_000003,
>> > NodeId: localhost:35711, NodeHttpAddress: localhost:8042, Resource:
>> > <memory:512, vCores:1>, Priority: 5, Token: Token { kind:
>> ContainerToken,
>> > service: 127.0.0.1:35711 }, ] to fast fail map
>> > 2016-01-19 20:04:20,286 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned
>> from
>> > earlierFailedMaps
>> > 2016-01-19 20:04:20,286 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned
>> > container container_1453244277886_0001_01_000003 to
>> > attempt_1453244277886_0001_m_000000_1
>> > 2016-01-19 20:04:20,287 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> Recalculating
>> > schedule, headroom=768
>> > 2016-01-19 20:04:20,287 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
>> > start threshold not met. completedMapsForReduceSlowstart 1
>> > 2016-01-19 20:04:20,287 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
>> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1
>> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:2 ContRel:0
>> > HostLocal:1 RackLocal:0
>> > 2016-01-19 20:04:20,287 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.yarn.util.RackResolver: Resolved localhost to
>> > /default-rack
>> > 2016-01-19 20:04:20,289 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_1 TaskAttempt Transitioned from
>> > UNASSIGNED to ASSIGNED
>> > 2016-01-19 20:04:20,292 INFO [ContainerLauncher #2]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> > Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container
>> > container_1453244277886_0001_01_000003 taskAttempt
>> > attempt_1453244277886_0001_m_000000_1
>> > 2016-01-19 20:04:20,292 INFO [ContainerLauncher #2]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> Launching
>> > attempt_1453244277886_0001_m_000000_1
>> > 2016-01-19 20:04:20,301 INFO [ContainerLauncher #2]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> Shuffle
>> > port returned by ContainerManager for
>> attempt_1453244277886_0001_m_000000_1
>> > : 13562
>> > 2016-01-19 20:04:20,302 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> TaskAttempt:
>> > [attempt_1453244277886_0001_m_000000_1] using containerId:
>> > [container_1453244277886_0001_01_000003 on NM: [localhost:35711]
>> > 2016-01-19 20:04:20,303 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_1 TaskAttempt Transitioned from
>> ASSIGNED
>> > to RUNNING
>> > 2016-01-19 20:04:20,304 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.speculate.DefaultSpeculator:
>> > ATTEMPT_START task_1453244277886_0001_m_000000
>> > 2016-01-19 20:04:21,295 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
>> getResources()
>> > for application_1453244277886_0001: ask=1 release= 0 newContainers=0
>> > finishedContainers=1 resourcelimit=<memory:768, vCores:-1> knownNMs=1
>> > 2016-01-19 20:04:21,296 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received
>> > completed container container_1453244277886_0001_01_000003
>> > 2016-01-19 20:04:21,296 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> Recalculating
>> > schedule, headroom=768
>> > 2016-01-19 20:04:21,296 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
>> > start threshold not met. completedMapsForReduceSlowstart 1
>> > 2016-01-19 20:04:21,297 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
>> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0
>> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:2 ContRel:0
>> > HostLocal:1 RackLocal:0
>> > 2016-01-19 20:04:21,297 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_1 TaskAttempt Transitioned from
>> RUNNING
>> > to FAIL_CONTAINER_CLEANUP
>> > 2016-01-19 20:04:21,298 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics
>> > report from attempt_1453244277886_0001_m_000000_1: Exception from
>> > container-launch: org.apache.hadoop.util.Shell$ExitCodeException:
>> > org.apache.hadoop.util.Shell$ExitCodeException:
>> >     at org.apache.hadoop.util.Shell.runCommand(Shell.java:505)
>> >     at org.apache.hadoop.util.Shell.run(Shell.java:418)
>> >     at
>> >
>> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:650)
>> >     at
>> >
>> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195)
>> >     at
>> >
>> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:300)
>> >     at
>> >
>> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:81)
>> >     at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>> >     at
>> >
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>> >     at
>> >
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>> >     at java.lang.Thread.run(Thread.java:745)
>> >
>> >
>> > Container exited with a non-zero exit code 1
>> >
>> > 2016-01-19 20:04:21,300 INFO [ContainerLauncher #3]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> > Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container
>> > container_1453244277886_0001_01_000003 taskAttempt
>> > attempt_1453244277886_0001_m_000000_1
>> > 2016-01-19 20:04:21,301 INFO [ContainerLauncher #3]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> KILLING
>> > attempt_1453244277886_0001_m_000000_1
>> > 2016-01-19 20:04:21,307 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_1 TaskAttempt Transitioned from
>> > FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
>> > 2016-01-19 20:04:21,308 INFO [CommitterEvent Processor #2]
>> > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler:
>> Processing
>> > the event EventType: TASK_ABORT
>> > 2016-01-19 20:04:21,312 WARN [CommitterEvent Processor #2]
>> > org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not
>> delete
>> >
>> hdfs://hdnode01:54310/user/hduser/output/wordcount/_temporary/1/_temporary/attempt_1453244277886_0001_m_000000_1
>> > 2016-01-19 20:04:21,312 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_1 TaskAttempt Transitioned from
>> > FAIL_TASK_CLEANUP to FAILED
>> > 2016-01-19 20:04:21,313 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.yarn.util.RackResolver: Resolved localhost to
>> > /default-rack
>> > 2016-01-19 20:04:21,313 INFO [Thread-50]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 2 failures
>> on
>> > node localhost
>> > 2016-01-19 20:04:21,313 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_2 TaskAttempt Transitioned from NEW
>> to
>> > UNASSIGNED
>> > 2016-01-19 20:04:21,314 INFO [Thread-50]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added
>> > attempt_1453244277886_0001_m_000000_2 to list of failed maps
>> > 2016-01-19 20:04:22,297 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before
>> > Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:0
>> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:2 ContRel:0
>> > HostLocal:1 RackLocal:0
>> > 2016-01-19 20:04:22,304 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
>> getResources()
>> > for application_1453244277886_0001: ask=1 release= 0 newContainers=0
>> > finishedContainers=0 resourcelimit=<memory:768, vCores:-1> knownNMs=1
>> > 2016-01-19 20:04:22,305 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> Recalculating
>> > schedule, headroom=768
>> > 2016-01-19 20:04:22,305 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
>> > start threshold not met. completedMapsForReduceSlowstart 1
>> > 2016-01-19 20:04:23,316 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got
>> allocated
>> > containers 1
>> > 2016-01-19 20:04:23,316 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning
>> > container Container: [ContainerId:
>> container_1453244277886_0001_01_000004,
>> > NodeId: localhost:35711, NodeHttpAddress: localhost:8042, Resource:
>> > <memory:512, vCores:1>, Priority: 5, Token: Token { kind:
>> ContainerToken,
>> > service: 127.0.0.1:35711 }, ] to fast fail map
>> > 2016-01-19 20:04:23,317 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned
>> from
>> > earlierFailedMaps
>> > 2016-01-19 20:04:23,317 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned
>> > container container_1453244277886_0001_01_000004 to
>> > attempt_1453244277886_0001_m_000000_2
>> > 2016-01-19 20:04:23,318 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> Recalculating
>> > schedule, headroom=768
>> > 2016-01-19 20:04:23,318 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
>> > start threshold not met. completedMapsForReduceSlowstart 1
>> > 2016-01-19 20:04:23,318 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
>> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1
>> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:3 ContRel:0
>> > HostLocal:1 RackLocal:0
>> > 2016-01-19 20:04:23,318 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.yarn.util.RackResolver: Resolved localhost to
>> > /default-rack
>> > 2016-01-19 20:04:23,320 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_2 TaskAttempt Transitioned from
>> > UNASSIGNED to ASSIGNED
>> > 2016-01-19 20:04:23,323 INFO [ContainerLauncher #4]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> > Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container
>> > container_1453244277886_0001_01_000004 taskAttempt
>> > attempt_1453244277886_0001_m_000000_2
>> > 2016-01-19 20:04:23,323 INFO [ContainerLauncher #4]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> Launching
>> > attempt_1453244277886_0001_m_000000_2
>> > 2016-01-19 20:04:23,334 INFO [ContainerLauncher #4]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> Shuffle
>> > port returned by ContainerManager for
>> attempt_1453244277886_0001_m_000000_2
>> > : 13562
>> > 2016-01-19 20:04:23,335 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> TaskAttempt:
>> > [attempt_1453244277886_0001_m_000000_2] using containerId:
>> > [container_1453244277886_0001_01_000004 on NM: [localhost:35711]
>> > 2016-01-19 20:04:23,335 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_2 TaskAttempt Transitioned from
>> ASSIGNED
>> > to RUNNING
>> > 2016-01-19 20:04:23,335 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.speculate.DefaultSpeculator:
>> > ATTEMPT_START task_1453244277886_0001_m_000000
>> > 2016-01-19 20:04:24,326 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
>> getResources()
>> > for application_1453244277886_0001: ask=1 release= 0 newContainers=0
>> > finishedContainers=1 resourcelimit=<memory:768, vCores:-1> knownNMs=1
>> > 2016-01-19 20:04:24,327 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received
>> > completed container container_1453244277886_0001_01_000004
>> > 2016-01-19 20:04:24,327 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> Recalculating
>> > schedule, headroom=768
>> > 2016-01-19 20:04:24,327 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
>> > start threshold not met. completedMapsForReduceSlowstart 1
>> > 2016-01-19 20:04:24,328 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
>> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0
>> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:3 ContRel:0
>> > HostLocal:1 RackLocal:0
>> > 2016-01-19 20:04:24,328 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_2 TaskAttempt Transitioned from
>> RUNNING
>> > to FAIL_CONTAINER_CLEANUP
>> > 2016-01-19 20:04:24,328 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics
>> > report from attempt_1453244277886_0001_m_000000_2: Exception from
>> > container-launch: org.apache.hadoop.util.Shell$ExitCodeException:
>> > org.apache.hadoop.util.Shell$ExitCodeException:
>> >     at org.apache.hadoop.util.Shell.runCommand(Shell.java:505)
>> >     at org.apache.hadoop.util.Shell.run(Shell.java:418)
>> >     at
>> >
>> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:650)
>> >     at
>> >
>> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195)
>> >     at
>> >
>> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:300)
>> >     at
>> >
>> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:81)
>> >     at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>> >     at
>> >
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>> >     at
>> >
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>> >     at java.lang.Thread.run(Thread.java:745)
>> >
>> >
>> > Container exited with a non-zero exit code 1
>> >
>> > 2016-01-19 20:04:24,331 INFO [ContainerLauncher #5]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> > Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container
>> > container_1453244277886_0001_01_000004 taskAttempt
>> > attempt_1453244277886_0001_m_000000_2
>> > 2016-01-19 20:04:24,331 INFO [ContainerLauncher #5]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> KILLING
>> > attempt_1453244277886_0001_m_000000_2
>> > 2016-01-19 20:04:24,335 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_2 TaskAttempt Transitioned from
>> > FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
>> > 2016-01-19 20:04:24,336 INFO [CommitterEvent Processor #3]
>> > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler:
>> Processing
>> > the event EventType: TASK_ABORT
>> > 2016-01-19 20:04:24,340 WARN [CommitterEvent Processor #3]
>> > org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not
>> delete
>> >
>> hdfs://hdnode01:54310/user/hduser/output/wordcount/_temporary/1/_temporary/attempt_1453244277886_0001_m_000000_2
>> > 2016-01-19 20:04:24,341 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_2 TaskAttempt Transitioned from
>> > FAIL_TASK_CLEANUP to FAILED
>> > 2016-01-19 20:04:24,341 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.yarn.util.RackResolver: Resolved localhost to
>> > /default-rack
>> > 2016-01-19 20:04:24,342 INFO [Thread-50]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 3 failures
>> on
>> > node localhost
>> > 2016-01-19 20:04:24,342 INFO [Thread-50]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: Blacklisted
>> host
>> > localhost
>> > 2016-01-19 20:04:24,342 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_3 TaskAttempt Transitioned from NEW
>> to
>> > UNASSIGNED
>> > 2016-01-19 20:04:24,343 INFO [Thread-50]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added
>> > attempt_1453244277886_0001_m_000000_3 to list of failed maps
>> > 2016-01-19 20:04:25,328 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before
>> > Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:0
>> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:3 ContRel:0
>> > HostLocal:1 RackLocal:0
>> > 2016-01-19 20:04:25,336 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
>> getResources()
>> > for application_1453244277886_0001: ask=1 release= 0 newContainers=0
>> > finishedContainers=0 resourcelimit=<memory:768, vCores:-1> knownNMs=1
>> > 2016-01-19 20:04:25,336 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: Update the
>> > blacklist for application_1453244277886_0001: blacklistAdditions=1
>> > blacklistRemovals=0
>> > 2016-01-19 20:04:25,337 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: Ignore
>> > blacklisting set to true. Known: 1, Blacklisted: 1, 100%
>> > 2016-01-19 20:04:25,337 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> Recalculating
>> > schedule, headroom=768
>> > 2016-01-19 20:04:25,337 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
>> > start threshold not met. completedMapsForReduceSlowstart 1
>> > 2016-01-19 20:04:26,342 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: Update the
>> > blacklist for application_1453244277886_0001: blacklistAdditions=0
>> > blacklistRemovals=1
>> > 2016-01-19 20:04:26,342 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> Recalculating
>> > schedule, headroom=768
>> > 2016-01-19 20:04:26,342 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
>> > start threshold not met. completedMapsForReduceSlowstart 1
>> > 2016-01-19 20:04:27,351 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got
>> allocated
>> > containers 1
>> > 2016-01-19 20:04:27,352 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning
>> > container Container: [ContainerId:
>> container_1453244277886_0001_01_000005,
>> > NodeId: localhost:35711, NodeHttpAddress: localhost:8042, Resource:
>> > <memory:512, vCores:1>, Priority: 5, Token: Token { kind:
>> ContainerToken,
>> > service: 127.0.0.1:35711 }, ] to fast fail map
>> > 2016-01-19 20:04:27,352 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned
>> from
>> > earlierFailedMaps
>> > 2016-01-19 20:04:27,352 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned
>> > container container_1453244277886_0001_01_000005 to
>> > attempt_1453244277886_0001_m_000000_3
>> > 2016-01-19 20:04:27,353 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> Recalculating
>> > schedule, headroom=768
>> > 2016-01-19 20:04:27,353 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
>> > start threshold not met. completedMapsForReduceSlowstart 1
>> > 2016-01-19 20:04:27,353 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.yarn.util.RackResolver: Resolved localhost to
>> > /default-rack
>> > 2016-01-19 20:04:27,353 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
>> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1
>> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:4 ContRel:0
>> > HostLocal:1 RackLocal:0
>> > 2016-01-19 20:04:27,354 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_3 TaskAttempt Transitioned from
>> > UNASSIGNED to ASSIGNED
>> > 2016-01-19 20:04:27,355 INFO [ContainerLauncher #6]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> > Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container
>> > container_1453244277886_0001_01_000005 taskAttempt
>> > attempt_1453244277886_0001_m_000000_3
>> > 2016-01-19 20:04:27,355 INFO [ContainerLauncher #6]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> Launching
>> > attempt_1453244277886_0001_m_000000_3
>> > 2016-01-19 20:04:27,365 INFO [ContainerLauncher #6]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> Shuffle
>> > port returned by ContainerManager for
>> attempt_1453244277886_0001_m_000000_3
>> > : 13562
>> > 2016-01-19 20:04:27,365 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> TaskAttempt:
>> > [attempt_1453244277886_0001_m_000000_3] using containerId:
>> > [container_1453244277886_0001_01_000005 on NM: [localhost:35711]
>> > 2016-01-19 20:04:27,366 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_3 TaskAttempt Transitioned from
>> ASSIGNED
>> > to RUNNING
>> > 2016-01-19 20:04:27,366 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.speculate.DefaultSpeculator:
>> > ATTEMPT_START task_1453244277886_0001_m_000000
>> > 2016-01-19 20:04:28,361 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
>> getResources()
>> > for application_1453244277886_0001: ask=1 release= 0 newContainers=0
>> > finishedContainers=1 resourcelimit=<memory:768, vCores:-1> knownNMs=1
>> > 2016-01-19 20:04:28,361 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received
>> > completed container container_1453244277886_0001_01_000005
>> > 2016-01-19 20:04:28,361 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> Recalculating
>> > schedule, headroom=768
>> > 2016-01-19 20:04:28,361 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
>> > start threshold not met. completedMapsForReduceSlowstart 1
>> > 2016-01-19 20:04:28,362 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
>> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0
>> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:4 ContRel:0
>> > HostLocal:1 RackLocal:0
>> > 2016-01-19 20:04:28,362 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_3 TaskAttempt Transitioned from
>> RUNNING
>> > to FAIL_CONTAINER_CLEANUP
>> > 2016-01-19 20:04:28,362 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics
>> > report from attempt_1453244277886_0001_m_000000_3: Exception from
>> > container-launch: org.apache.hadoop.util.Shell$ExitCodeException:
>> > org.apache.hadoop.util.Shell$ExitCodeException:
>> >     at org.apache.hadoop.util.Shell.runCommand(Shell.java:505)
>> >     at org.apache.hadoop.util.Shell.run(Shell.java:418)
>> >     at
>> >
>> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:650)
>> >     at
>> >
>> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195)
>> >     at
>> >
>> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:300)
>> >     at
>> >
>> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:81)
>> >     at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>> >     at
>> >
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>> >     at
>> >
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>> >     at java.lang.Thread.run(Thread.java:745)
>> >
>> >
>> > Container exited with a non-zero exit code 1
>> >
>> > 2016-01-19 20:04:28,364 INFO [ContainerLauncher #7]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> > Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container
>> > container_1453244277886_0001_01_000005 taskAttempt
>> > attempt_1453244277886_0001_m_000000_3
>> > 2016-01-19 20:04:28,365 INFO [ContainerLauncher #7]
>> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
>> KILLING
>> > attempt_1453244277886_0001_m_000000_3
>> > 2016-01-19 20:04:28,373 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_3 TaskAttempt Transitioned from
>> > FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
>> > 2016-01-19 20:04:28,374 INFO [CommitterEvent Processor #4]
>> > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler:
>> Processing
>> > the event EventType: TASK_ABORT
>> > 2016-01-19 20:04:28,377 WARN [CommitterEvent Processor #4]
>> > org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not
>> delete
>> >
>> hdfs://hdnode01:54310/user/hduser/output/wordcount/_temporary/1/_temporary/attempt_1453244277886_0001_m_000000_3
>> > 2016-01-19 20:04:28,378 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_m_000000_3 TaskAttempt Transitioned from
>> > FAIL_TASK_CLEANUP to FAILED
>> > 2016-01-19 20:04:28,380 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl:
>> > task_1453244277886_0001_m_000000 Task Transitioned from RUNNING to
>> FAILED
>> > 2016-01-19 20:04:28,381 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Num completed
>> Tasks: 1
>> > 2016-01-19 20:04:28,381 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Job failed as tasks
>> > failed. failedMaps:1 failedReduces:0
>> > 2016-01-19 20:04:28,382 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl:
>> > job_1453244277886_0001Job Transitioned from RUNNING to FAIL_WAIT
>> > 2016-01-19 20:04:28,383 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl:
>> > task_1453244277886_0001_r_000000 Task Transitioned from SCHEDULED to
>> > KILL_WAIT
>> > 2016-01-19 20:04:28,383 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
>> > attempt_1453244277886_0001_r_000000_0 TaskAttempt Transitioned from
>> > UNASSIGNED to KILLED
>> > 2016-01-19 20:04:28,383 INFO [Thread-50]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Processing
>> the
>> > event EventType: CONTAINER_DEALLOCATE
>> > 2016-01-19 20:04:28,383 ERROR [Thread-50]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Could not
>> > deallocate container for task attemptId
>> > attempt_1453244277886_0001_r_000000_0
>> > 2016-01-19 20:04:28,383 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl:
>> > task_1453244277886_0001_r_000000 Task Transitioned from KILL_WAIT to
>> KILLED
>> > 2016-01-19 20:04:28,384 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl:
>> > job_1453244277886_0001Job Transitioned from FAIL_WAIT to FAIL_ABORT
>> > 2016-01-19 20:04:28,390 INFO [CommitterEvent Processor #0]
>> > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler:
>> Processing
>> > the event EventType: JOB_ABORT
>> > 2016-01-19 20:04:28,432 INFO [AsyncDispatcher event handler]
>> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl:
>> > job_1453244277886_0001Job Transitioned from FAIL_ABORT to FAILED
>> > 2016-01-19 20:04:28,433 INFO [Thread-61]
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: We are finishing
>> cleanly so
>> > this is the last retry
>> > 2016-01-19 20:04:28,433 INFO [Thread-61]
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Notify RMCommunicator
>> > isAMLastRetry: true
>> > 2016-01-19 20:04:28,433 INFO [Thread-61]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> RMCommunicator
>> > notified that shouldUnregistered is: true
>> > 2016-01-19 20:04:28,433 INFO [Thread-61]
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Notify JHEH
>> isAMLastRetry:
>> > true
>> > 2016-01-19 20:04:28,433 INFO [Thread-61]
>> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler:
>> > JobHistoryEventHandler notified that forceJobCompletion is true
>> > 2016-01-19 20:04:28,434 INFO [Thread-61]
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Calling stop for all the
>> > services
>> > 2016-01-19 20:04:28,435 INFO [Thread-61]
>> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Stopping
>> > JobHistoryEventHandler. Size of the outstanding queue size is 0
>> > 2016-01-19 20:04:29,362 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before
>> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0
>> > AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:4 ContRel:0
>> > HostLocal:1 RackLocal:0
>> > 2016-01-19 20:04:29,362 INFO [eventHandlingThread]
>> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copying
>> >
>> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/hduser/.staging/job_1453244277886_0001/job_1453244277886_0001_1.jhist
>> > to
>> >
>> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001-1453244648033-hduser-word+count-1453244668381-0-0-FAILED-default-1453244654166.jhist_tmp
>> > 2016-01-19 20:04:29,366 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
>> Recalculating
>> > schedule, headroom=768
>> > 2016-01-19 20:04:29,366 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
>> > start threshold reached. Scheduling reduces.
>> > 2016-01-19 20:04:29,366 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: All maps
>> > assigned. Ramping up all remaining reduces:1
>> > 2016-01-19 20:04:29,366 INFO [RMCommunicator Allocator]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
>> > Scheduling: PendingReds:0 ScheduledMaps:0 ScheduledReds:1 AssignedMaps:0
>> > AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:4 ContRel:0
>> > HostLocal:1 RackLocal:0
>> > 2016-01-19 20:04:29,544 INFO [eventHandlingThread]
>> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copied to
>> > done location:
>> >
>> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001-1453244648033-hduser-word+count-1453244668381-0-0-FAILED-default-1453244654166.jhist_tmp
>> > 2016-01-19 20:04:29,598 INFO [eventHandlingThread]
>> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copying
>> >
>> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/hduser/.staging/job_1453244277886_0001/job_1453244277886_0001_1_conf.xml
>> > to
>> >
>> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001_conf.xml_tmp
>> > 2016-01-19 20:04:29,801 INFO [eventHandlingThread]
>> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copied to
>> > done location:
>> >
>> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001_conf.xml_tmp
>> > 2016-01-19 20:04:29,907 INFO [eventHandlingThread]
>> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Moved
>> tmp to
>> > done:
>> >
>> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001.summary_tmp
>> > to
>> >
>> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001.summary
>> > 2016-01-19 20:04:30,008 INFO [eventHandlingThread]
>> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Moved
>> tmp to
>> > done:
>> >
>> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001_conf.xml_tmp
>> > to
>> >
>> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001_conf.xml
>> > 2016-01-19 20:04:30,070 INFO [eventHandlingThread]
>> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Moved
>> tmp to
>> > done:
>> >
>> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001-1453244648033-hduser-word+count-1453244668381-0-0-FAILED-default-1453244654166.jhist_tmp
>> > to
>> >
>> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001-1453244648033-hduser-word+count-1453244668381-0-0-FAILED-default-1453244654166.jhist
>> > 2016-01-19 20:04:30,071 INFO [Thread-61]
>> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Stopped
>> > JobHistoryEventHandler. super.stop()
>> > 2016-01-19 20:04:30,078 INFO [Thread-61]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Setting job
>> > diagnostics to Task failed task_1453244277886_0001_m_000000
>> > Job failed as tasks failed. failedMaps:1 failedReduces:0
>> >
>> > 2016-01-19 20:04:30,080 INFO [Thread-61]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: History url
>> is
>> > http://localhost:19888/jobhistory/job/job_1453244277886_0001
>> > 2016-01-19 20:04:30,094 INFO [Thread-61]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Waiting for
>> > application to be successfully unregistered.
>> > 2016-01-19 20:04:31,099 INFO [Thread-61]
>> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Final Stats:
>> > PendingReds:0 ScheduledMaps:0 ScheduledReds:1 AssignedMaps:0
>> AssignedReds:0
>> > CompletedMaps:1 CompletedReds:0 ContAlloc:4 ContRel:0 HostLocal:1
>> > RackLocal:0
>> > 2016-01-19 20:04:31,104 INFO [Thread-61]
>> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Deleting staging
>> directory
>> > hdfs://hdnode01:54310
>> > /tmp/hadoop-yarn/staging/hduser/.staging/job_1453244277886_0001
>> > 2016-01-19 20:04:31,133 INFO [Thread-61] org.apache.hadoop.ipc.Server:
>> > Stopping server on 45584
>> > 2016-01-19 20:04:31,135 INFO [IPC Server listener on 45584]
>> > org.apache.hadoop.ipc.Server: Stopping IPC Server listener on 45584
>> > 2016-01-19 20:04:31,135 INFO [IPC Server Responder]
>> > org.apache.hadoop.ipc.Server: Stopping IPC Server Responder
>> > 2016-01-19 20:04:31,136 INFO [TaskHeartbeatHandler PingChecker]
>> > org.apache.hadoop.mapreduce.v2.app.TaskHeartbeatHandler:
>> > TaskHeartbeatHandler thread interrupted
>> >
>> >
>> > Jps results, i believe that everything is ok, right?:
>> > 21267 DataNode
>> > 21609 ResourceManager
>> > 21974 JobHistoryServer
>> > 21735 NodeManager
>> > 24546 Jps
>> > 16532 org.eclipse.equinox.launcher_1.3.0.v20140415-2008.jar
>> > 21121 NameNode
>> > 22098 QuorumPeerMain
>> > 21456 SecondaryNameNode
>> >
>> >
>>
>
>

Mime
View raw message