hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From José Luis Larroque <larroques...@gmail.com>
Subject Re: Can't run hadoop examples with YARN Single node cluster
Date Tue, 26 Jan 2016 00:19:08 GMT
Thanks Namikaze for keep trying, don't give up!! :D

- I have these lines in *$HOME/.bashrc*


export HADOOP_PREFIX=/usr/local/hadoop

# Others variables

export HADOOP_COMMON_HOME=${HADOOP_PREFIX}

export HADOOP_MAPRED_HOME=${HADOOP_PREFIX}

export HADOOP_HDFS_HOME=${HADOOP_PREFIX}

export HADOOP_YARN_HOME=${HADOOP_PREFIX}


  - in *hadoop-env.sh* i have:

export HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/usr/local/hadoop/etc/hadoop"}


  - I read that SO question and all answers to it. The only useful answer
in my opinion was checking yarn classpath. I have three times the following
line:

/usr/local/hadoop/etc/hadoop:


I put yarn.application.classpath on yarn-site.xml because i don't know any
other way to fix it, with the value recomended for default in this
<https://hadoop.apache.org/docs/r2.4.1/hadoop-yarn/hadoop-yarn-common/yarn-default.xml>
(see for yarn.application.classpath):


$HADOOP_CONF_DIR, $HADOOP_COMMON_HOME/share/hadoop/common/*,
$HADOOP_COMMON_HOME/share/hadoop/common/lib/*,
$HADOOP_HDFS_HOME/share/hadoop/hdfs/*,
$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*,
$HADOOP_YARN_HOME/share/hadoop/yarn/*,
$HADOOP_YARN_HOME/share/hadoop/yarn/lib/*


But the classpath remains the same. And i can't find any other way to fix
it. Maybe this is the problem?


 - yarn.log-aggregation-enable was always set to true. I couldn't find
nothing in *datanodes logs*, here they are:

2016-01-25 21:13:07,006 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting DataNode
STARTUP_MSG:   host = jose-ubuntu/127.0.0.1
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 2.4.0
STARTUP_MSG:   classpath =
/usr/local/hadoop/etc/hadoop:/usr/local/hadoop/share/hadoop/common/lib/jackson-jaxrs-1.8.8.jar:/usr/local/hadoop/share/hadoop/common/lib/jersey-json-1.9.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-compress-1.4.1.jar:/usr/local/hadoop/share/hadoop/common/lib/httpcore-4.2.5.jar:/usr/local/hadoop/share/hadoop/common/lib/httpclient-4.2.5.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-beanutils-1.7.0.jar:/usr/local/hadoop/share/hadoop/common/lib/jettison-1.1.jar:/usr/local/hadoop/share/hadoop/common/lib/jaxb-impl-2.2.3-1.jar:/usr/local/hadoop/share/hadoop/common/lib/activation-1.1.jar:/usr/local/hadoop/share/hadoop/common/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-net-3.1.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-cli-1.2.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-codec-1.4.jar:/usr/local/hadoop/share/hadoop/common/lib/junit-4.8.2.jar:/usr/local/hadoop/share/hadoop/common/lib/paranamer-2.3.jar:/usr/local/hadoop/share/hadoop/common/lib/jasper-compiler-5.5.23.jar:/usr/local/hadoop/share/hadoop/common/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-digester-1.8.jar:/usr/local/hadoop/share/hadoop/common/lib/jets3t-0.9.0.jar:/usr/local/hadoop/share/hadoop/common/lib/jackson-xc-1.8.8.jar:/usr/local/hadoop/share/hadoop/common/lib/jersey-server-1.9.jar:/usr/local/hadoop/share/hadoop/common/lib/jetty-util-6.1.26.jar:/usr/local/hadoop/share/hadoop/common/lib/mockito-all-1.8.5.jar:/usr/local/hadoop/share/hadoop/common/lib/xmlenc-0.52.jar:/usr/local/hadoop/share/hadoop/common/lib/slf4j-api-1.7.5.jar:/usr/local/hadoop/share/hadoop/common/lib/stax-api-1.0-2.jar:/usr/local/hadoop/share/hadoop/common/lib/avro-1.7.4.jar:/usr/local/hadoop/share/hadoop/common/lib/hadoop-annotations-2.4.0.jar:/usr/local/hadoop/share/hadoop/common/lib/hadoop-auth-2.4.0.jar:/usr/local/hadoop/share/hadoop/common/lib/netty-3.6.2.Final.jar:/usr/local/hadoop/share/hadoop/common/lib/jasper-runtime-5.5.23.jar:/usr/local/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-logging-1.1.3.jar:/usr/local/hadoop/share/hadoop/common/lib/guava-11.0.2.jar:/usr/local/hadoop/share/hadoop/common/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop/share/hadoop/common/lib/jsp-api-2.1.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-configuration-1.6.jar:/usr/local/hadoop/share/hadoop/common/lib/snappy-java-1.0.4.1.jar:/usr/local/hadoop/share/hadoop/common/lib/log4j-1.2.17.jar:/usr/local/hadoop/share/hadoop/common/lib/jsch-0.1.42.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-io-2.4.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-beanutils-core-1.8.0.jar:/usr/local/hadoop/share/hadoop/common/lib/jsr305-1.3.9.jar:/usr/local/hadoop/share/hadoop/common/lib/jersey-core-1.9.jar:/usr/local/hadoop/share/hadoop/common/lib/asm-3.2.jar:/usr/local/hadoop/share/hadoop/common/lib/servlet-api-2.5.jar:/usr/local/hadoop/share/hadoop/common/lib/jaxb-api-2.2.2.jar:/usr/local/hadoop/share/hadoop/common/lib/jetty-6.1.26.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-lang-2.6.jar:/usr/local/hadoop/share/hadoop/common/lib/java-xmlbuilder-0.4.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-collections-3.2.1.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-el-1.0.jar:/usr/local/hadoop/share/hadoop/common/lib/zookeeper-3.4.5.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-httpclient-3.1.jar:/usr/local/hadoop/share/hadoop/common/lib/xz-1.0.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-math3-3.1.1.jar:/usr/local/hadoop/share/hadoop/common/hadoop-common-2.4.0-tests.jar:/usr/local/hadoop/share/hadoop/common/hadoop-nfs-2.4.0.jar:/usr/local/hadoop/share/hadoop/common/hadoop-common-2.4.0.jar:/usr/local/hadoop/share/hadoop/hdfs:/usr/local/hadoop/share/hadoop/hdfs/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-cli-1.2.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-codec-1.4.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jersey-server-1.9.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jetty-util-6.1.26.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/xmlenc-0.52.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-daemon-1.0.13.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/netty-3.6.2.Final.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jasper-runtime-5.5.23.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-logging-1.1.3.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/guava-11.0.2.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jsp-api-2.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/log4j-1.2.17.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-io-2.4.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jsr305-1.3.9.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jersey-core-1.9.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/asm-3.2.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/servlet-api-2.5.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jetty-6.1.26.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-lang-2.6.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-el-1.0.jar:/usr/local/hadoop/share/hadoop/hdfs/hadoop-hdfs-2.4.0-tests.jar:/usr/local/hadoop/share/hadoop/hdfs/hadoop-hdfs-2.4.0.jar:/usr/local/hadoop/share/hadoop/hdfs/hadoop-hdfs-nfs-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jackson-jaxrs-1.8.8.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jersey-json-1.9.jar:/usr/local/hadoop/share/hadoop/yarn/lib/commons-compress-1.4.1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jersey-client-1.9.jar:/usr/local/hadoop/share/hadoop/yarn/lib/guice-servlet-3.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/giraph-examples-1.1.0-for-hadoop-2.4.0-jar-with-dependencies.jar:/usr/local/hadoop/share/hadoop/yarn/lib/hadoop-mapreduce-examples-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jettison-1.1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jaxb-impl-2.2.3-1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/activation-1.1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop/share/hadoop/yarn/lib/commons-cli-1.2.jar:/usr/local/hadoop/share/hadoop/yarn/lib/commons-codec-1.4.jar:/usr/local/hadoop/share/hadoop/yarn/lib/aopalliance-1.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/leveldbjni-all-1.8.jar:/usr/local/hadoop/share/hadoop/yarn/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jackson-xc-1.8.8.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jersey-server-1.9.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jetty-util-6.1.26.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jline-0.9.94.jar:/usr/local/hadoop/share/hadoop/yarn/lib/stax-api-1.0-2.jar:/usr/local/hadoop/share/hadoop/yarn/lib/guice-3.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/commons-logging-1.1.3.jar:/usr/local/hadoop/share/hadoop/yarn/lib/guava-11.0.2.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop/share/hadoop/yarn/lib/log4j-1.2.17.jar:/usr/local/hadoop/share/hadoop/yarn/lib/commons-io-2.4.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jsr305-1.3.9.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jersey-core-1.9.jar:/usr/local/hadoop/share/hadoop/yarn/lib/asm-3.2.jar:/usr/local/hadoop/share/hadoop/yarn/lib/servlet-api-2.5.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jaxb-api-2.2.2.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jetty-6.1.26.jar:/usr/local/hadoop/share/hadoop/yarn/lib/commons-lang-2.6.jar:/usr/local/hadoop/share/hadoop/yarn/lib/commons-collections-3.2.1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jersey-guice-1.9.jar:/usr/local/hadoop/share/hadoop/yarn/lib/zookeeper-3.4.5.jar:/usr/local/hadoop/share/hadoop/yarn/lib/commons-httpclient-3.1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/javax.inject-1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/xz-1.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-common-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-client-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-api-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-applications-unmanaged-am-launcher-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-nodemanager-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-tests-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-resourcemanager-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-web-proxy-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-applications-distributedshell-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-applicationhistoryservice-2.4.0.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-common-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/commons-compress-1.4.1.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/guice-servlet-3.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/hamcrest-core-1.1.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/paranamer-2.3.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/aopalliance-1.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/jersey-server-1.9.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/avro-1.7.4.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/hadoop-annotations-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/guice-3.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/junit-4.10.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/netty-3.6.2.Final.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/snappy-java-1.0.4.1.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/log4j-1.2.17.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/commons-io-2.4.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/jersey-core-1.9.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/asm-3.2.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/jersey-guice-1.9.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/javax.inject-1.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/xz-1.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/giraph-examples-1.1.0-for-hadoop-2.4.0-jar-with-dependencies.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/procesamiento_salida_grafo_caminos_navegacionales-0.0.1.jar:/usr/local/hadoop/share/hadoop/mapreduce/zookeeper-3.4.6.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-shuffle-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/procesamiento_grafo_wikiquote-0.0.1.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-app-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-common-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-core-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/generacion_grafo_wikiquote-0.0.1-SNAPSHOT.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-2.4.0.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.4.0-tests.jar:/usr/local/hadoop/contrib/capacity-scheduler/*.jar:/usr/local/hadoop/contrib/capacity-scheduler/*.jar:/usr/local/hadoop/contrib/capacity-scheduler/*.jar
STARTUP_MSG:   build = http://svn.apache.org/repos/asf/hadoop/common
-r 1583262; compiled by 'jenkins' on 2014-03-31T08:29Z
STARTUP_MSG:   java = 1.7.0_79
************************************************************/
2016-01-25 21:13:07,015 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: registered UNIX
signal handlers for [TERM, HUP, INT]
2016-01-25 21:13:07,188 WARN
org.apache.hadoop.hdfs.server.common.Util: Path
/usr/local/hadoop/dfs/name/data should be specified as a URI in
configuration files. Please update hdfs configuration.
2016-01-25 21:13:07,356 WARN org.apache.hadoop.util.NativeCodeLoader:
Unable to load native-hadoop library for your platform... using
builtin-java classes where applicable
2016-01-25 21:13:07,648 INFO
org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from
hadoop-metrics2.properties
2016-01-25 21:13:07,723 INFO
org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
period at 10 second(s).
2016-01-25 21:13:07,723 INFO
org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics
system started
2016-01-25 21:13:07,727 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Configured hostname
is localhost
2016-01-25 21:13:07,728 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Starting DataNode
with maxLockedMemory = 0
2016-01-25 21:13:07,757 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Opened streaming
server at /0.0.0.0:50010
2016-01-25 21:13:07,760 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwith is
1048576 bytes/s
2016-01-25 21:13:07,839 INFO org.mortbay.log: Logging to
org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
org.mortbay.log.Slf4jLog
2016-01-25 21:13:07,843 INFO org.apache.hadoop.http.HttpRequestLog:
Http request log for http.requests.datanode is not defined
2016-01-25 21:13:07,853 INFO org.apache.hadoop.http.HttpServer2: Added
global filter 'safety'
(class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)
2016-01-25 21:13:07,856 INFO org.apache.hadoop.http.HttpServer2: Added
filter static_user_filter
(class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter)
to context datanode
2016-01-25 21:13:07,856 INFO org.apache.hadoop.http.HttpServer2: Added
filter static_user_filter
(class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter)
to context logs
2016-01-25 21:13:07,856 INFO org.apache.hadoop.http.HttpServer2: Added
filter static_user_filter
(class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter)
to context static
2016-01-25 21:13:07,872 INFO org.apache.hadoop.http.HttpServer2:
addJerseyResourcePackage:
packageName=org.apache.hadoop.hdfs.server.datanode.web.resources;org.apache.hadoop.hdfs.web.resources,
pathSpec=/webhdfs/v1/*
2016-01-25 21:13:07,875 INFO org.apache.hadoop.http.HttpServer2: Jetty
bound to port 50075
2016-01-25 21:13:07,875 INFO org.mortbay.log: jetty-6.1.26
2016-01-25 21:13:08,137 INFO org.mortbay.log: Started
SelectChannelConnector@0.0.0.0:50075
2016-01-25 21:13:08,225 INFO org.apache.hadoop.ipc.CallQueueManager:
Using callQueue class java.util.concurrent.LinkedBlockingQueue
2016-01-25 21:13:08,239 INFO org.apache.hadoop.ipc.Server: Starting
Socket Reader #1 for port 50020
2016-01-25 21:13:08,288 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Opened IPC server at
/0.0.0.0:50020
2016-01-25 21:13:08,300 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Refresh request
received for nameservices: null
2016-01-25 21:13:08,316 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Starting
BPOfferServices for nameservices: <default>
2016-01-25 21:13:08,321 WARN
org.apache.hadoop.hdfs.server.common.Util: Path
/usr/local/hadoop/dfs/name/data should be specified as a URI in
configuration files. Please update hdfs configuration.
2016-01-25 21:13:08,325 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool
<registering> (Datanode Uuid unassigned) service to
hdnode01/192.168.0.10:54310 starting to offer service
2016-01-25 21:13:08,326 INFO org.apache.hadoop.ipc.Server: IPC Server
Responder: starting
2016-01-25 21:13:08,326 INFO org.apache.hadoop.ipc.Server: IPC Server
listener on 50020: starting
2016-01-25 21:13:08,719 INFO
org.apache.hadoop.hdfs.server.common.Storage: Data-node version: -55
and name-node layout version: -56
2016-01-25 21:13:08,828 INFO
org.apache.hadoop.hdfs.server.common.Storage: Lock on
/usr/local/hadoop/dfs/name/data/in_use.lock acquired by nodename
10365@jose-ubuntu
2016-01-25 21:13:08,833 INFO
org.apache.hadoop.hdfs.server.common.Storage: Storage directory
/usr/local/hadoop/dfs/name/data is not formatted
2016-01-25 21:13:08,833 INFO
org.apache.hadoop.hdfs.server.common.Storage: Formatting ...
2016-01-25 21:13:09,017 INFO
org.apache.hadoop.hdfs.server.common.Storage: Analyzing storage
directories for bpid BP-216406264-127.0.0.1-1453767164845
2016-01-25 21:13:09,017 INFO
org.apache.hadoop.hdfs.server.common.Storage: Locking is disabled
2016-01-25 21:13:09,018 INFO
org.apache.hadoop.hdfs.server.common.Storage: Storage directory
/usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845
is not formatted.
2016-01-25 21:13:09,018 INFO
org.apache.hadoop.hdfs.server.common.Storage: Formatting ...
2016-01-25 21:13:09,018 INFO
org.apache.hadoop.hdfs.server.common.Storage: Formatting block pool
BP-216406264-127.0.0.1-1453767164845 directory
/usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current
2016-01-25 21:13:09,072 INFO
org.apache.hadoop.hdfs.server.common.Storage: Restored 0 block files
from trash.
2016-01-25 21:13:09,198 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Setting up storage:
nsid=1479061672;bpid=BP-216406264-127.0.0.1-1453767164845;lv=-55;nsInfo=lv=-56;cid=CID-8fa0e75b-6942-452a-a4e6-8cd0c24de652;nsid=1479061672;c=0;bpid=BP-216406264-127.0.0.1-1453767164845;dnuuid=null
2016-01-25 21:13:09,248 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Generated and
persisted new Datanode UUID 6b4236c8-2183-49ba-84d7-a273298ba37a
2016-01-25 21:13:09,268 WARN
org.apache.hadoop.hdfs.server.common.Util: Path
/usr/local/hadoop/dfs/name/data should be specified as a URI in
configuration files. Please update hdfs configuration.
2016-01-25 21:13:09,270 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl:
Added volume - /usr/local/hadoop/dfs/name/data/current, StorageType:
DISK
2016-01-25 21:13:09,279 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl:
Registered FSDatasetState MBean
2016-01-25 21:13:09,282 INFO
org.apache.hadoop.hdfs.server.datanode.DirectoryScanner: Periodic
Directory Tree Verification scan starting at 1453784080282 with
interval 21600000
2016-01-25 21:13:09,283 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl:
Adding block pool BP-216406264-127.0.0.1-1453767164845
2016-01-25 21:13:09,284 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl:
Scanning block pool BP-216406264-127.0.0.1-1453767164845 on volume
/usr/local/hadoop/dfs/name/data/current...
2016-01-25 21:13:09,299 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl:
Time taken to scan block pool BP-216406264-127.0.0.1-1453767164845 on
/usr/local/hadoop/dfs/name/data/current: 15ms
2016-01-25 21:13:09,300 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl:
Total time to scan all replicas for block pool
BP-216406264-127.0.0.1-1453767164845: 17ms
2016-01-25 21:13:09,301 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl:
Adding replicas to map for block pool
BP-216406264-127.0.0.1-1453767164845 on volume
/usr/local/hadoop/dfs/name/data/current...
2016-01-25 21:13:09,301 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl:
Time to add replicas to map for block pool
BP-216406264-127.0.0.1-1453767164845 on volume
/usr/local/hadoop/dfs/name/data/current: 0ms
2016-01-25 21:13:09,301 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl:
Total time to add all replicas to map: 1ms
2016-01-25 21:13:09,305 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool
BP-216406264-127.0.0.1-1453767164845 (Datanode Uuid null) service to
hdnode01/192.168.0.10:54310 beginning handshake with NN
2016-01-25 21:13:09,355 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool Block pool
BP-216406264-127.0.0.1-1453767164845 (Datanode Uuid null) service to
hdnode01/192.168.0.10:54310 successfully registered with NN
2016-01-25 21:13:09,356 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: For namenode
hdnode01/192.168.0.10:54310 using DELETEREPORT_INTERVAL of 300000 msec
 BLOCKREPORT_INTERVAL of 21600000msec CACHEREPORT_INTERVAL of
10000msec Initial delay: 0msec; heartBeatInterval=3000
2016-01-25 21:13:09,444 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Namenode Block pool
BP-216406264-127.0.0.1-1453767164845 (Datanode Uuid
6b4236c8-2183-49ba-84d7-a273298ba37a) service to
hdnode01/192.168.0.10:54310 trying to claim ACTIVE state with txid=1
2016-01-25 21:13:09,444 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Acknowledging ACTIVE
Namenode Block pool BP-216406264-127.0.0.1-1453767164845 (Datanode
Uuid 6b4236c8-2183-49ba-84d7-a273298ba37a) service to
hdnode01/192.168.0.10:54310
2016-01-25 21:13:09,487 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Sent 1 blockreports 0
blocks total. Took 1 msec to generate and 42 msecs for RPC and NN
processing.  Got back commands none
2016-01-25 21:13:09,492 INFO org.apache.hadoop.util.GSet: Computing
capacity for map BlockMap
2016-01-25 21:13:09,493 INFO org.apache.hadoop.util.GSet: VM type       = 64-bit
2016-01-25 21:13:09,494 INFO org.apache.hadoop.util.GSet: 0.5% max
memory 1.8 GB = 9.1 MB
2016-01-25 21:13:09,494 INFO org.apache.hadoop.util.GSet: capacity
 = 2^20 = 1048576 entries
2016-01-25 21:13:09,495 INFO
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Periodic
Block Verification Scanner initialized with interval 504 hours for
block pool BP-216406264-127.0.0.1-1453767164845
2016-01-25 21:13:09,499 INFO
org.apache.hadoop.hdfs.server.datanode.DataBlockScanner: Added
bpid=BP-216406264-127.0.0.1-1453767164845 to blockPoolScannerMap, new
size=1
2016-01-25 21:13:32,355 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving
BP-216406264-127.0.0.1-1453767164845:blk_1073741825_1001 src:
/192.168.0.10:58649 dest: /192.168.0.10:50010
2016-01-25 21:13:32,482 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:58649, dest: /192.168.0.10:50010, bytes: 343691, op:
HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_538002429_1, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741825_1001, duration:
98632367
2016-01-25 21:13:32,482 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder:
BP-216406264-127.0.0.1-1453767164845:blk_1073741825_1001,
type=LAST_IN_PIPELINE, downstreams=0:[] terminating
2016-01-25 21:13:34,291 INFO
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner:
Verification succeeded for
BP-216406264-127.0.0.1-1453767164845:blk_1073741825_1001
2016-01-25 21:14:10,176 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving
BP-216406264-127.0.0.1-1453767164845:blk_1073741826_1002 src:
/192.168.0.10:58663 dest: /192.168.0.10:50010
2016-01-25 21:14:10,220 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:58663, dest: /192.168.0.10:50010, bytes: 270263, op:
HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_342504113_1, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741826_1002, duration:
42378742
2016-01-25 21:14:10,221 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder:
BP-216406264-127.0.0.1-1453767164845:blk_1073741826_1002,
type=LAST_IN_PIPELINE, downstreams=0:[] terminating
2016-01-25 21:14:10,714 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving
BP-216406264-127.0.0.1-1453767164845:blk_1073741827_1003 src:
/192.168.0.10:58664 dest: /192.168.0.10:50010
2016-01-25 21:14:10,721 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:58664, dest: /192.168.0.10:50010, bytes: 121, op:
HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_342504113_1, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741827_1003, duration:
2656758
2016-01-25 21:14:10,721 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder:
BP-216406264-127.0.0.1-1453767164845:blk_1073741827_1003,
type=LAST_IN_PIPELINE, downstreams=0:[] terminating
2016-01-25 21:14:10,853 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving
BP-216406264-127.0.0.1-1453767164845:blk_1073741828_1004 src:
/192.168.0.10:58665 dest: /192.168.0.10:50010
2016-01-25 21:14:10,860 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:58665, dest: /192.168.0.10:50010, bytes: 26, op:
HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_342504113_1, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741828_1004, duration:
3257396
2016-01-25 21:14:10,861 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder:
BP-216406264-127.0.0.1-1453767164845:blk_1073741828_1004,
type=LAST_IN_PIPELINE, downstreams=0:[] terminating
2016-01-25 21:14:11,717 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving
BP-216406264-127.0.0.1-1453767164845:blk_1073741829_1005 src:
/192.168.0.10:58666 dest: /192.168.0.10:50010
2016-01-25 21:14:11,726 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:58666, dest: /192.168.0.10:50010, bytes: 77957, op:
HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_342504113_1, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741829_1005, duration:
6180229
2016-01-25 21:14:11,727 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder:
BP-216406264-127.0.0.1-1453767164845:blk_1073741829_1005,
type=LAST_IN_PIPELINE, downstreams=0:[] terminating
2016-01-25 21:14:14,298 INFO
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner:
Verification succeeded for
BP-216406264-127.0.0.1-1453767164845:blk_1073741829_1005
2016-01-25 21:14:14,299 INFO
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner:
Verification succeeded for
BP-216406264-127.0.0.1-1453767164845:blk_1073741827_1003
2016-01-25 21:14:14,305 INFO
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner:
Verification succeeded for
BP-216406264-127.0.0.1-1453767164845:blk_1073741826_1002
2016-01-25 21:14:14,305 INFO
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner:
Verification succeeded for
BP-216406264-127.0.0.1-1453767164845:blk_1073741828_1004
2016-01-25 21:14:16,099 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:50010, dest: /192.168.0.10:58674, bytes: 272375, op:
HDFS_READ, cliID: DFSClient_NONMAPREDUCE_76231625_102, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741826_1002, duration:
2878920
2016-01-25 21:14:16,253 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:50010, dest: /192.168.0.10:58674, bytes: 30, op:
HDFS_READ, cliID: DFSClient_NONMAPREDUCE_76231625_102, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741828_1004, duration:
236423
2016-01-25 21:14:16,312 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:50010, dest: /192.168.0.10:58674, bytes: 125, op:
HDFS_READ, cliID: DFSClient_NONMAPREDUCE_76231625_102, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741827_1003, duration:
909236
2016-01-25 21:14:16,364 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:50010, dest: /192.168.0.10:58674, bytes: 78569, op:
HDFS_READ, cliID: DFSClient_NONMAPREDUCE_76231625_102, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741829_1005, duration:
1489437
2016-01-25 21:14:20,174 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:50010, dest: /192.168.0.10:58676, bytes: 30, op:
HDFS_READ, cliID: DFSClient_NONMAPREDUCE_694066886_1, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741828_1004, duration:
899980
2016-01-25 21:14:22,692 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving
BP-216406264-127.0.0.1-1453767164845:blk_1073741830_1006 src:
/192.168.0.10:58679 dest: /192.168.0.10:50010
2016-01-25 21:14:22,754 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:58679, dest: /192.168.0.10:50010, bytes: 92684, op:
HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_694066886_1, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741830_1006, duration:
60114851
2016-01-25 21:14:22,754 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder:
BP-216406264-127.0.0.1-1453767164845:blk_1073741830_1006,
type=LAST_IN_PIPELINE, downstreams=0:[] terminating
2016-01-25 21:14:24,319 INFO
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner:
Verification succeeded for
BP-216406264-127.0.0.1-1453767164845:blk_1073741830_1006
2016-01-25 21:14:25,808 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving
BP-216406264-127.0.0.1-1453767164845:blk_1073741831_1007 src:
/192.168.0.10:58681 dest: /192.168.0.10:50010
2016-01-25 21:14:35,846 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:58681, dest: /192.168.0.10:50010, bytes: 21176, op:
HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_694066886_1, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741831_1007, duration:
9975409048
2016-01-25 21:14:35,846 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder:
BP-216406264-127.0.0.1-1453767164845:blk_1073741831_1007,
type=LAST_IN_PIPELINE, downstreams=0:[] terminating
2016-01-25 21:14:36,066 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving
BP-216406264-127.0.0.1-1453767164845:blk_1073741832_1008 src:
/192.168.0.10:58682 dest: /192.168.0.10:50010
2016-01-25 21:14:36,075 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:58682, dest: /192.168.0.10:50010, bytes: 332, op:
HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_694066886_1, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741832_1008, duration:
4992595
2016-01-25 21:14:36,075 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder:
BP-216406264-127.0.0.1-1453767164845:blk_1073741832_1008,
type=LAST_IN_PIPELINE, downstreams=0:[] terminating
2016-01-25 21:14:36,548 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:50010, dest: /192.168.0.10:58683, bytes: 21344, op:
HDFS_READ, cliID: DFSClient_NONMAPREDUCE_694066886_1, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741831_1007, duration:
497225
2016-01-25 21:14:36,564 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving
BP-216406264-127.0.0.1-1453767164845:blk_1073741833_1009 src:
/192.168.0.10:58684 dest: /192.168.0.10:50010
2016-01-25 21:14:36,572 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:58684, dest: /192.168.0.10:50010, bytes: 21176, op:
HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_694066886_1, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741833_1009, duration:
2649337
2016-01-25 21:14:36,573 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder:
BP-216406264-127.0.0.1-1453767164845:blk_1073741833_1009,
type=LAST_IN_PIPELINE, downstreams=0:[] terminating
2016-01-25 21:14:36,622 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:50010, dest: /192.168.0.10:58683, bytes: 93412, op:
HDFS_READ, cliID: DFSClient_NONMAPREDUCE_694066886_1, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741830_1006, duration:
379439
2016-01-25 21:14:36,638 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving
BP-216406264-127.0.0.1-1453767164845:blk_1073741834_1010 src:
/192.168.0.10:58685 dest: /192.168.0.10:50010
2016-01-25 21:14:36,646 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:58685, dest: /192.168.0.10:50010, bytes: 92684, op:
HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_694066886_1, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741834_1010, duration:
3135698
2016-01-25 21:14:36,646 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder:
BP-216406264-127.0.0.1-1453767164845:blk_1073741834_1010,
type=LAST_IN_PIPELINE, downstreams=0:[] terminating
2016-01-25 21:14:39,335 INFO
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner:
Verification succeeded for
BP-216406264-127.0.0.1-1453767164845:blk_1073741833_1009
2016-01-25 21:14:39,336 INFO
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner:
Verification succeeded for
BP-216406264-127.0.0.1-1453767164845:blk_1073741831_1007
2016-01-25 21:14:39,337 INFO
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner:
Verification succeeded for
BP-216406264-127.0.0.1-1453767164845:blk_1073741832_1008
2016-01-25 21:14:39,338 INFO
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner:
Verification succeeded for
BP-216406264-127.0.0.1-1453767164845:blk_1073741834_1010
2016-01-25 21:14:39,376 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
Scheduling blk_1073741826_1002 file
/usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741826
for deletion
2016-01-25 21:14:39,379 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
Scheduling blk_1073741827_1003 file
/usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741827
for deletion
2016-01-25 21:14:39,379 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
Deleted BP-216406264-127.0.0.1-1453767164845 blk_1073741826_1002 file
/usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741826
2016-01-25 21:14:39,379 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
Scheduling blk_1073741828_1004 file
/usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741828
for deletion
2016-01-25 21:14:39,380 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
Deleted BP-216406264-127.0.0.1-1453767164845 blk_1073741827_1003 file
/usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741827
2016-01-25 21:14:39,380 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
Scheduling blk_1073741829_1005 file
/usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741829
for deletion
2016-01-25 21:14:39,380 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
Deleted BP-216406264-127.0.0.1-1453767164845 blk_1073741828_1004 file
/usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741828
2016-01-25 21:14:39,380 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
Scheduling blk_1073741830_1006 file
/usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741830
for deletion
2016-01-25 21:14:39,381 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
Deleted BP-216406264-127.0.0.1-1453767164845 blk_1073741829_1005 file
/usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741829
2016-01-25 21:14:39,381 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
Scheduling blk_1073741831_1007 file
/usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741831
for deletion
2016-01-25 21:14:39,382 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
Deleted BP-216406264-127.0.0.1-1453767164845 blk_1073741830_1006 file
/usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741830
2016-01-25 21:14:39,382 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
Deleted BP-216406264-127.0.0.1-1453767164845 blk_1073741831_1007 file
/usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741831
2016-01-25 21:14:44,797 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving
BP-216406264-127.0.0.1-1453767164845:blk_1073741835_1011 src:
/192.168.0.10:58688 dest: /192.168.0.10:50010
2016-01-25 21:14:44,834 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:58688, dest: /192.168.0.10:50010, bytes: 57450, op:
HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_-433405715_88, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741835_1011, duration:
34522284
2016-01-25 21:14:44,834 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder:
BP-216406264-127.0.0.1-1453767164845:blk_1073741835_1011,
type=LAST_IN_PIPELINE, downstreams=0:[] terminating
2016-01-25 21:14:49,343 INFO
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner:
Verification succeeded for
BP-216406264-127.0.0.1-1453767164845:blk_1073741835_1011
2016-01-25 21:16:33,785 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.0.10:50010, dest: /192.168.0.10:58694, bytes: 336, op:
HDFS_READ, cliID: DFSClient_NONMAPREDUCE_-1832227986_1, offset: 0,
srvID: 6b4236c8-2183-49ba-84d7-a273298ba37a, blockid:
BP-216406264-127.0.0.1-1453767164845:blk_1073741832_1008, duration:
284719
2016-01-25 21:16:36,371 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
Scheduling blk_1073741832_1008 file
/usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741832
for deletion
2016-01-25 21:16:36,372 INFO
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
Deleted BP-216406264-127.0.0.1-1453767164845 blk_1073741832_1008 file
/usr/local/hadoop/dfs/name/data/current/BP-216406264-127.0.0.1-1453767164845/current/finalized/blk_1073741832




2016-01-21 18:52 GMT-03:00 Namikaze Minato <lloydsensei@gmail.com>:

> It could be a classpath issue (see
> http://stackoverflow.com/a/25090151/4486184), I'm strongly thinking
> this is the case.
>
> You could drill down to the exact root cause by looking at the
> datanode logs (see
>
> http://mail-archives.apache.org/mod_mbox/hadoop-user/201410.mbox/%3CCAEMetGubzq12LXbLRk6N4ejOoKse9dLEWMW8_WE6aRj=+RQtVw@mail.gmail.com%3E
> )
> But I'm not sure we would get another error than what we had...
>
> Check if your application has the correct values for the following
> variables:
> HADOOP_CONF_DIR
> HADOOP_COMMON_HOME
> HADOOP_HDFS_HOME
> HADOOP_MAPRED_HOME
> HADOOP_YARN_HOME
>
> I'm afraid I can't help you much more than this myself, sorry...
>
> LLoyd
>
> On 20 January 2016 at 02:08, José Luis Larroque <larroquester@gmail.com>
> wrote:
> > Hi guys, thanks for your answers.
> >
> > Wordcount logs:
> >
> > 16/01/19 21:58:32 INFO client.RMProxy: Connecting to ResourceManager at
> > hdnode01/192.168.0.10:8050
> > SLF4J: Class path contains multiple SLF4J bindings.
> > SLF4J: Found binding in
> >
> [jar:file:/usr/local/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> > SLF4J: Found binding in
> >
> [jar:file:/usr/local/hadoop/share/hadoop/yarn/lib/giraph-examples-1.1.0-for-hadoop-2.4.0-jar-with-dependencies.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> > SLF4J: Found binding in
> >
> [jar:file:/usr/local/hadoop/share/hadoop/mapreduce/giraph-examples-1.1.0-for-hadoop-2.4.0-jar-with-dependencies.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> > explanation.
> > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> > 16/01/19 21:58:32 WARN util.NativeCodeLoader: Unable to load
> native-hadoop
> > library for your platform... using builtin-java classes where applicable
> > hduser@jose-ubuntu:/usr/local/hadoop$ nano
> >
> /home/hduser/Desktop/Tesina/casos_de_prueba/resultados/resultado_cluster_modo_yarn_wordcount
> > hduser@jose-ubuntu:/usr/local/hadoop$ nano
> >
> /home/hduser/Desktop/Tesina/casos_de_prueba/resultados/resultado_cluster_modo_yarn_wordcount
> > hduser@jose-ubuntu:/usr/local/hadoop$ cat
> >
> /home/hduser/Desktop/Tesina/casos_de_prueba/resultados/resultado_cluster_modo_yarn_wordcount
> >
> >
> > Container: container_1453244277886_0001_01_000002 on localhost_35711
> > ======================================================================
> > LogType: stderr
> > LogLength: 45
> > Log Contents:
> > Error: Could not find or load main class 256
> >
> > LogType: stdout
> > LogLength: 0
> > Log Contents:
> >
> >
> >
> > Container: container_1453244277886_0001_01_000003 on localhost_35711
> > ======================================================================
> > LogType: stderr
> > LogLength: 45
> > Log Contents:
> > Error: Could not find or load main class 256
> >
> > LogType: stdout
> > LogLength: 0
> > Log Contents:
> >
> >
> >
> > Container: container_1453244277886_0001_01_000004 on localhost_35711
> > ======================================================================
> > LogType: stderr
> > LogLength: 45
> > Log Contents:
> > Error: Could not find or load main class 256
> >
> > LogType: stdout
> > LogLength: 0
> > Log Contents:
> >
> >
> >
> > Container: container_1453244277886_0001_01_000005 on localhost_35711
> > ======================================================================
> > LogType: stderr
> > LogLength: 45
> > Log Contents:
> > Error: Could not find or load main class 256
> >
> > LogType: stdout
> > LogLength: 0
> > Log Contents:
> >
> >
> >
> > Container: container_1453244277886_0001_01_000001 on localhost_35711
> > ======================================================================
> > LogType: stderr
> > LogLength: 929
> > Log Contents:
> > SLF4J: Class path contains multiple SLF4J bindings.
> > SLF4J: Found binding in
> >
> [jar:file:/usr/local/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> > SLF4J: Found binding in
> >
> [jar:file:/usr/local/hadoop/share/hadoop/yarn/lib/giraph-examples-1.1.0-for-hadoop-2.4.0-jar-with-dependencies.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> > SLF4J: Found binding in
> >
> [jar:file:/usr/local/hadoop/share/hadoop/mapreduce/giraph-examples-1.1.0-for-hadoop-2.4.0-jar-with-dependencies.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> > explanation.
> > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> > log4j:WARN No appenders could be found for logger
> > (org.apache.hadoop.ipc.Server).
> > log4j:WARN Please initialize the log4j system properly.
> > log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for
> > more info.
> >
> > LogType: stdout
> > LogLength: 0
> > Log Contents:
> >
> > LogType: syslog
> > LogLength: 56780
> > Log Contents:
> > 2016-01-19 20:04:11,329 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Created MRAppMaster for
> > application appattempt_1453244277886_0001_000001
> > 2016-01-19 20:04:11,657 WARN [main] org.apache.hadoop.conf.Configuration:
> > job.xml:an attempt to override final parameter:
> > mapreduce.job.end-notification.max.retry.interval;  Ignoring.
> > 2016-01-19 20:04:11,674 WARN [main] org.apache.hadoop.conf.Configuration:
> > job.xml:an attempt to override final parameter:
> > mapreduce.job.end-notification.max.attempts;  Ignoring.
> > 2016-01-19 20:04:11,765 WARN [main]
> org.apache.hadoop.util.NativeCodeLoader:
> > Unable to load native-hadoop library for your platform... using
> builtin-java
> > classes where applicable
> > 2016-01-19 20:04:11,776 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Executing with tokens:
> > 2016-01-19 20:04:11,776 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Kind: YARN_AM_RM_TOKEN,
> > Service: , Ident:
> > (org.apache.hadoop.yarn.security.AMRMTokenIdentifier@73e8f4b9)
> > 2016-01-19 20:04:11,801 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: The specific max
> attempts: 2
> > for application: 1. Attempt num: 1 is last retry: false
> > 2016-01-19 20:04:11,806 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Using mapred
> > newApiCommitter.
> > 2016-01-19 20:04:11,934 WARN [main] org.apache.hadoop.conf.Configuration:
> > job.xml:an attempt to override final parameter: dfs.datanode.data.dir;
> > Ignoring.
> > 2016-01-19 20:04:11,939 WARN [main] org.apache.hadoop.conf.Configuration:
> > job.xml:an attempt to override final parameter:
> > mapreduce.job.end-notification.max.retry.interval;  Ignoring.
> > 2016-01-19 20:04:11,948 WARN [main] org.apache.hadoop.conf.Configuration:
> > job.xml:an attempt to override final parameter: dfs.namenode.name.dir;
> > Ignoring.
> > 2016-01-19 20:04:11,953 WARN [main] org.apache.hadoop.conf.Configuration:
> > job.xml:an attempt to override final parameter:
> > mapreduce.job.end-notification.max.attempts;  Ignoring.
> > 2016-01-19 20:04:12,464 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter set in
> > config null
> > 2016-01-19 20:04:12,526 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter is
> > org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
> > 2016-01-19 20:04:12,548 INFO [main]
> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
> > org.apache.hadoop.mapreduce.jobhistory.EventType for class
> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler
> > 2016-01-19 20:04:12,549 INFO [main]
> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
> > org.apache.hadoop.mapreduce.v2.app.job.event.JobEventType for class
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobEventDispatcher
> > 2016-01-19 20:04:12,550 INFO [main]
> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
> > org.apache.hadoop.mapreduce.v2.app.job.event.TaskEventType for class
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskEventDispatcher
> > 2016-01-19 20:04:12,551 INFO [main]
> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
> > org.apache.hadoop.mapreduce.v2.app.job.event.TaskAttemptEventType for
> class
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskAttemptEventDispatcher
> > 2016-01-19 20:04:12,552 INFO [main]
> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
> > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventType for class
> > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler
> > 2016-01-19 20:04:12,557 INFO [main]
> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
> > org.apache.hadoop.mapreduce.v2.app.speculate.Speculator$EventType for
> class
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster$SpeculatorEventDispatcher
> > 2016-01-19 20:04:12,558 INFO [main]
> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
> > org.apache.hadoop.mapreduce.v2.app.rm.ContainerAllocator$EventType for
> class
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerAllocatorRouter
> > 2016-01-19 20:04:12,559 INFO [main]
> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncher$EventType
> for
> > class
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerLauncherRouter
> > 2016-01-19 20:04:12,615 INFO [main]
> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Perms
> after
> > creating 488, Expected: 504
> > 2016-01-19 20:04:12,615 INFO [main]
> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Explicitly
> > setting permissions to : 504, rwxrwx---
> > 2016-01-19 20:04:12,731 INFO [main]
> > org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class
> > org.apache.hadoop.mapreduce.v2.app.job.event.JobFinishEvent$Type for
> class
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobFinishEventHandler
> > 2016-01-19 20:04:12,956 INFO [main]
> > org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from
> > hadoop-metrics2.properties
> > 2016-01-19 20:04:13,018 INFO [main]
> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
> period
> > at 10 second(s).
> > 2016-01-19 20:04:13,018 INFO [main]
> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: MRAppMaster metrics
> > system started
> > 2016-01-19 20:04:13,026 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Adding job token for
> > job_1453244277886_0001 to jobTokenSecretManager
> > 2016-01-19 20:04:13,139 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Not uberizing
> > job_1453244277886_0001 because: not enabled;
> > 2016-01-19 20:04:13,154 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Input size for job
> > job_1453244277886_0001 = 343691. Number of splits = 1
> > 2016-01-19 20:04:13,156 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Number of reduces
> for
> > job job_1453244277886_0001 = 1
> > 2016-01-19 20:04:13,156 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl:
> > job_1453244277886_0001Job Transitioned from NEW to INITED
> > 2016-01-19 20:04:13,157 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: MRAppMaster launching
> > normal, non-uberized, multi-container job job_1453244277886_0001.
> > 2016-01-19 20:04:13,186 INFO [main]
> org.apache.hadoop.ipc.CallQueueManager:
> > Using callQueue class java.util.concurrent.LinkedBlockingQueue
> > 2016-01-19 20:04:13,195 INFO [Socket Reader #1 for port 56461]
> > org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 56461
> > 2016-01-19 20:04:13,237 INFO [main]
> > org.apache.hadoop.yarn.factories.impl.pb.RpcServerFactoryPBImpl: Adding
> > protocol org.apache.hadoop.mapreduce.v2.api.MRClientProtocolPB to the
> server
> > 2016-01-19 20:04:13,238 INFO [IPC Server Responder]
> > org.apache.hadoop.ipc.Server: IPC Server Responder: starting
> > 2016-01-19 20:04:13,238 INFO [IPC Server listener on 56461]
> > org.apache.hadoop.ipc.Server: IPC Server listener on 56461: starting
> > 2016-01-19 20:04:13,239 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.client.MRClientService: Instantiated
> > MRClientService at jose-ubuntu/127.0.0.1:56461
> > 2016-01-19 20:04:13,300 INFO [main] org.mortbay.log: Logging to
> > org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
> > org.mortbay.log.Slf4jLog
> > 2016-01-19 20:04:13,304 INFO [main]
> org.apache.hadoop.http.HttpRequestLog:
> > Http request log for http.requests.mapreduce is not defined
> > 2016-01-19 20:04:13,315 INFO [main] org.apache.hadoop.http.HttpServer2:
> > Added global filter 'safety'
> > (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)
> > 2016-01-19 20:04:13,320 INFO [main] org.apache.hadoop.http.HttpServer2:
> > Added filter AM_PROXY_FILTER
> > (class=org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter) to
> > context mapreduce
> > 2016-01-19 20:04:13,321 INFO [main] org.apache.hadoop.http.HttpServer2:
> > Added filter AM_PROXY_FILTER
> > (class=org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter) to
> > context static
> > 2016-01-19 20:04:13,324 INFO [main] org.apache.hadoop.http.HttpServer2:
> > adding path spec: /mapreduce/*
> > 2016-01-19 20:04:13,324 INFO [main] org.apache.hadoop.http.HttpServer2:
> > adding path spec: /ws/*
> > 2016-01-19 20:04:13,335 INFO [main] org.apache.hadoop.http.HttpServer2:
> > Jetty bound to port 44070
> > 2016-01-19 20:04:13,335 INFO [main] org.mortbay.log: jetty-6.1.26
> > 2016-01-19 20:04:13,370 INFO [main] org.mortbay.log: Extract
> >
> jar:file:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-common-2.4.0.jar!/webapps/mapreduce
> > to /tmp/Jetty_0_0_0_0_44070_mapreduce____rdpvio/webapp
> > 2016-01-19 20:04:13,647 INFO [main] org.mortbay.log: Started
> > SelectChannelConnector@0.0.0.0:44070
> > 2016-01-19 20:04:13,647 INFO [main]
> org.apache.hadoop.yarn.webapp.WebApps:
> > Web app /mapreduce started at 44070
> > 2016-01-19 20:04:13,956 INFO [main]
> org.apache.hadoop.yarn.webapp.WebApps:
> > Registered webapp guice modules
> > 2016-01-19 20:04:13,960 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.speculate.DefaultSpeculator:
> JOB_CREATE
> > job_1453244277886_0001
> > 2016-01-19 20:04:13,961 INFO [main]
> org.apache.hadoop.ipc.CallQueueManager:
> > Using callQueue class java.util.concurrent.LinkedBlockingQueue
> > 2016-01-19 20:04:13,961 INFO [Socket Reader #1 for port 45584]
> > org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 45584
> > 2016-01-19 20:04:13,966 INFO [IPC Server Responder]
> > org.apache.hadoop.ipc.Server: IPC Server Responder: starting
> > 2016-01-19 20:04:13,966 INFO [IPC Server listener on 45584]
> > org.apache.hadoop.ipc.Server: IPC Server listener on 45584: starting
> > 2016-01-19 20:04:13,987 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
> > nodeBlacklistingEnabled:true
> > 2016-01-19 20:04:13,987 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
> > maxTaskFailuresPerNode is 3
> > 2016-01-19 20:04:13,988 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
> > blacklistDisablePercent is 33
> > 2016-01-19 20:04:14,052 WARN [main] org.apache.hadoop.conf.Configuration:
> > job.xml:an attempt to override final parameter: dfs.datanode.data.dir;
> > Ignoring.
> > 2016-01-19 20:04:14,054 WARN [main] org.apache.hadoop.conf.Configuration:
> > job.xml:an attempt to override final parameter:
> > mapreduce.job.end-notification.max.retry.interval;  Ignoring.
> > 2016-01-19 20:04:14,057 WARN [main] org.apache.hadoop.conf.Configuration:
> > job.xml:an attempt to override final parameter: dfs.namenode.name.dir;
> > Ignoring.
> > 2016-01-19 20:04:14,059 WARN [main] org.apache.hadoop.conf.Configuration:
> > job.xml:an attempt to override final parameter:
> > mapreduce.job.end-notification.max.attempts;  Ignoring.
> > 2016-01-19 20:04:14,062 INFO [main]
> org.apache.hadoop.yarn.client.RMProxy:
> > Connecting to ResourceManager at hdnode01/192.168.0.10:8030
> > 2016-01-19 20:04:14,158 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
> > maxContainerCapability: 2000
> > 2016-01-19 20:04:14,158 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: queue:
> default
> > 2016-01-19 20:04:14,162 INFO [main]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Upper
> > limit on the thread pool size is 500
> > 2016-01-19 20:04:14,164 INFO [main]
> > org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy:
> > yarn.client.max-nodemanagers-proxies : 500
> > 2016-01-19 20:04:14,172 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl:
> > job_1453244277886_0001Job Transitioned from INITED to SETUP
> > 2016-01-19 20:04:14,174 INFO [CommitterEvent Processor #0]
> > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler:
> Processing
> > the event EventType: JOB_SETUP
> > 2016-01-19 20:04:14,210 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl:
> > job_1453244277886_0001Job Transitioned from SETUP to RUNNING
> > 2016-01-19 20:04:14,227 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.yarn.util.RackResolver: Resolved localhost to
> > /default-rack
> > 2016-01-19 20:04:14,230 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl:
> > task_1453244277886_0001_m_000000 Task Transitioned from NEW to SCHEDULED
> > 2016-01-19 20:04:14,230 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl:
> > task_1453244277886_0001_r_000000 Task Transitioned from NEW to SCHEDULED
> > 2016-01-19 20:04:14,232 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_0 TaskAttempt Transitioned from NEW
> to
> > UNASSIGNED
> > 2016-01-19 20:04:14,232 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_r_000000_0 TaskAttempt Transitioned from NEW
> to
> > UNASSIGNED
> > 2016-01-19 20:04:14,233 INFO [Thread-50]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
> > mapResourceReqt:512
> > 2016-01-19 20:04:14,245 INFO [Thread-50]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
> > reduceResourceReqt:512
> > 2016-01-19 20:04:14,324 INFO [eventHandlingThread]
> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Event
> Writer
> > setup for JobId: job_1453244277886_0001, File:
> >
> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/hduser/.staging/job_1453244277886_0001/job_1453244277886_0001_1.jhist
> > 2016-01-19 20:04:15,162 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before
> > Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:0
> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:0 ContRel:0
> > HostLocal:0 RackLocal:0
> > 2016-01-19 20:04:15,217 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
> getResources()
> > for application_1453244277886_0001: ask=3 release= 0 newContainers=0
> > finishedContainers=0 resourcelimit=<memory:1280, vCores:0> knownNMs=1
> > 2016-01-19 20:04:15,217 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating
> > schedule, headroom=1280
> > 2016-01-19 20:04:15,217 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
> > start threshold not met. completedMapsForReduceSlowstart 1
> > 2016-01-19 20:04:16,240 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated
> > containers 1
> > 2016-01-19 20:04:16,241 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned
> > container container_1453244277886_0001_01_000002 to
> > attempt_1453244277886_0001_m_000000_0
> > 2016-01-19 20:04:16,243 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating
> > schedule, headroom=768
> > 2016-01-19 20:04:16,243 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
> > start threshold not met. completedMapsForReduceSlowstart 1
> > 2016-01-19 20:04:16,243 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1
> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:1 ContRel:0
> > HostLocal:1 RackLocal:0
> > 2016-01-19 20:04:16,291 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.yarn.util.RackResolver: Resolved localhost to
> > /default-rack
> > 2016-01-19 20:04:16,316 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: The job-jar
> > file on the remote FS is
> >
> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/hduser/.staging/job_1453244277886_0001/job.jar
> > 2016-01-19 20:04:16,322 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: The job-conf
> > file on the remote FS is
> > /tmp/hadoop-yarn/staging/hduser/.staging/job_1453244277886_0001/job.xml
> > 2016-01-19 20:04:16,325 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Adding #0
> > tokens and #1 secret keys for NM use for launching container
> > 2016-01-19 20:04:16,325 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Size of
> > containertokens_dob is 1
> > 2016-01-19 20:04:16,325 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Putting
> shuffle
> > token in serviceData
> > 2016-01-19 20:04:16,350 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_0 TaskAttempt Transitioned from
> > UNASSIGNED to ASSIGNED
> > 2016-01-19 20:04:16,354 INFO [ContainerLauncher #0]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> > Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container
> > container_1453244277886_0001_01_000002 taskAttempt
> > attempt_1453244277886_0001_m_000000_0
> > 2016-01-19 20:04:16,356 INFO [ContainerLauncher #0]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> Launching
> > attempt_1453244277886_0001_m_000000_0
> > 2016-01-19 20:04:16,357 INFO [ContainerLauncher #0]
> > org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy:
> > Opening proxy : localhost:35711
> > 2016-01-19 20:04:16,411 INFO [ContainerLauncher #0]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> Shuffle
> > port returned by ContainerManager for
> attempt_1453244277886_0001_m_000000_0
> > : 13562
> > 2016-01-19 20:04:16,413 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt:
> > [attempt_1453244277886_0001_m_000000_0] using containerId:
> > [container_1453244277886_0001_01_000002 on NM: [localhost:35711]
> > 2016-01-19 20:04:16,418 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_0 TaskAttempt Transitioned from
> ASSIGNED
> > to RUNNING
> > 2016-01-19 20:04:16,419 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.speculate.DefaultSpeculator:
> > ATTEMPT_START task_1453244277886_0001_m_000000
> > 2016-01-19 20:04:16,419 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl:
> > task_1453244277886_0001_m_000000 Task Transitioned from SCHEDULED to
> RUNNING
> > 2016-01-19 20:04:17,251 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
> getResources()
> > for application_1453244277886_0001: ask=3 release= 0 newContainers=0
> > finishedContainers=0 resourcelimit=<memory:768, vCores:-1> knownNMs=1
> > 2016-01-19 20:04:18,269 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received
> > completed container container_1453244277886_0001_01_000002
> > 2016-01-19 20:04:18,269 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating
> > schedule, headroom=768
> > 2016-01-19 20:04:18,269 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
> > start threshold not met. completedMapsForReduceSlowstart 1
> > 2016-01-19 20:04:18,270 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0
> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:1 ContRel:0
> > HostLocal:1 RackLocal:0
> > 2016-01-19 20:04:18,280 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_0 TaskAttempt Transitioned from
> RUNNING
> > to FAIL_CONTAINER_CLEANUP
> > 2016-01-19 20:04:18,280 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics
> > report from attempt_1453244277886_0001_m_000000_0: Exception from
> > container-launch: org.apache.hadoop.util.Shell$ExitCodeException:
> > org.apache.hadoop.util.Shell$ExitCodeException:
> >     at org.apache.hadoop.util.Shell.runCommand(Shell.java:505)
> >     at org.apache.hadoop.util.Shell.run(Shell.java:418)
> >     at
> > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:650)
> >     at
> >
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195)
> >     at
> >
> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:300)
> >     at
> >
> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:81)
> >     at java.util.concurrent.FutureTask.run(FutureTask.java:262)
> >     at
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> >     at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> >     at java.lang.Thread.run(Thread.java:745)
> >
> >
> > Container exited with a non-zero exit code 1
> >
> > 2016-01-19 20:04:18,281 INFO [ContainerLauncher #1]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> > Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container
> > container_1453244277886_0001_01_000002 taskAttempt
> > attempt_1453244277886_0001_m_000000_0
> > 2016-01-19 20:04:18,281 INFO [ContainerLauncher #1]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> KILLING
> > attempt_1453244277886_0001_m_000000_0
> > 2016-01-19 20:04:18,299 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_0 TaskAttempt Transitioned from
> > FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
> > 2016-01-19 20:04:18,300 INFO [CommitterEvent Processor #1]
> > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler:
> Processing
> > the event EventType: TASK_ABORT
> > 2016-01-19 20:04:18,313 WARN [CommitterEvent Processor #1]
> > org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not
> delete
> >
> hdfs://hdnode01:54310/user/hduser/output/wordcount/_temporary/1/_temporary/attempt_1453244277886_0001_m_000000_0
> > 2016-01-19 20:04:18,317 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_0 TaskAttempt Transitioned from
> > FAIL_TASK_CLEANUP to FAILED
> > 2016-01-19 20:04:18,326 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.yarn.util.RackResolver: Resolved localhost to
> > /default-rack
> > 2016-01-19 20:04:18,327 INFO [Thread-50]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 1 failures on
> > node localhost
> > 2016-01-19 20:04:18,329 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_1 TaskAttempt Transitioned from NEW
> to
> > UNASSIGNED
> > 2016-01-19 20:04:18,329 INFO [Thread-50]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added
> > attempt_1453244277886_0001_m_000000_1 to list of failed maps
> > 2016-01-19 20:04:19,270 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before
> > Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:0
> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:1 ContRel:0
> > HostLocal:1 RackLocal:0
> > 2016-01-19 20:04:19,277 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
> getResources()
> > for application_1453244277886_0001: ask=1 release= 0 newContainers=0
> > finishedContainers=0 resourcelimit=<memory:768, vCores:-1> knownNMs=1
> > 2016-01-19 20:04:19,277 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating
> > schedule, headroom=768
> > 2016-01-19 20:04:19,278 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
> > start threshold not met. completedMapsForReduceSlowstart 1
> > 2016-01-19 20:04:20,285 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated
> > containers 1
> > 2016-01-19 20:04:20,285 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning
> > container Container: [ContainerId:
> container_1453244277886_0001_01_000003,
> > NodeId: localhost:35711, NodeHttpAddress: localhost:8042, Resource:
> > <memory:512, vCores:1>, Priority: 5, Token: Token { kind: ContainerToken,
> > service: 127.0.0.1:35711 }, ] to fast fail map
> > 2016-01-19 20:04:20,286 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned from
> > earlierFailedMaps
> > 2016-01-19 20:04:20,286 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned
> > container container_1453244277886_0001_01_000003 to
> > attempt_1453244277886_0001_m_000000_1
> > 2016-01-19 20:04:20,287 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating
> > schedule, headroom=768
> > 2016-01-19 20:04:20,287 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
> > start threshold not met. completedMapsForReduceSlowstart 1
> > 2016-01-19 20:04:20,287 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1
> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:2 ContRel:0
> > HostLocal:1 RackLocal:0
> > 2016-01-19 20:04:20,287 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.yarn.util.RackResolver: Resolved localhost to
> > /default-rack
> > 2016-01-19 20:04:20,289 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_1 TaskAttempt Transitioned from
> > UNASSIGNED to ASSIGNED
> > 2016-01-19 20:04:20,292 INFO [ContainerLauncher #2]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> > Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container
> > container_1453244277886_0001_01_000003 taskAttempt
> > attempt_1453244277886_0001_m_000000_1
> > 2016-01-19 20:04:20,292 INFO [ContainerLauncher #2]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> Launching
> > attempt_1453244277886_0001_m_000000_1
> > 2016-01-19 20:04:20,301 INFO [ContainerLauncher #2]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> Shuffle
> > port returned by ContainerManager for
> attempt_1453244277886_0001_m_000000_1
> > : 13562
> > 2016-01-19 20:04:20,302 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt:
> > [attempt_1453244277886_0001_m_000000_1] using containerId:
> > [container_1453244277886_0001_01_000003 on NM: [localhost:35711]
> > 2016-01-19 20:04:20,303 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_1 TaskAttempt Transitioned from
> ASSIGNED
> > to RUNNING
> > 2016-01-19 20:04:20,304 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.speculate.DefaultSpeculator:
> > ATTEMPT_START task_1453244277886_0001_m_000000
> > 2016-01-19 20:04:21,295 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
> getResources()
> > for application_1453244277886_0001: ask=1 release= 0 newContainers=0
> > finishedContainers=1 resourcelimit=<memory:768, vCores:-1> knownNMs=1
> > 2016-01-19 20:04:21,296 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received
> > completed container container_1453244277886_0001_01_000003
> > 2016-01-19 20:04:21,296 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating
> > schedule, headroom=768
> > 2016-01-19 20:04:21,296 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
> > start threshold not met. completedMapsForReduceSlowstart 1
> > 2016-01-19 20:04:21,297 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0
> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:2 ContRel:0
> > HostLocal:1 RackLocal:0
> > 2016-01-19 20:04:21,297 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_1 TaskAttempt Transitioned from
> RUNNING
> > to FAIL_CONTAINER_CLEANUP
> > 2016-01-19 20:04:21,298 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics
> > report from attempt_1453244277886_0001_m_000000_1: Exception from
> > container-launch: org.apache.hadoop.util.Shell$ExitCodeException:
> > org.apache.hadoop.util.Shell$ExitCodeException:
> >     at org.apache.hadoop.util.Shell.runCommand(Shell.java:505)
> >     at org.apache.hadoop.util.Shell.run(Shell.java:418)
> >     at
> > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:650)
> >     at
> >
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195)
> >     at
> >
> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:300)
> >     at
> >
> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:81)
> >     at java.util.concurrent.FutureTask.run(FutureTask.java:262)
> >     at
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> >     at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> >     at java.lang.Thread.run(Thread.java:745)
> >
> >
> > Container exited with a non-zero exit code 1
> >
> > 2016-01-19 20:04:21,300 INFO [ContainerLauncher #3]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> > Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container
> > container_1453244277886_0001_01_000003 taskAttempt
> > attempt_1453244277886_0001_m_000000_1
> > 2016-01-19 20:04:21,301 INFO [ContainerLauncher #3]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> KILLING
> > attempt_1453244277886_0001_m_000000_1
> > 2016-01-19 20:04:21,307 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_1 TaskAttempt Transitioned from
> > FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
> > 2016-01-19 20:04:21,308 INFO [CommitterEvent Processor #2]
> > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler:
> Processing
> > the event EventType: TASK_ABORT
> > 2016-01-19 20:04:21,312 WARN [CommitterEvent Processor #2]
> > org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not
> delete
> >
> hdfs://hdnode01:54310/user/hduser/output/wordcount/_temporary/1/_temporary/attempt_1453244277886_0001_m_000000_1
> > 2016-01-19 20:04:21,312 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_1 TaskAttempt Transitioned from
> > FAIL_TASK_CLEANUP to FAILED
> > 2016-01-19 20:04:21,313 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.yarn.util.RackResolver: Resolved localhost to
> > /default-rack
> > 2016-01-19 20:04:21,313 INFO [Thread-50]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 2 failures on
> > node localhost
> > 2016-01-19 20:04:21,313 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_2 TaskAttempt Transitioned from NEW
> to
> > UNASSIGNED
> > 2016-01-19 20:04:21,314 INFO [Thread-50]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added
> > attempt_1453244277886_0001_m_000000_2 to list of failed maps
> > 2016-01-19 20:04:22,297 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before
> > Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:0
> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:2 ContRel:0
> > HostLocal:1 RackLocal:0
> > 2016-01-19 20:04:22,304 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
> getResources()
> > for application_1453244277886_0001: ask=1 release= 0 newContainers=0
> > finishedContainers=0 resourcelimit=<memory:768, vCores:-1> knownNMs=1
> > 2016-01-19 20:04:22,305 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating
> > schedule, headroom=768
> > 2016-01-19 20:04:22,305 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
> > start threshold not met. completedMapsForReduceSlowstart 1
> > 2016-01-19 20:04:23,316 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated
> > containers 1
> > 2016-01-19 20:04:23,316 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning
> > container Container: [ContainerId:
> container_1453244277886_0001_01_000004,
> > NodeId: localhost:35711, NodeHttpAddress: localhost:8042, Resource:
> > <memory:512, vCores:1>, Priority: 5, Token: Token { kind: ContainerToken,
> > service: 127.0.0.1:35711 }, ] to fast fail map
> > 2016-01-19 20:04:23,317 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned from
> > earlierFailedMaps
> > 2016-01-19 20:04:23,317 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned
> > container container_1453244277886_0001_01_000004 to
> > attempt_1453244277886_0001_m_000000_2
> > 2016-01-19 20:04:23,318 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating
> > schedule, headroom=768
> > 2016-01-19 20:04:23,318 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
> > start threshold not met. completedMapsForReduceSlowstart 1
> > 2016-01-19 20:04:23,318 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1
> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:3 ContRel:0
> > HostLocal:1 RackLocal:0
> > 2016-01-19 20:04:23,318 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.yarn.util.RackResolver: Resolved localhost to
> > /default-rack
> > 2016-01-19 20:04:23,320 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_2 TaskAttempt Transitioned from
> > UNASSIGNED to ASSIGNED
> > 2016-01-19 20:04:23,323 INFO [ContainerLauncher #4]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> > Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container
> > container_1453244277886_0001_01_000004 taskAttempt
> > attempt_1453244277886_0001_m_000000_2
> > 2016-01-19 20:04:23,323 INFO [ContainerLauncher #4]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> Launching
> > attempt_1453244277886_0001_m_000000_2
> > 2016-01-19 20:04:23,334 INFO [ContainerLauncher #4]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> Shuffle
> > port returned by ContainerManager for
> attempt_1453244277886_0001_m_000000_2
> > : 13562
> > 2016-01-19 20:04:23,335 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt:
> > [attempt_1453244277886_0001_m_000000_2] using containerId:
> > [container_1453244277886_0001_01_000004 on NM: [localhost:35711]
> > 2016-01-19 20:04:23,335 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_2 TaskAttempt Transitioned from
> ASSIGNED
> > to RUNNING
> > 2016-01-19 20:04:23,335 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.speculate.DefaultSpeculator:
> > ATTEMPT_START task_1453244277886_0001_m_000000
> > 2016-01-19 20:04:24,326 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
> getResources()
> > for application_1453244277886_0001: ask=1 release= 0 newContainers=0
> > finishedContainers=1 resourcelimit=<memory:768, vCores:-1> knownNMs=1
> > 2016-01-19 20:04:24,327 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received
> > completed container container_1453244277886_0001_01_000004
> > 2016-01-19 20:04:24,327 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating
> > schedule, headroom=768
> > 2016-01-19 20:04:24,327 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
> > start threshold not met. completedMapsForReduceSlowstart 1
> > 2016-01-19 20:04:24,328 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0
> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:3 ContRel:0
> > HostLocal:1 RackLocal:0
> > 2016-01-19 20:04:24,328 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_2 TaskAttempt Transitioned from
> RUNNING
> > to FAIL_CONTAINER_CLEANUP
> > 2016-01-19 20:04:24,328 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics
> > report from attempt_1453244277886_0001_m_000000_2: Exception from
> > container-launch: org.apache.hadoop.util.Shell$ExitCodeException:
> > org.apache.hadoop.util.Shell$ExitCodeException:
> >     at org.apache.hadoop.util.Shell.runCommand(Shell.java:505)
> >     at org.apache.hadoop.util.Shell.run(Shell.java:418)
> >     at
> > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:650)
> >     at
> >
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195)
> >     at
> >
> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:300)
> >     at
> >
> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:81)
> >     at java.util.concurrent.FutureTask.run(FutureTask.java:262)
> >     at
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> >     at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> >     at java.lang.Thread.run(Thread.java:745)
> >
> >
> > Container exited with a non-zero exit code 1
> >
> > 2016-01-19 20:04:24,331 INFO [ContainerLauncher #5]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> > Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container
> > container_1453244277886_0001_01_000004 taskAttempt
> > attempt_1453244277886_0001_m_000000_2
> > 2016-01-19 20:04:24,331 INFO [ContainerLauncher #5]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> KILLING
> > attempt_1453244277886_0001_m_000000_2
> > 2016-01-19 20:04:24,335 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_2 TaskAttempt Transitioned from
> > FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
> > 2016-01-19 20:04:24,336 INFO [CommitterEvent Processor #3]
> > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler:
> Processing
> > the event EventType: TASK_ABORT
> > 2016-01-19 20:04:24,340 WARN [CommitterEvent Processor #3]
> > org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not
> delete
> >
> hdfs://hdnode01:54310/user/hduser/output/wordcount/_temporary/1/_temporary/attempt_1453244277886_0001_m_000000_2
> > 2016-01-19 20:04:24,341 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_2 TaskAttempt Transitioned from
> > FAIL_TASK_CLEANUP to FAILED
> > 2016-01-19 20:04:24,341 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.yarn.util.RackResolver: Resolved localhost to
> > /default-rack
> > 2016-01-19 20:04:24,342 INFO [Thread-50]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 3 failures on
> > node localhost
> > 2016-01-19 20:04:24,342 INFO [Thread-50]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: Blacklisted
> host
> > localhost
> > 2016-01-19 20:04:24,342 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_3 TaskAttempt Transitioned from NEW
> to
> > UNASSIGNED
> > 2016-01-19 20:04:24,343 INFO [Thread-50]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added
> > attempt_1453244277886_0001_m_000000_3 to list of failed maps
> > 2016-01-19 20:04:25,328 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before
> > Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:0
> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:3 ContRel:0
> > HostLocal:1 RackLocal:0
> > 2016-01-19 20:04:25,336 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
> getResources()
> > for application_1453244277886_0001: ask=1 release= 0 newContainers=0
> > finishedContainers=0 resourcelimit=<memory:768, vCores:-1> knownNMs=1
> > 2016-01-19 20:04:25,336 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: Update the
> > blacklist for application_1453244277886_0001: blacklistAdditions=1
> > blacklistRemovals=0
> > 2016-01-19 20:04:25,337 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: Ignore
> > blacklisting set to true. Known: 1, Blacklisted: 1, 100%
> > 2016-01-19 20:04:25,337 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating
> > schedule, headroom=768
> > 2016-01-19 20:04:25,337 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
> > start threshold not met. completedMapsForReduceSlowstart 1
> > 2016-01-19 20:04:26,342 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: Update the
> > blacklist for application_1453244277886_0001: blacklistAdditions=0
> > blacklistRemovals=1
> > 2016-01-19 20:04:26,342 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating
> > schedule, headroom=768
> > 2016-01-19 20:04:26,342 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
> > start threshold not met. completedMapsForReduceSlowstart 1
> > 2016-01-19 20:04:27,351 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated
> > containers 1
> > 2016-01-19 20:04:27,352 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning
> > container Container: [ContainerId:
> container_1453244277886_0001_01_000005,
> > NodeId: localhost:35711, NodeHttpAddress: localhost:8042, Resource:
> > <memory:512, vCores:1>, Priority: 5, Token: Token { kind: ContainerToken,
> > service: 127.0.0.1:35711 }, ] to fast fail map
> > 2016-01-19 20:04:27,352 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned from
> > earlierFailedMaps
> > 2016-01-19 20:04:27,352 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned
> > container container_1453244277886_0001_01_000005 to
> > attempt_1453244277886_0001_m_000000_3
> > 2016-01-19 20:04:27,353 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating
> > schedule, headroom=768
> > 2016-01-19 20:04:27,353 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
> > start threshold not met. completedMapsForReduceSlowstart 1
> > 2016-01-19 20:04:27,353 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.yarn.util.RackResolver: Resolved localhost to
> > /default-rack
> > 2016-01-19 20:04:27,353 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1
> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:4 ContRel:0
> > HostLocal:1 RackLocal:0
> > 2016-01-19 20:04:27,354 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_3 TaskAttempt Transitioned from
> > UNASSIGNED to ASSIGNED
> > 2016-01-19 20:04:27,355 INFO [ContainerLauncher #6]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> > Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container
> > container_1453244277886_0001_01_000005 taskAttempt
> > attempt_1453244277886_0001_m_000000_3
> > 2016-01-19 20:04:27,355 INFO [ContainerLauncher #6]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> Launching
> > attempt_1453244277886_0001_m_000000_3
> > 2016-01-19 20:04:27,365 INFO [ContainerLauncher #6]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> Shuffle
> > port returned by ContainerManager for
> attempt_1453244277886_0001_m_000000_3
> > : 13562
> > 2016-01-19 20:04:27,365 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt:
> > [attempt_1453244277886_0001_m_000000_3] using containerId:
> > [container_1453244277886_0001_01_000005 on NM: [localhost:35711]
> > 2016-01-19 20:04:27,366 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_3 TaskAttempt Transitioned from
> ASSIGNED
> > to RUNNING
> > 2016-01-19 20:04:27,366 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.speculate.DefaultSpeculator:
> > ATTEMPT_START task_1453244277886_0001_m_000000
> > 2016-01-19 20:04:28,361 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor:
> getResources()
> > for application_1453244277886_0001: ask=1 release= 0 newContainers=0
> > finishedContainers=1 resourcelimit=<memory:768, vCores:-1> knownNMs=1
> > 2016-01-19 20:04:28,361 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received
> > completed container container_1453244277886_0001_01_000005
> > 2016-01-19 20:04:28,361 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating
> > schedule, headroom=768
> > 2016-01-19 20:04:28,361 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
> > start threshold not met. completedMapsForReduceSlowstart 1
> > 2016-01-19 20:04:28,362 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0
> > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:4 ContRel:0
> > HostLocal:1 RackLocal:0
> > 2016-01-19 20:04:28,362 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_3 TaskAttempt Transitioned from
> RUNNING
> > to FAIL_CONTAINER_CLEANUP
> > 2016-01-19 20:04:28,362 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics
> > report from attempt_1453244277886_0001_m_000000_3: Exception from
> > container-launch: org.apache.hadoop.util.Shell$ExitCodeException:
> > org.apache.hadoop.util.Shell$ExitCodeException:
> >     at org.apache.hadoop.util.Shell.runCommand(Shell.java:505)
> >     at org.apache.hadoop.util.Shell.run(Shell.java:418)
> >     at
> > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:650)
> >     at
> >
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195)
> >     at
> >
> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:300)
> >     at
> >
> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:81)
> >     at java.util.concurrent.FutureTask.run(FutureTask.java:262)
> >     at
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> >     at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> >     at java.lang.Thread.run(Thread.java:745)
> >
> >
> > Container exited with a non-zero exit code 1
> >
> > 2016-01-19 20:04:28,364 INFO [ContainerLauncher #7]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> > Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container
> > container_1453244277886_0001_01_000005 taskAttempt
> > attempt_1453244277886_0001_m_000000_3
> > 2016-01-19 20:04:28,365 INFO [ContainerLauncher #7]
> > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> KILLING
> > attempt_1453244277886_0001_m_000000_3
> > 2016-01-19 20:04:28,373 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_3 TaskAttempt Transitioned from
> > FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
> > 2016-01-19 20:04:28,374 INFO [CommitterEvent Processor #4]
> > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler:
> Processing
> > the event EventType: TASK_ABORT
> > 2016-01-19 20:04:28,377 WARN [CommitterEvent Processor #4]
> > org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not
> delete
> >
> hdfs://hdnode01:54310/user/hduser/output/wordcount/_temporary/1/_temporary/attempt_1453244277886_0001_m_000000_3
> > 2016-01-19 20:04:28,378 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_m_000000_3 TaskAttempt Transitioned from
> > FAIL_TASK_CLEANUP to FAILED
> > 2016-01-19 20:04:28,380 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl:
> > task_1453244277886_0001_m_000000 Task Transitioned from RUNNING to FAILED
> > 2016-01-19 20:04:28,381 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Num completed
> Tasks: 1
> > 2016-01-19 20:04:28,381 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Job failed as tasks
> > failed. failedMaps:1 failedReduces:0
> > 2016-01-19 20:04:28,382 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl:
> > job_1453244277886_0001Job Transitioned from RUNNING to FAIL_WAIT
> > 2016-01-19 20:04:28,383 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl:
> > task_1453244277886_0001_r_000000 Task Transitioned from SCHEDULED to
> > KILL_WAIT
> > 2016-01-19 20:04:28,383 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> > attempt_1453244277886_0001_r_000000_0 TaskAttempt Transitioned from
> > UNASSIGNED to KILLED
> > 2016-01-19 20:04:28,383 INFO [Thread-50]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Processing
> the
> > event EventType: CONTAINER_DEALLOCATE
> > 2016-01-19 20:04:28,383 ERROR [Thread-50]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Could not
> > deallocate container for task attemptId
> > attempt_1453244277886_0001_r_000000_0
> > 2016-01-19 20:04:28,383 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl:
> > task_1453244277886_0001_r_000000 Task Transitioned from KILL_WAIT to
> KILLED
> > 2016-01-19 20:04:28,384 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl:
> > job_1453244277886_0001Job Transitioned from FAIL_WAIT to FAIL_ABORT
> > 2016-01-19 20:04:28,390 INFO [CommitterEvent Processor #0]
> > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler:
> Processing
> > the event EventType: JOB_ABORT
> > 2016-01-19 20:04:28,432 INFO [AsyncDispatcher event handler]
> > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl:
> > job_1453244277886_0001Job Transitioned from FAIL_ABORT to FAILED
> > 2016-01-19 20:04:28,433 INFO [Thread-61]
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: We are finishing cleanly
> so
> > this is the last retry
> > 2016-01-19 20:04:28,433 INFO [Thread-61]
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Notify RMCommunicator
> > isAMLastRetry: true
> > 2016-01-19 20:04:28,433 INFO [Thread-61]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator:
> RMCommunicator
> > notified that shouldUnregistered is: true
> > 2016-01-19 20:04:28,433 INFO [Thread-61]
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Notify JHEH
> isAMLastRetry:
> > true
> > 2016-01-19 20:04:28,433 INFO [Thread-61]
> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler:
> > JobHistoryEventHandler notified that forceJobCompletion is true
> > 2016-01-19 20:04:28,434 INFO [Thread-61]
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Calling stop for all the
> > services
> > 2016-01-19 20:04:28,435 INFO [Thread-61]
> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Stopping
> > JobHistoryEventHandler. Size of the outstanding queue size is 0
> > 2016-01-19 20:04:29,362 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before
> > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0
> > AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:4 ContRel:0
> > HostLocal:1 RackLocal:0
> > 2016-01-19 20:04:29,362 INFO [eventHandlingThread]
> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copying
> >
> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/hduser/.staging/job_1453244277886_0001/job_1453244277886_0001_1.jhist
> > to
> >
> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001-1453244648033-hduser-word+count-1453244668381-0-0-FAILED-default-1453244654166.jhist_tmp
> > 2016-01-19 20:04:29,366 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating
> > schedule, headroom=768
> > 2016-01-19 20:04:29,366 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
> > start threshold reached. Scheduling reduces.
> > 2016-01-19 20:04:29,366 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: All maps
> > assigned. Ramping up all remaining reduces:1
> > 2016-01-19 20:04:29,366 INFO [RMCommunicator Allocator]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After
> > Scheduling: PendingReds:0 ScheduledMaps:0 ScheduledReds:1 AssignedMaps:0
> > AssignedReds:0 CompletedMaps:1 CompletedReds:0 ContAlloc:4 ContRel:0
> > HostLocal:1 RackLocal:0
> > 2016-01-19 20:04:29,544 INFO [eventHandlingThread]
> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copied to
> > done location:
> >
> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001-1453244648033-hduser-word+count-1453244668381-0-0-FAILED-default-1453244654166.jhist_tmp
> > 2016-01-19 20:04:29,598 INFO [eventHandlingThread]
> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copying
> >
> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/hduser/.staging/job_1453244277886_0001/job_1453244277886_0001_1_conf.xml
> > to
> >
> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001_conf.xml_tmp
> > 2016-01-19 20:04:29,801 INFO [eventHandlingThread]
> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copied to
> > done location:
> >
> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001_conf.xml_tmp
> > 2016-01-19 20:04:29,907 INFO [eventHandlingThread]
> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Moved tmp
> to
> > done:
> >
> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001.summary_tmp
> > to
> >
> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001.summary
> > 2016-01-19 20:04:30,008 INFO [eventHandlingThread]
> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Moved tmp
> to
> > done:
> >
> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001_conf.xml_tmp
> > to
> >
> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001_conf.xml
> > 2016-01-19 20:04:30,070 INFO [eventHandlingThread]
> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Moved tmp
> to
> > done:
> >
> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001-1453244648033-hduser-word+count-1453244668381-0-0-FAILED-default-1453244654166.jhist_tmp
> > to
> >
> hdfs://hdnode01:54310/tmp/hadoop-yarn/staging/history/done_intermediate/hduser/job_1453244277886_0001-1453244648033-hduser-word+count-1453244668381-0-0-FAILED-default-1453244654166.jhist
> > 2016-01-19 20:04:30,071 INFO [Thread-61]
> > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Stopped
> > JobHistoryEventHandler. super.stop()
> > 2016-01-19 20:04:30,078 INFO [Thread-61]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Setting job
> > diagnostics to Task failed task_1453244277886_0001_m_000000
> > Job failed as tasks failed. failedMaps:1 failedReduces:0
> >
> > 2016-01-19 20:04:30,080 INFO [Thread-61]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: History url
> is
> > http://localhost:19888/jobhistory/job/job_1453244277886_0001
> > 2016-01-19 20:04:30,094 INFO [Thread-61]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Waiting for
> > application to be successfully unregistered.
> > 2016-01-19 20:04:31,099 INFO [Thread-61]
> > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Final Stats:
> > PendingReds:0 ScheduledMaps:0 ScheduledReds:1 AssignedMaps:0
> AssignedReds:0
> > CompletedMaps:1 CompletedReds:0 ContAlloc:4 ContRel:0 HostLocal:1
> > RackLocal:0
> > 2016-01-19 20:04:31,104 INFO [Thread-61]
> > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Deleting staging
> directory
> > hdfs://hdnode01:54310
> > /tmp/hadoop-yarn/staging/hduser/.staging/job_1453244277886_0001
> > 2016-01-19 20:04:31,133 INFO [Thread-61] org.apache.hadoop.ipc.Server:
> > Stopping server on 45584
> > 2016-01-19 20:04:31,135 INFO [IPC Server listener on 45584]
> > org.apache.hadoop.ipc.Server: Stopping IPC Server listener on 45584
> > 2016-01-19 20:04:31,135 INFO [IPC Server Responder]
> > org.apache.hadoop.ipc.Server: Stopping IPC Server Responder
> > 2016-01-19 20:04:31,136 INFO [TaskHeartbeatHandler PingChecker]
> > org.apache.hadoop.mapreduce.v2.app.TaskHeartbeatHandler:
> > TaskHeartbeatHandler thread interrupted
> >
> >
> > Jps results, i believe that everything is ok, right?:
> > 21267 DataNode
> > 21609 ResourceManager
> > 21974 JobHistoryServer
> > 21735 NodeManager
> > 24546 Jps
> > 16532 org.eclipse.equinox.launcher_1.3.0.v20140415-2008.jar
> > 21121 NameNode
> > 22098 QuorumPeerMain
> > 21456 SecondaryNameNode
> >
> >
>

Mime
View raw message