hadoop-mapreduce-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Laxman (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (MAPREDUCE-6659) Mapreduce App master waits long to kill containers on lost nodes.
Date Thu, 24 Mar 2016 13:20:25 GMT

    [ https://issues.apache.org/jira/browse/MAPREDUCE-6659?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15210196#comment-15210196
] 

Laxman commented on MAPREDUCE-6659:
-----------------------------------

Please note that this issue happens with lost nodes (i.e, Unreachable hosts). NM crash with
a reachable host is exhibiting a totally a different expected retry behavior. There liveness
configurations are coming into play (yarn.resourcemanager.container.liveness-monitor.interval-ms,
yarn.nm.liveness-monitor.expiry-interval-ms, yarn.am.liveness-monitor.expiry-interval-ms)
as expected.

> Mapreduce App master waits long to kill containers on lost nodes.
> -----------------------------------------------------------------
>
>                 Key: MAPREDUCE-6659
>                 URL: https://issues.apache.org/jira/browse/MAPREDUCE-6659
>             Project: Hadoop Map/Reduce
>          Issue Type: Bug
>          Components: mr-am
>    Affects Versions: 2.6.0
>            Reporter: Laxman
>
> MR Application master waits for very long time to cleanup and relaunch the tasks on lost
nodes. Wait time is actually 2.5 hours (ipc.client.connect.max.retries * ipc.client.connect.max.retries.on.timeouts
* ipc.client.connect.timeout = 10 * 45 * 20 = 9000 seconds = 2.5 hours)
> Some similar issue related in RM-AM rpc protocol is fixed in YARN-3809.
> As fixed in YARN-3809, we may need to introduce new configurations to control this RPC
retry behavior.
> Also, I feel this total retry time should honor and capped maximum to global task time
out (mapreduce.task.timeout = 600000 default)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Mime
View raw message