hadoop-common-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Vinod Kumar Vavilapalli (JIRA)" <j...@apache.org>
Subject [jira] Commented: (HADOOP-3581) Prevent memory intensive user tasks from taking down nodes
Date Thu, 17 Jul 2008 05:57:32 GMT

    [ https://issues.apache.org/jira/browse/HADOOP-3581?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12614231#action_12614231

Vinod Kumar Vavilapalli commented on HADOOP-3581:

Summarizing all the discussion that went so far ..

 - TaskTracker tracks the memory usage of all the tasks and their sub-processes (irrespective
of which user runs which tasks).
 - It uses per-task-objects of classes implementing ProcessTree as described earlier. Currently,
we implement ProcfsBasedProcessTree, which works on both Linux and Cygwin. For other OS' we
need classes extending ProcessTree.
 - We will have two configuration properties - per-tracker property mapred.tasktracker.tasks.maxmemory
specifying maximum memory usable across all tasks on a tasktracker; and per-job property mapred.map.memlimit.percent
for specifying memory usable across all maps in terms of percentage of mapred.tasktracker.tasks.maxmemory.
Maximum memory usable by reduce tasks = (100-mapred.map.memlimit.percent )% of mapred.tasktracker.tasks.maxmemory.
All these are virtual memory limits, not working set. By default, we can set mapred.tasktracker.tasks.maxmemory
to 12GB(4GB RAM + 8GB swap) and mapred.map.memlimit.percent to 33%. Should be ok?
 - After every heartbeat, TT scans through the list of running tasks and finds if any task
trees are transgressing limits and kills them by issuing 'kill <pid>' (SIGTERM) to each
process in each of the concerned process-trees. TT will monitor if a process-tree is killed
successfully. If not, it issues a subsequent SIGKILL.
 - ProcfsBasedProcessTree: Constructs process-tree information from /proc file system. TT
obtains pid from the task via the rpc method getPid and then constructs the process tree using

Please put forward your objections to the above proposal, if any.

> Prevent memory intensive user tasks from taking down nodes
> ----------------------------------------------------------
>                 Key: HADOOP-3581
>                 URL: https://issues.apache.org/jira/browse/HADOOP-3581
>             Project: Hadoop Core
>          Issue Type: Improvement
>          Components: mapred
>            Reporter: Hemanth Yamijala
>            Assignee: Vinod Kumar Vavilapalli
>         Attachments: patch_3581_0.1.txt
> Sometimes user Map/Reduce applications can get extremely memory intensive, maybe due
to some inadvertent bugs in the user code, or the amount of data processed. When this happens,
the user tasks start to interfere with the proper execution of other processes on the node,
including other Hadoop daemons like the DataNode and TaskTracker. Thus, the node would become
unusable for any Hadoop tasks. There should be a way to prevent such tasks from bringing down
the node.

This message is automatically generated by JIRA.
You can reply to this email to add a comment to the issue online.

View raw message