Return-Path: Delivered-To: apmail-hadoop-core-dev-archive@www.apache.org Received: (qmail 64335 invoked from network); 17 Jul 2008 11:04:26 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (140.211.11.2) by minotaur.apache.org with SMTP; 17 Jul 2008 11:04:26 -0000 Received: (qmail 69170 invoked by uid 500); 17 Jul 2008 11:04:23 -0000 Delivered-To: apmail-hadoop-core-dev-archive@hadoop.apache.org Received: (qmail 69145 invoked by uid 500); 17 Jul 2008 11:04:23 -0000 Mailing-List: contact core-dev-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: core-dev@hadoop.apache.org Delivered-To: mailing list core-dev@hadoop.apache.org Received: (qmail 69134 invoked by uid 99); 17 Jul 2008 11:04:23 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 17 Jul 2008 04:04:23 -0700 X-ASF-Spam-Status: No, hits=-2000.0 required=10.0 tests=ALL_TRUSTED X-Spam-Check-By: apache.org Received: from [140.211.11.140] (HELO brutus.apache.org) (140.211.11.140) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 17 Jul 2008 11:03:38 +0000 Received: from brutus (localhost [127.0.0.1]) by brutus.apache.org (Postfix) with ESMTP id 20918234C177 for ; Thu, 17 Jul 2008 04:03:32 -0700 (PDT) Message-ID: <843780437.1216292612132.JavaMail.jira@brutus> Date: Thu, 17 Jul 2008 04:03:32 -0700 (PDT) From: "Hemanth Yamijala (JIRA)" To: core-dev@hadoop.apache.org Subject: [jira] Commented: (HADOOP-3581) Prevent memory intensive user tasks from taking down nodes In-Reply-To: <1009886343.1213704226494.JavaMail.jira@brutus> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit X-Virus-Checked: Checked by ClamAV on apache.org [ https://issues.apache.org/jira/browse/HADOOP-3581?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12614295#action_12614295 ] Hemanth Yamijala commented on HADOOP-3581: ------------------------------------------ bq. A user should specify the MAX RAM in GB or MB that the tasks will use. +1. I think that is much easier for a user to specify. Here's what I propose with respect to the configuration variables: - mapred.tasktracker.tasks.maxmemory: Cumulative memory that can be used by all map/reduce tasks. - mapred.map.task.maxmemory: (Overridable per job) Maximum memory any map task of a job can take. By default, mapred.tasktracker.tasks.maxmemory / number of slots on a node - mapred.reduce.task.maxmemory: (Overridable per job) Maximum memory any reduce of a job can take. By default, mapred.tasktracker.tasks.maxmemory / number of slots on a node Thoughts ? Specifically, on the default values, is it OK to give the same amount of max memory to map tasks and reduce tasks ? Or should we look to divide the max memory so that there's more (say twice) given to the reduce tasks, than to the map tasks ? > Prevent memory intensive user tasks from taking down nodes > ---------------------------------------------------------- > > Key: HADOOP-3581 > URL: https://issues.apache.org/jira/browse/HADOOP-3581 > Project: Hadoop Core > Issue Type: Improvement > Components: mapred > Reporter: Hemanth Yamijala > Assignee: Vinod Kumar Vavilapalli > Attachments: patch_3581_0.1.txt > > > Sometimes user Map/Reduce applications can get extremely memory intensive, maybe due to some inadvertent bugs in the user code, or the amount of data processed. When this happens, the user tasks start to interfere with the proper execution of other processes on the node, including other Hadoop daemons like the DataNode and TaskTracker. Thus, the node would become unusable for any Hadoop tasks. There should be a way to prevent such tasks from bringing down the node. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.