Return-Path: X-Original-To: apmail-hadoop-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id F1A70C8C7 for ; Tue, 18 Nov 2014 16:02:21 +0000 (UTC) Received: (qmail 38417 invoked by uid 500); 18 Nov 2014 16:02:16 -0000 Delivered-To: apmail-hadoop-user-archive@hadoop.apache.org Received: (qmail 38288 invoked by uid 500); 18 Nov 2014 16:02:16 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 38277 invoked by uid 99); 18 Nov 2014 16:02:16 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 18 Nov 2014 16:02:16 +0000 X-ASF-Spam-Status: No, hits=-0.7 required=5.0 tests=RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of ozawa.tsuyoshi@gmail.com designates 209.85.212.178 as permitted sender) Received: from [209.85.212.178] (HELO mail-wi0-f178.google.com) (209.85.212.178) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 18 Nov 2014 16:01:49 +0000 Received: by mail-wi0-f178.google.com with SMTP id hi2so4655633wib.17 for ; Tue, 18 Nov 2014 08:01:49 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type; bh=bUOKcE/W4UPQSVCOu9PegLxIn44CizYl2tZMpZntLHA=; b=MGRj9HtQvdY6QYX00+OqRgfXYsPo+j6hGRAgojSu0IO5PhI7q1KGDUdqz5Uxq7rbYH /f2HleBe5/ZEWSUbt/eMbjPcbBNlLjC1UdXeK03D1uamR9l57U668tEq4dAWrK6uptSN +lViN/1yEiDKO30C1cmavdIqO54zS04Nq4SX+dlZShc9H5UaAx3oxBOujbZrxMcixBJO 5HscAb4cNQ9OnGxxzyFJ89Ow/ExIw0RPgpFVhEpzhzdtRB27gISFn7/N+3bBPHrQhEIR IzTYW5vxjzrOSOoZl0d5dGmG/qUH3AdV1FEEtqMo1hee5GczntmqFAoVFO0O1qd12E+m N/Wg== MIME-Version: 1.0 X-Received: by 10.180.187.234 with SMTP id fv10mr41788396wic.25.1416326508924; Tue, 18 Nov 2014 08:01:48 -0800 (PST) Received: by 10.216.154.6 with HTTP; Tue, 18 Nov 2014 08:01:48 -0800 (PST) In-Reply-To: References: Date: Wed, 19 Nov 2014 01:01:48 +0900 Message-ID: Subject: Re: MR job fails with too many mappers From: Tsuyoshi OZAWA To: "user@hadoop.apache.org" Content-Type: text/plain; charset=UTF-8 X-Virus-Checked: Checked by ClamAV on apache.org Hi, Could you share following configurations? It can be failures because of out of memory at mapper side. yarn.app.mapreduce.am.resource.mb mapreduce.map.memory.mb mapreduce.reduce.memory.mb mapreduce.map.java.opts mapreduce.reduce.java.opts On Wed, Nov 19, 2014 at 12:23 AM, francexo83 wrote: > Hi All, > > I have a small hadoop cluster with three nodes and HBase 0.98.1 installed > on it. > > The hadoop version is 2.3.0 and below my use case scenario. > > I wrote a map reduce program that reads data from an hbase table and does > some transformations on these data. > Jobs are very simple so they didn't need the reduce phase. I also wrote a > TableInputFormat extension in order to maximize the number of concurrent > maps on the cluster. > In other words, each row should be processed by a single map task. > > Everything goes well until the number of rows and consequently mappers > exceeds 300000 quota. > > This is the only exception I see when the job fails: > > Application application_1416304409718_0032 failed 2 times due to AM > Container for appattempt_1416304409718_0032_000002 exited with exitCode: 1 > due to: > > > Exception from container-launch: > org.apache.hadoop.util.Shell$ExitCodeException: > org.apache.hadoop.util.Shell$ExitCodeException: > at org.apache.hadoop.util.Shell.runCommand(Shell.java:511) > at org.apache.hadoop.util.Shell.run(Shell.java:424) > at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:656) > at > org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195) > at > org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:300) > at > org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:81) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > Container exited with a non-zero exit code 1 > > > Cluster configuration details: > Node1: 12 GB, 4 core > Node2: 6 GB, 4 core > Node3: 6 GB, 4 core > > yarn.scheduler.minimum-allocation-mb=2048 > yarn.scheduler.maximum-allocation-mb=4096 > yarn.nodemanager.resource.memory-mb=6144 > > > > Regards -- - Tsuyoshi