Return-Path: X-Original-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id EC3F310C30 for ; Mon, 6 May 2013 09:01:20 +0000 (UTC) Received: (qmail 93330 invoked by uid 500); 6 May 2013 09:01:16 -0000 Delivered-To: apmail-hadoop-hdfs-user-archive@hadoop.apache.org Received: (qmail 93195 invoked by uid 500); 6 May 2013 09:01:15 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 93182 invoked by uid 99); 6 May 2013 09:01:15 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 06 May 2013 09:01:15 +0000 X-ASF-Spam-Status: No, hits=1.7 required=5.0 tests=FREEMAIL_ENVFROM_END_DIGIT,HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of nitinpawar432@gmail.com designates 209.85.217.173 as permitted sender) Received: from [209.85.217.173] (HELO mail-lb0-f173.google.com) (209.85.217.173) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 06 May 2013 09:01:10 +0000 Received: by mail-lb0-f173.google.com with SMTP id t10so3149955lbi.4 for ; Mon, 06 May 2013 02:00:49 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:x-received:in-reply-to:references:date:message-id :subject:from:to:content-type; bh=c02CwXrrNfxqCRp8nnBV/exuxB/JX5lZEWkOqiziDbM=; b=dStDGdYZgixq1exq3WVOwTjcN97dN8wwswAFuqHxKoZspJg/5S65dggFW7GWpHag2A 1DGWowx3nxWkkdM6pQ/q1P8pXZp8DY09ZAxwMh7O765bw7BwQSqkOsoahblzQ4hy3/tZ h/e/QnE8kpMgcXSlyx0fJ7pYULEMGwqlRdQiiocjka9zlE2Woba05HlmtQhnLX6erO1j /d2DAjiNftoty4wQ/xg4Hs/o99JOcBavKHmS3A5u6XQkeYIx/5vuf/TsnnW8N4eVbop2 zBHmT5oGKfEge08E2v1qD42t5ETqOrsYVWzSzu/kgoeHWisE+0DJrcAEz9JrMHFsoi00 UaIw== MIME-Version: 1.0 X-Received: by 10.152.26.225 with SMTP id o1mr7744351lag.43.1367830849156; Mon, 06 May 2013 02:00:49 -0700 (PDT) Received: by 10.114.96.196 with HTTP; Mon, 6 May 2013 02:00:49 -0700 (PDT) Received: by 10.114.96.196 with HTTP; Mon, 6 May 2013 02:00:49 -0700 (PDT) In-Reply-To: References: Date: Mon, 6 May 2013 14:30:49 +0530 Message-ID: Subject: Re: Namenode going to safe mode on YARN From: Nitin Pawar To: user@hadoop.apache.org Content-Type: multipart/alternative; boundary=089e0158c738e07dc804dc08efa4 X-Virus-Checked: Checked by ClamAV on apache.org --089e0158c738e07dc804dc08efa4 Content-Type: text/plain; charset=ISO-8859-1 What is your replication factor on hdfs? Did any of your datanode go down recently and is not back in rotation? Did you delete any hdfs blocks directly from datanodes? On May 6, 2013 2:28 PM, "Krishna Kishore Bonagiri" wrote: > Hi, > > I have been running application on my YARN cluster since around 20 days, > about 5000 applications a day. I am getting the following error today. > Please let me know how can I avoid this, is this happening because of a bug? > > org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.SafeModeException): > Cannot create file/1066/AppMaster.jar. Name node is in safe mode. > The reported blocks 4775 needs additional 880 blocks to reach the > threshold 0.9990 of total blocks 5660. Safe mode will be turned off > automatically. > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInternal(FSNamesystem.java:1786) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInt(FSNamesystem.java:1737) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFile(FSNamesystem.java:1719) > at > org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.create(NameNodeRpcServer.java:429) > at > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.create(ClientNamenodeProtocolServerSideTranslatorPB.java:271) > at > org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:40732) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:454) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1014) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1741) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1737) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1478) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1735) > > > Thanks, > Kishore > --089e0158c738e07dc804dc08efa4 Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable

What is your replication factor on hdfs?
Did any of your datanode go down recently and is not back in rotation?
Did you delete any hdfs blocks directly from datanodes?

On May 6, 2013 2:28 PM, "Krishna Kishore Bo= nagiri" <write2kishore@g= mail.com> wrote:
Hi,=A0

=A0 I have been running application on= my YARN cluster since around 20 days, about 5000 applications a day. I am = getting the following error today. Please let me know how can I avoid this,= is this happening because of a bug?

org.apache.hadoop.ipc.RemoteException(org.apache.h= adoop.hdfs.server.namenode.SafeModeException): Cannot create file/1066/AppM= aster.jar. Name node is in safe mode.
The reported blocks 4775 ne= eds additional 880 blocks to reach the threshold 0.9990 of total blocks 566= 0. Safe mode will be turned off automatically.
=A0 =A0 =A0 =A0 at org.apache.hadoop.hdfs.server.namenode.FSNamesystem= .startFileInternal(FSNamesystem.java:1786)
=A0 =A0 =A0 =A0 at org= .apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInt(FSNamesystem.= java:1737)
=A0 =A0 =A0 =A0 at org.apache.hadoop.hdfs.server.namenode.FSNamesystem= .startFile(FSNamesystem.java:1719)
=A0 =A0 =A0 =A0 at org.apache.= hadoop.hdfs.server.namenode.NameNodeRpcServer.create(NameNodeRpcServer.java= :429)
=A0 =A0 =A0 =A0 at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocol= ServerSideTranslatorPB.create(ClientNamenodeProtocolServerSideTranslatorPB.= java:271)
=A0 =A0 =A0 =A0 at org.apache.hadoop.hdfs.protocol.prot= o.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(= ClientNamenodeProtocolProtos.java:40732)
=A0 =A0 =A0 =A0 at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$Prot= oBufRpcInvoker.call(ProtobufRpcEngine.java:454)
=A0 =A0 =A0 =A0 a= t org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1014)
=A0 =A0 = =A0 =A0 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1741)
=A0 =A0 =A0 =A0 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.j= ava:1737)
=A0 =A0 =A0 =A0 at java.security.AccessController.doPri= vileged(Native Method)
=A0 =A0 =A0 =A0 at javax.security.auth.Sub= ject.doAs(Subject.java:415)
=A0 =A0 =A0 =A0 at org.apache.hadoop.security.UserGroupInformation.doA= s(UserGroupInformation.java:1478)
=A0 =A0 =A0 =A0 at org.apache.h= adoop.ipc.Server$Handler.run(Server.java:1735)

Thanks,
Kishore
--089e0158c738e07dc804dc08efa4--