Return-Path: X-Original-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id B1EADE7BB for ; Mon, 18 Mar 2013 15:28:42 +0000 (UTC) Received: (qmail 72674 invoked by uid 500); 18 Mar 2013 15:28:37 -0000 Delivered-To: apmail-hadoop-hdfs-user-archive@hadoop.apache.org Received: (qmail 72517 invoked by uid 500); 18 Mar 2013 15:28:37 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 72507 invoked by uid 99); 18 Mar 2013 15:28:37 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 18 Mar 2013 15:28:37 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of dechouxb@gmail.com designates 209.85.217.176 as permitted sender) Received: from [209.85.217.176] (HELO mail-lb0-f176.google.com) (209.85.217.176) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 18 Mar 2013 15:28:31 +0000 Received: by mail-lb0-f176.google.com with SMTP id s4so4764677lbc.35 for ; Mon, 18 Mar 2013 08:28:11 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:x-received:in-reply-to:references:date:message-id :subject:from:to:content-type; bh=sij7EqIIaHqB+rxK3TsLHFNwTxtClh+pd0v1HLJ5qFs=; b=wvJhyxoJWKv18K80e09APrE2bocQaP/M5pzqPHxqrtKsShN31vq4Ixo903Pz1KKBtH TTyHGMKUE7IvNI7vCNfseBoYOBqELeFslIHb4vEZ5hql4mJKXmJKIuY7+48WrQy9ecFP CiUEvCd/Z5/iRxBKk24uLjkbfSOXV3nrhLM99LJ91Zmd4pwJPmnFe2aRiG5mwq9Moux3 tXEEd5QXDH4IR/Vat7e8nnFMu0fzeDFwglPQC4xqU6+heOgrFbY7XfmCv647Uhch+vN3 K8zouLE+PUUftVhkzKI/KqP4yENN+QRhdh1ZkACZ9N/Ub4fQBSsfBUfcW+86VScpUvQN opJA== MIME-Version: 1.0 X-Received: by 10.112.26.232 with SMTP id o8mr6646055lbg.51.1363620491206; Mon, 18 Mar 2013 08:28:11 -0700 (PDT) Received: by 10.112.135.131 with HTTP; Mon, 18 Mar 2013 08:28:11 -0700 (PDT) In-Reply-To: <51472A17.2090001@getjar.com> References: <51472A17.2090001@getjar.com> Date: Mon, 18 Mar 2013 16:28:11 +0100 Message-ID: Subject: Re: namenode directory failure question From: Bertrand Dechoux To: user@hadoop.apache.org Content-Type: multipart/alternative; boundary=bcaec554dbc8fcaa7104d834a282 X-Virus-Checked: Checked by ClamAV on apache.org --bcaec554dbc8fcaa7104d834a282 Content-Type: text/plain; charset=ISO-8859-1 You may want to check this JIRA: https://issues.apache.org/jira/browse/HADOOP-4885 It won't help you right know but it could allow you next time to avoid restarting. Regards Bertrand On Mon, Mar 18, 2013 at 3:52 PM, Brennon Church wrote: > Hello all, > > We have our dfs.name.dir configured to write to two local and one NFS > directories. The NFS server in question had to be restarted a couple days > back and that copy of the namenode data fell behind as a result. As I > understand it, restarting hadoop will take the most recent copy of the > namenode data, in this case one of the two local copies, and write that to > all three locations going forward. So that solves the problem. > > My question is this, is there a way to get the NFS copy of the data back > in sync without having to shut down and restart the namenode? I'd prefer to > not take an outage if I can help it. > > Thanks. > > --Brennon > > --bcaec554dbc8fcaa7104d834a282 Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable You may want to check this JIRA: https://issues.apache.org/jira/browse/HADOOP-4885
It won't help you right know but it could allow you next time to = avoid restarting.

Regards

Bertrand

On Mon, Mar 1= 8, 2013 at 3:52 PM, Brennon Church <brennon@getjar.com> wro= te:
Hello all,

We have our dfs.name.dir configured to write to two local and one NFS direc= tories. =A0The NFS server in question had to be restarted a couple days bac= k and that copy of the namenode data fell behind as a result. =A0As I under= stand it, restarting hadoop will take the most recent copy of the namenode = data, in this case one of the two local copies, and write that to all three= locations going forward. =A0So that solves the problem.

My question is this, is there a way to get the NFS copy of the data back in= sync without having to shut down and restart the namenode? I'd prefer = to not take an outage if I can help it.

Thanks.

--Brennon



--bcaec554dbc8fcaa7104d834a282--