Return-Path: X-Original-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id EC375E4CE for ; Tue, 22 Jan 2013 17:39:06 +0000 (UTC) Received: (qmail 6745 invoked by uid 500); 22 Jan 2013 17:39:02 -0000 Delivered-To: apmail-hadoop-hdfs-user-archive@hadoop.apache.org Received: (qmail 6419 invoked by uid 500); 22 Jan 2013 17:39:02 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 6408 invoked by uid 99); 22 Jan 2013 17:39:02 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 22 Jan 2013 17:39:02 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of harsh@cloudera.com designates 209.85.223.179 as permitted sender) Received: from [209.85.223.179] (HELO mail-ie0-f179.google.com) (209.85.223.179) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 22 Jan 2013 17:38:55 +0000 Received: by mail-ie0-f179.google.com with SMTP id k14so11979646iea.38 for ; Tue, 22 Jan 2013 09:38:35 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20120113; h=x-received:mime-version:in-reply-to:references:from:date:message-id :subject:to:content-type:x-gm-message-state; bh=0Dym8tYXl/6lXgkbCOlto4w+Z2vG5sRQv/8ckNupDLY=; b=TscNU+LNPzkcXD4ZPTGeP6cpXNe1caLSmF/KzYfVCGiZZdQR/BXES/kkO9l56whtwx cwGaFqxx/6W4G4jbPhpMbXsN1PFGEbftuccMmccX3VpcqAIXPSy/QFt9MvVBEEC+QvR2 K4amYCRyDItzeNXH2PvoLoGGW5ZMItB/QiQLYQEI08YOPe7RoW6PyT3++0pTm2A9h4b4 ouS55MhLP/AfVlJ3q/LUmInai33KEe+WITQ1tqDtpbbEVofAMjv06l+t4F4rFqO8KGsw UZTh0SQCIxFV2Ot19B/1i6xNmS7pJOGmFoXFDMKKxeZVfS/30pxZhde7BPKgiCawHpyH nM8g== X-Received: by 10.50.193.167 with SMTP id hp7mr12285284igc.18.1358876314747; Tue, 22 Jan 2013 09:38:34 -0800 (PST) MIME-Version: 1.0 Received: by 10.64.32.166 with HTTP; Tue, 22 Jan 2013 09:38:14 -0800 (PST) In-Reply-To: <50FDC9C0.40703@salesforce.com> References: <50FDC9C0.40703@salesforce.com> From: Harsh J Date: Tue, 22 Jan 2013 23:08:14 +0530 Message-ID: Subject: Re: CDH412/Hadoop 2.0.3 Upgrade instructions To: dbebortha@salesforce.com, "cdh-user@cloudera.org" Content-Type: multipart/alternative; boundary=14dae9340c0108c1ab04d3e40c6c X-Gm-Message-State: ALoCoQl5y0XmV8cWOOgQwJa+Ger2DHp8latDR5VCXATauGMhnRwda2PG1X+nVwO7LIkJBXiJNLRj X-Virus-Checked: Checked by ClamAV on apache.org --14dae9340c0108c1ab04d3e40c6c Content-Type: text/plain; charset=ISO-8859-1 Moving to cdh-user@cloudera.org as your question is CDH related. My answers inline: On Tue, Jan 22, 2013 at 4:35 AM, Dheeren bebortha wrote: > I am trying to upgrade a Hadoop Cluster with 0.20.X and MRv1 to a hadoop > Cluster with CDH412 with HA+QJM+YARN (aka Hadoop 2.0.3) without any data > loss and minimal down time. The documentation on cloudera site iis OK, but > very confusing. BTW I do not plan on using Cloudera manager. Has anyone > attempted a clean upgrade using hadoop native commands? > The upgrade process for any 0.20/1.x/CDH3 release to CDH4 is documented at https://ccp.cloudera.com/display/CDH4DOC/Upgrading+from+CDH3+to+CDH4. The only difference you may see is in use of packaging (tarballs or RPMs/DEBs?) and therefore, of usernames used in the guide. The basic process is to stop the older HDFS, remove older installation and all its traces carefully, and start the newer HDFS with the -upgrade flag. This takes care of HDFS metadata upgrades. Once done and you've verified that files/etc. are all perfectly readable and state's good, you can dfsadmin -finalizeUpgrade your cluster to commit the upgrade permanently. QJM is documented in a separate guide, found on the same portal mentioned above and can be upgraded in a second step after upgrade, to achieve full HA. For MR side, all your MR1 jobs will need to be recompiled before they may be run on the newer YARN+MR2 cluster due to some binary incompatible changes made between the versions you're upgrading. Other than a recompile, you may mostly not require to do anything else. May we also know your reason to not use CM when its aimed to make all this much easier to do and manage? We appreciate any form of feedback, thanks! -- Harsh J --14dae9340c0108c1ab04d3e40c6c Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable
Moving to cdh-use= r@cloudera.org as your question is CDH related.

My answers inline:

On Tue, Jan 22, 2013 at 4:35 AM, Dheeren bebortha <dbebortha@salesf= orce.com> wrote:
I am trying to upgrade a Hadoop Cluster with 0.20.X and MRv1 to a hadoop Cl= uster with CDH412 with HA+QJM+YARN (aka Hadoop 2.0.3) without any =A0data l= oss and minimal down time. The documentation on cloudera site iis OK, but v= ery confusing. BTW I do not plan on using Cloudera manager. Has anyone atte= mpted a clean upgrade using hadoop native commands?

The upgrade process for any 0.20/1.x= /CDH3 release to CDH4 is documented at=A0https://ccp.cloudera.com/dis= play/CDH4DOC/Upgrading+from+CDH3+to+CDH4. The only difference you may s= ee is in use of packaging (tarballs or RPMs/DEBs?) and therefore, of userna= mes used in the guide.

The basic process is to stop the older HDFS= , remove older installation and all its traces carefully, and start the new= er HDFS with the -upgrade flag. This takes care of HDFS metadata upgrades. = Once done and you've verified that files/etc. are all perfectly readabl= e and state's good, you can dfsadmin -finalizeUpgrade your cluster to c= ommit the upgrade permanently. QJM is documented in a separate guide, found= on the same portal mentioned above and can be upgraded in a second step af= ter upgrade, to achieve full HA.

For MR side, all your MR1 jobs will need to= be recompiled before they may be run on the newer YARN+MR2 cluster due to = some binary incompatible changes made between the versions you're upgra= ding. Other than a recompile, you may mostly not require to do anything els= e.

May we also know your reason to not use CM = when its aimed to make all this much easier to do and manage? We appreciate= any form of feedback, thanks!

--
Harsh J
--14dae9340c0108c1ab04d3e40c6c--