Return-Path: Delivered-To: apmail-hadoop-common-user-archive@www.apache.org Received: (qmail 62576 invoked from network); 11 Aug 2010 14:44:35 -0000 Received: from unknown (HELO mail.apache.org) (140.211.11.3) by 140.211.11.9 with SMTP; 11 Aug 2010 14:44:35 -0000 Received: (qmail 2659 invoked by uid 500); 11 Aug 2010 14:44:33 -0000 Delivered-To: apmail-hadoop-common-user-archive@hadoop.apache.org Received: (qmail 2369 invoked by uid 500); 11 Aug 2010 14:44:30 -0000 Mailing-List: contact common-user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: common-user@hadoop.apache.org Delivered-To: mailing list common-user@hadoop.apache.org Received: (qmail 2361 invoked by uid 99); 11 Aug 2010 14:44:29 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 11 Aug 2010 14:44:29 +0000 X-ASF-Spam-Status: No, hits=0.0 required=10.0 tests=MIME_QP_LONG_LINE,RCVD_IN_DNSWL_NONE,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of charmalloc@allthingshadoop.com designates 209.85.161.176 as permitted sender) Received: from [209.85.161.176] (HELO mail-gx0-f176.google.com) (209.85.161.176) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 11 Aug 2010 14:44:24 +0000 Received: by gxk7 with SMTP id 7so90615gxk.35 for ; Wed, 11 Aug 2010 07:44:04 -0700 (PDT) Received: by 10.101.195.37 with SMTP id x37mr21792774anp.38.1281537843777; Wed, 11 Aug 2010 07:44:03 -0700 (PDT) Received: from [10.65.84.173] (mobile-166-137-138-199.mycingular.net [166.137.138.199]) by mx.google.com with ESMTPS id w6sm270861anb.23.2010.08.11.07.43.56 (version=TLSv1/SSLv3 cipher=RC4-MD5); Wed, 11 Aug 2010 07:44:01 -0700 (PDT) Subject: Re: Best practices - Large Hadoop Cluster References: <982106.53616.qm@web33504.mail.mud.yahoo.com> <8B12A3D37FC04AEB9AE98ED290EB937E@china.huawei.com> <997223.34532.qm@web33503.mail.mud.yahoo.com> <4C628C2A.9060005@apache.org> From: Joe Stein Content-Type: text/plain; charset=us-ascii X-Mailer: iPhone Mail (8A306) In-Reply-To: <4C628C2A.9060005@apache.org> Message-Id: <1C89D00A-5195-4CDF-AA89-EE49792EDCF6@allthingshadoop.com> Date: Wed, 11 Aug 2010 10:43:48 -0400 To: "common-user@hadoop.apache.org" Content-Transfer-Encoding: quoted-printable Mime-Version: 1.0 (iPhone Mail 8A306) Not sure this was mentioned already but Adobe open sourced their puppet impl= http://github.com/hstack/puppet as well as a nice post in regards to it h= ttp://hstack.org/hstack-automated-deployment-using-puppet/ /* Joe Stein http://www.linkedin.com/in/charmalloc Twitter: @allthingshadoop */ On Aug 11, 2010, at 7:40 AM, Steve Loughran wrote: > On 10/08/10 21:06, Raj V wrote: >> Mike >> 512 nodes, even a minute for each node ( ssh-ing to each node, typing a 8= >> character password, ensuring that everything looks ok) is about 8.5 hours= . After >> that if something does not work, that is a different level of pain altoge= ther. >>=20 >> Using scp to exchange keys simply does not scale. >>=20 >> My question was simple, how do other people in the group who run large cl= usters >> manage this? Brian put it better; Whats is the best, duplicatable way o= f >> running hadoop when the cluster is large. I agree, this is not a hadoop >> question per se, but hadoop is really what I care about now. >>=20 >=20 >=20 > SSH is great, but you still shouldn't be playing around trying to do thing= s by hand, even those parallel SSH tools break the moment you have a hint of= inconsistency between machines. >=20 >=20 > Instead general practise in managing *any large datacentre scale applicati= on*, be it hadoop or not is automate things so the machines do the work them= selves, leaving sysadmins to deal with important issues like why all packets= are being routed via singapore or whether the HDD failure rate is statistic= ally significant. >=20 > The standard techniques usually one of >=20 > * build your own RPMs, deb files, push out stuff with kickstart, change a m= achine by rebuilding its root disk. > Strengths: good for clean builds > Weaknesses: a lot of work, doesn't do recovery >=20 > * Model driven tools. I know most people now say "yes, puppet", but actual= ly cfEngine and bcfg2 have been around for a while, SmartFrog is what we use= . In these tools, you specify what you want, they keep an eye on things and p= ush the machines back into the desired state. > Strengths: recovers from bad state, keeps the machines close to the desire= d state > Weaknesses: if the desired state is not consistent, they tend to circle be= tween the various unreachable states. >=20 > * Scripts. People end up doing this without thinking. > Strengths: take your commands and script them, strong order to operations > Weaknesses: bad at recovery. >=20 > * VM images, maintained by hand or another technique > Strengths: OK if you have one gold image that can be pushed out every time= a VM is created -and VMs are short lived. > Weaknesses: Unless your VMs are short lived, you've just created a mainten= ance nightmare worse than before. >=20 >=20 > Hadoop itself is not too bad at handling failures of individual machines, b= ut the general best practices in large cluster management (look at LISA proc= eedings) are pretty much foundational. >=20 > http://wiki.smartfrog.org/wiki/display/sf/Patterns+of+Hadoop+Deployment >=20 > -Steve