Return-Path: X-Original-To: apmail-hadoop-general-archive@minotaur.apache.org Delivered-To: apmail-hadoop-general-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 4D4FA94E5 for ; Fri, 7 Oct 2011 09:18:14 +0000 (UTC) Received: (qmail 10257 invoked by uid 500); 7 Oct 2011 09:18:12 -0000 Delivered-To: apmail-hadoop-general-archive@hadoop.apache.org Received: (qmail 10191 invoked by uid 500); 7 Oct 2011 09:18:12 -0000 Mailing-List: contact general-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: general@hadoop.apache.org Delivered-To: mailing list general@hadoop.apache.org Received: (qmail 10182 invoked by uid 99); 7 Oct 2011 09:18:11 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 07 Oct 2011 09:18:11 +0000 X-ASF-Spam-Status: No, hits=-1.6 required=5.0 tests=RCVD_IN_DNSWL_MED,SPF_NEUTRAL X-Spam-Check-By: apache.org Received-SPF: neutral (nike.apache.org: local policy) Received: from [192.6.10.2] (HELO colossus.hpl.hp.com) (192.6.10.2) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 07 Oct 2011 09:18:03 +0000 Received: from localhost (localhost [127.0.0.1]) by colossus.hpl.hp.com (Postfix) with ESMTP id 51AB11BA509 for ; Fri, 7 Oct 2011 10:17:42 +0100 (BST) X-Virus-Scanned: Debian amavisd-new at hpl.hp.com Received: from colossus.hpl.hp.com ([127.0.0.1]) by localhost (colossus.hpl.hp.com [127.0.0.1]) (amavisd-new, port 10024) with LMTP id 2DgYKBT-s6Tv for ; Fri, 7 Oct 2011 10:17:42 +0100 (BST) Received: from 0-imap-br1.hpl.hp.com (0-imap-br1.hpl.hp.com [16.25.144.60]) by colossus.hpl.hp.com (Postfix) with ESMTP id CC5CD1BA352 for ; Fri, 7 Oct 2011 10:17:41 +0100 (BST) MailScanner-NULL-Check: 1318583846.62015@P297mJIyApE2GF/DIvN8hQ Received: from [16.98.140.84] ([16.98.140.84]) by 0-imap-br1.hpl.hp.com (8.14.1/8.13.4) with ESMTP id p979HPV9009005 for ; Fri, 7 Oct 2011 10:17:25 +0100 (BST) Message-ID: <4E8EC39F.904@apache.org> Date: Fri, 07 Oct 2011 10:17:19 +0100 From: Steve Loughran User-Agent: Mozilla/5.0 (Windows; U; Windows NT 5.2; en-US; rv:1.9.2.17) Gecko/20110414 Thunderbird/3.1.10 MIME-Version: 1.0 To: general@hadoop.apache.org Subject: Re: Which proposed distro of Hadoop, 0.20.206 or 0.22, will be better for HBase? References: In-Reply-To: Content-Type: text/plain; charset=ISO-8859-1; format=flowed Content-Transfer-Encoding: 7bit X-HPL-MailScanner-Information: Please contact the ISP for more information X-MailScanner-ID: p979HPV9009005 X-HPL-MailScanner: Found to be clean X-HPL-MailScanner-From: stevel@apache.org X-Virus-Checked: Checked by ClamAV on apache.org On 06/10/2011 17:49, Milind.Bhandarkar@emc.com wrote: > Steve, > >> Summary: I'm not sure that HDFS is the right FS in this world, as it >> contains a lot of assumptions about system stability and HDD persistence >> that aren't valid any more. With the ability to plug in new placers you >> could do tricks like ensure 1 replica lives in a persistent blockstore >> (and rely on it always being there), and add other replicas in transient >> storage if the data is about to be needed in jobs. > > Can you please shed more light on the statement "... as it > contains a lot of assumptions about system stability and HDD persistence > that aren't valid any more..." ? > > I know that you were doing some analysis of disk failure modes sometime > ago. Is this the result of that research ? I am very interested. no, it's unrelated -experience in hosting virtual hadoop infrastructures. Which is how my short-lived clusters exist today -you don't know the hostname of the master nodes until allocated, so you need to allocate them and dynamically push out configs to the workers -the Datanodes spin when the namenode goes down, forever, rather than checking somewhere to see if its changed. HDFS HA may fix that. -It's dangerously easy to have >1 DN on the same physical host, losing independence of that replica. -It's possible for the entire cluster to go down without warning. MR-layer issues -again, the TaskTrackers spin when the JT goes down, rather than look to see if its moved. -Blacklisting isn't the right way to deal with task tracker failures: termination of VM is. -if the TT's are idle, VM termination may be the best action Hadoop is optimised for large physical clusters. If you look at the Stratosphere work at TuBerlin, they've designed something that includes VM allocation in the execution plan. you can improve Hadoop to make it more agile; my defunct Hadoop lifecycle branch did a lot of that, but you have to have everyone else using Hadoop to be willing to let the changes go in -and those changes mustn't impose a cost or risk to the physical cluster model.