Return-Path: Delivered-To: apmail-hadoop-core-user-archive@www.apache.org Received: (qmail 68235 invoked from network); 11 Jan 2009 21:23:40 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (140.211.11.2) by minotaur.apache.org with SMTP; 11 Jan 2009 21:23:40 -0000 Received: (qmail 3949 invoked by uid 500); 11 Jan 2009 21:23:34 -0000 Delivered-To: apmail-hadoop-core-user-archive@hadoop.apache.org Received: (qmail 3913 invoked by uid 500); 11 Jan 2009 21:23:34 -0000 Mailing-List: contact core-user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: core-user@hadoop.apache.org Delivered-To: mailing list core-user@hadoop.apache.org Received: (qmail 3902 invoked by uid 99); 11 Jan 2009 21:23:34 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 11 Jan 2009 13:23:34 -0800 X-ASF-Spam-Status: No, hits=-0.0 required=10.0 tests=SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of david.ritch@gmail.com designates 209.85.198.226 as permitted sender) Received: from [209.85.198.226] (HELO rv-out-0506.google.com) (209.85.198.226) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 11 Jan 2009 21:23:26 +0000 Received: by rv-out-0506.google.com with SMTP id k40so10299695rvb.29 for ; Sun, 11 Jan 2009 13:23:06 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:received:received:message-id:date:user-agent :mime-version:to:subject:x-enigmail-version:x-enigmail-version :content-type:content-transfer-encoding:from; bh=u/9g5ywqFMSH/WAstJ4a2yWgTYb2QmF6le0ub1HWdGE=; b=c6aKd3rP/CSLbbwtaOwbFm3UJ3+KgAeJYRAsNhuUzNZLF0spieO0M/ZZ6ohpk9QNDu DX6LpFY4QXh1x1U1HIIeBHDw4Dsbt3+88jMlt9J0rDkpl2Lu+zZnYN+XnxNbJYbknpi2 tzpzZTMacBIowoLcdVXKfE5JDYsoFWWkqvkNI= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=message-id:date:user-agent:mime-version:to:subject :x-enigmail-version:content-type:content-transfer-encoding:from; b=lU7HJ7ljH1+sVUKQ0qaI0G4kIydGeJTt5WL9zuO6rvod/Tzuvx8jOI0iUpk22+qliF GM0ak/BDbmUOxQ3ksvnLF7qBvc3GdOwE/UM0MGXvO/lQG7TxpeucM52bFqVPRKBTAmQn y7y8BzKXuXhOwtPnsLm4L06obCr3mC3MpeMvE= Received: by 10.140.164.1 with SMTP id m1mr14025251rve.66.1231708986280; Sun, 11 Jan 2009 13:23:06 -0800 (PST) Received: from ?192.168.2.110? (c-69-250-61-164.hsd1.md.comcast.net [69.250.61.164]) by mx.google.com with ESMTPS id k41sm7064430rvb.3.2009.01.11.13.23.04 (version=TLSv1/SSLv3 cipher=RC4-MD5); Sun, 11 Jan 2009 13:23:05 -0800 (PST) Message-ID: <496A6335.8000100@gmail.com> Date: Sun, 11 Jan 2009 16:23:01 -0500 User-Agent: Thunderbird 2.0.0.19 (Windows/20081209) MIME-Version: 1.0 To: core-user@hadoop.apache.org Subject: RAID vs. JBOD X-Enigmail-Version: 0.95.7 X-Enigmail-Version: 0.95.7 Content-Type: text/plain; charset=ISO-8859-1 Content-Transfer-Encoding: 7bit From: "David B. Ritch" X-Virus-Checked: Checked by ClamAV on apache.org How well does Hadoop handle multiple independent disks per node? I have a cluster with 4 identical disks per node. I plan to use one disk for OS and temporary storage, and dedicate the other three to HDFS. Our IT folks have some disagreement as to whether the three disks should be striped, or treated by HDFS as three independent disks. Could someone with more HDFS experience comment on the relative advantages and disadvantages to each approach? Here are some of my thoughts. It's a bit easier to manage a 3-disk striped partition, and we wouldn't have to worry about balancing files between them. Single-file I/O should be considerably faster. On the other hand, I would expect typical use to require multiple files reads or write simultaneously. I would expect Hadoop to be able to manage read/write to/from the disks independently. Managing 3 streams to 3 independent devices would likely result in less disk head movement, and therefore better performance. I would expect Hadoop to be able to balance load between the disks fairly well. Availability doesn't really differentiate between the two approaches - if a single disk dies, the striped array would go down, but all its data should be replicated on another datanode, anyway. And besides, I understand that datanode will shut down a node, even if only one of 3 independent disks crashes. So - any one want to agree or disagree with these thoughts? Anyone have any other ideas, or - better - benchmarks and experience with layouts like these two? Thanks! David