Return-Path: Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: (qmail 82745 invoked from network); 14 Apr 2011 16:35:52 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (140.211.11.3) by minotaur.apache.org with SMTP; 14 Apr 2011 16:35:52 -0000 Received: (qmail 89043 invoked by uid 500); 14 Apr 2011 16:35:49 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 88998 invoked by uid 500); 14 Apr 2011 16:35:49 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 88990 invoked by uid 99); 14 Apr 2011 16:35:49 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 14 Apr 2011 16:35:49 +0000 X-ASF-Spam-Status: No, hits=2.2 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_NEUTRAL X-Spam-Check-By: apache.org Received-SPF: neutral (nike.apache.org: 74.125.83.44 is neither permitted nor denied by domain of nathan@milford.io) Received: from [74.125.83.44] (HELO mail-gw0-f44.google.com) (74.125.83.44) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 14 Apr 2011 16:35:40 +0000 Received: by gwb20 with SMTP id 20so990974gwb.31 for ; Thu, 14 Apr 2011 09:35:19 -0700 (PDT) Received: by 10.42.134.131 with SMTP id l3mr1311558ict.412.1302798919089; Thu, 14 Apr 2011 09:35:19 -0700 (PDT) MIME-Version: 1.0 Received: by 10.42.130.197 with HTTP; Thu, 14 Apr 2011 09:34:59 -0700 (PDT) X-Originating-IP: [204.145.64.58] From: Nathan Milford Date: Thu, 14 Apr 2011 12:34:59 -0400 Message-ID: Subject: Stress testing disk configurations. Your thoughts? To: user@cassandra.apache.org Content-Type: multipart/alternative; boundary=90e6ba61379ec93f7a04a0e38213 X-Virus-Checked: Checked by ClamAV on apache.org --90e6ba61379ec93f7a04a0e38213 Content-Type: text/plain; charset=UTF-8 Ahoy, I'm building out a new 0.7.4 cluster to migrate our 0.6.6 cluster to. While I'm waiting for the dev-side to get time to work on their side of the project I have a 10 node cluster evenly split across two data centers (NY & LA) and was looking to do some testing while I could. My primary focus is on disk configurations. Space isn't a huge issue, our current data set is ~30G on each node and I imagine that'll go up since I intend on tweaking the RF on the new cluster. Each node has 6 x 146G 10K SAS drives. I want to test: 1) 6 disks in R0 where everything is written to the same stripe 2) 1 disk for OS+Commitlog and 5 disks in R0 for data. 3) 1 disk for OS+Commitlog and 5 individual disks defined as separate data_file_directories. I suspect I'll see best performance with option 3, but the issue has become political\religious and there are internal doubts that separating the commit log and data will truly improve performance despite documentation and logic indicating otherwise. Thus the test :) Right now I've been tinkering and not being very scientific while I work out a testing methodology and get used to the tools. I've just been running zznate's cassandra-stress against a single node and measuring the time it takes to read and write N rows. Unscientifically I've found that they all perform about the same. It is hard to judge because, when writing to a single node, reads take exponentially longer. Writing 10M rows may take ~500 seconds, but reading will take ~5000 seconds. I'm sure this will even out when I test across more than one node. Early next week I'll be able to test against all 10 nodes with a realistic replication factor. I'd really love to hear some people's thoughts on methodologies and what I should be looking at/for other than iostat and the time for the test to inset/read. Thanks, nathan --90e6ba61379ec93f7a04a0e38213 Content-Type: text/html; charset=UTF-8 Content-Transfer-Encoding: quoted-printable
Ahoy,

I'm building out a new 0.7.4 cluste= r to migrate our 0.6.6 cluster to.

While I'm w= aiting for the dev-side to get time to work on=C2=A0their=C2=A0side of the = project I have a 10 node cluster evenly split across two data centers (NY &= amp; LA) and was looking to do some testing while I could.

My primary focus is on disk configurations. =C2=A0Space= isn't a huge issue, our current data set is ~30G on each node and I im= agine that'll go up since I intend on tweaking the RF on the new cluste= r.

Each node has 6 x 146G 10K SAS drives. =C2=A0I want to = test:

1) 6 disks in R0 where everything is written= to the same stripe
2) 1 disk for OS+Commitlog and 5 disks in R0 = for data.
3) 1 disk for OS+Commitlog and 5 individual disks defined as=C2=A0sepa= rate=C2=A0data_file_directories.

I suspect I'l= l see best performance with option 3, but the issue has become political\re= ligious and there are internal doubts that=C2=A0separating=C2=A0the commit = log and data will truly improve performance despite documentation and logic= indicating otherwise. =C2=A0Thus the test :)

Right now I've been tinkering and not being very sc= ientific while I work out a testing methodology and get used to the tools. = =C2=A0I've just been running zznate's cassandra-stress against a si= ngle node and measuring the time it takes to read and write N rows.=C2=A0

Unscientifically I've found that they all perform a= bout the same. It is hard to judge because, when writing to a single node, = reads take exponentially longer. =C2=A0Writing 10M rows may take ~500 secon= ds, but reading will take ~5000 seconds. =C2=A0I'm sure this will even = out when I test across more than one node.

Early next week I'll be able to test against all 10= nodes with a realistic replication factor.

I'= d really love to hear some people's thoughts on methodologies and what = I should be looking at/for other than iostat and the time for the test to i= nset/read.

Thanks,
nathan

--90e6ba61379ec93f7a04a0e38213--