Return-Path: Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: (qmail 378 invoked from network); 19 Dec 2010 02:16:46 -0000 Received: from unknown (HELO mail.apache.org) (140.211.11.3) by 140.211.11.9 with SMTP; 19 Dec 2010 02:16:46 -0000 Received: (qmail 2878 invoked by uid 500); 19 Dec 2010 02:16:44 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 2854 invoked by uid 500); 19 Dec 2010 02:16:44 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 2846 invoked by uid 99); 19 Dec 2010 02:16:44 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 19 Dec 2010 02:16:44 +0000 X-ASF-Spam-Status: No, hits=2.2 required=10.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_NEUTRAL X-Spam-Check-By: apache.org Received-SPF: neutral (athena.apache.org: local policy) Received: from [209.85.213.172] (HELO mail-yx0-f172.google.com) (209.85.213.172) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 19 Dec 2010 02:16:38 +0000 Received: by yxt33 with SMTP id 33so926679yxt.31 for ; Sat, 18 Dec 2010 18:16:17 -0800 (PST) Received: by 10.236.108.43 with SMTP id p31mr4691066yhg.55.1292724977443; Sat, 18 Dec 2010 18:16:17 -0800 (PST) Received: from [192.168.1.14] (99-8-186-71.lightspeed.snfcca.sbcglobal.net [99.8.186.71]) by mx.google.com with ESMTPS id z5sm1197321yhc.31.2010.12.18.18.16.15 (version=SSLv3 cipher=RC4-MD5); Sat, 18 Dec 2010 18:16:16 -0800 (PST) From: Chris Goffinet Mime-Version: 1.0 (Apple Message framework v1082) Content-Type: multipart/alternative; boundary=Apple-Mail-1-885161600 Subject: Re: Read Latency Degradation Date: Sat, 18 Dec 2010 18:16:14 -0800 In-Reply-To: To: user@cassandra.apache.org References: <5B3E740C-DB61-47AE-A8D6-9D8086A01422@gmx.net> Message-Id: <3DAB334F-86CE-4F85-BFEE-B4D29E56507A@chrisgoffinet.com> X-Mailer: Apple Mail (2.1082) --Apple-Mail-1-885161600 Content-Transfer-Encoding: quoted-printable Content-Type: text/plain; charset=us-ascii You can disable compaction and enable it later. Use nodetool and = setcompactionthreshold to 0 0 -Chris On Dec 18, 2010, at 6:05 PM, Wayne wrote: > Rereading through everything again I am starting to wonder if the page = cache is being affected by compaction. We have been heavily loading data = for weeks and compaction is basically running non-stop. The manual = compaction should be done some time tomorrow, so when totally caught up = I will try again. What changes can be hoped for in 1470 or 1882 in terms = of isolating compactions (or writes) affects on read requests? >=20 > Thanks >=20 >=20 >=20 > On Sat, Dec 18, 2010 at 2:36 PM, Peter Schuller = wrote: > > You are absolutely back to my main concern. Initially we were = consistently > > seeing < 10ms read latency and now we see 25ms (30GB sstable file), = 50ms > > (100GB sstable file) and 65ms (330GB table file) read times for a = single > > read with nothing else going on in the cluster. Concurrency is not = our > > problem/concern (at this point), our problem is slow reads in total > > isolation. Frankly the concern is that a 2TB node with a 1TB sstable = (worst > > case scenario) will result in > 100ms read latency in total = isolation. >=20 > So if you have a single non-concurrent client, along, submitting these > reads that take 65 ms - are you disk bound (according to the last > column of iostat -x 1), and how many reads per second (rps column) are > you seeing relative to client reads? Is the number of disk reads per > client read consistent with the actual number of sstables at the time? >=20 > The behavior you're describing really does seem indicative of a > problem, unless the the bottleneck is legitimately reads from disk > from multiple sstables resulting from rows being spread over said > sstables. >=20 > -- > / Peter Schuller >=20 --Apple-Mail-1-885161600 Content-Transfer-Encoding: 7bit Content-Type: text/html; charset=us-ascii You can disable compaction and enable it later. Use nodetool and setcompactionthreshold to 0 0

-Chris

On Dec 18, 2010, at 6:05 PM, Wayne wrote:

Rereading through everything again I am starting to wonder if the page cache is being affected by compaction. We have been heavily loading data for weeks and compaction is basically running non-stop. The manual compaction should be done some time tomorrow, so when totally caught up I will try again. What changes can be hoped for in 1470 or 1882 in terms of isolating compactions (or writes) affects on read requests?

Thanks



On Sat, Dec 18, 2010 at 2:36 PM, Peter Schuller <peter.schuller@infidyne.com> wrote:
> You are absolutely back to my main concern. Initially we were consistently
> seeing < 10ms read latency and now we see 25ms (30GB sstable file), 50ms
> (100GB sstable file) and 65ms (330GB table file) read times for a single
> read with nothing else going on in the cluster. Concurrency is not our
> problem/concern (at this point), our problem is slow reads in total
> isolation. Frankly the concern is that a 2TB node with a 1TB sstable (worst
> case scenario) will result in > 100ms read latency in total isolation.

So if you have a single non-concurrent client, along, submitting these
reads that take 65 ms - are you disk bound (according to the last
column of iostat -x 1), and how many reads per second (rps column) are
you seeing relative to client reads? Is the number of disk reads per
client read consistent with the actual number of sstables at the time?

The behavior you're describing really does seem indicative of a
problem, unless the the bottleneck is legitimately reads from disk
from multiple sstables resulting from rows being spread over said
sstables.

--
/ Peter Schuller


--Apple-Mail-1-885161600--