Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id E540A10909 for ; Wed, 11 Sep 2013 16:46:35 +0000 (UTC) Received: (qmail 55294 invoked by uid 500); 11 Sep 2013 16:46:31 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 55228 invoked by uid 500); 11 Sep 2013 16:46:31 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 55160 invoked by uid 99); 11 Sep 2013 16:46:29 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 11 Sep 2013 16:46:29 +0000 X-ASF-Spam-Status: No, hits=-2.3 required=5.0 tests=RCVD_IN_DNSWL_MED,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of paul.cichonski@lithium.com designates 208.74.204.5 as permitted sender) Received: from [208.74.204.5] (HELO smtp.lithium.com) (208.74.204.5) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 11 Sep 2013 16:46:24 +0000 Received: from farseer.lithium.local (farseer.office.lithium.com [10.10.100.81]) by smtp.lithium.com (Postfix) with ESMTP id A5954108538 for ; Wed, 11 Sep 2013 09:46:02 -0700 (PDT) X-DomainKeys: Sendmail DomainKeys Filter v1.0.2 smtp.lithium.com A5954108538 Received: from farseer.lithium.local ([10.10.100.81]) by farseer.lithium.local ([10.10.100.81]) with mapi; Wed, 11 Sep 2013 09:46:02 -0700 From: Paul Cichonski To: "user@cassandra.apache.org" Date: Wed, 11 Sep 2013 09:46:00 -0700 Subject: RE: heavy insert load overloads CPUs, with MutationStage pending Thread-Topic: heavy insert load overloads CPUs, with MutationStage pending Thread-Index: Ac6vCkn9mDlckaJ7SySv6Tvi+YBdtAAA1QWA Message-ID: <541C7781A689464891C05251C07E8CCF3D9D10AA9C@farseer.lithium.local> References: <522F32FC.2030804@gmail.com> <522F5987.9030300@gmail.com> <52309741.9060407@gmail.com> In-Reply-To: <52309741.9060407@gmail.com> Accept-Language: en-US Content-Language: en-US X-MS-Has-Attach: X-MS-TNEF-Correlator: acceptlanguage: en-US Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: quoted-printable MIME-Version: 1.0 X-Virus-Checked: Checked by ClamAV on apache.org How much of the data you are writing is going against the same row key?=20 I've experienced some issues using CQL to write a full wide-row at once (ac= ross multiple threads) that exhibited some of the symptoms you have describ= ed (i.e., high cpu, dropped mutations).=20 This question goes into it a bit more: http://stackoverflow.com/questions/1= 8522191/using-cassandra-and-cql3-how-do-you-insert-an-entire-wide-row-in-a-= single-reque . I was able to solve my issue by switching to using the thrif= t batch_mutate to write a full wide-row at once instead of using many CQL I= NSERT statements.=20 -Paul > -----Original Message----- > From: Keith Freeman [mailto:8forty@gmail.com] > Sent: Wednesday, September 11, 2013 9:16 AM > To: user@cassandra.apache.org > Subject: Re: heavy insert load overloads CPUs, with MutationStage pending >=20 >=20 > On 09/10/2013 11:42 AM, Nate McCall wrote: > > With SSDs, you can turn up memtable_flush_writers - try 3 initially (1 > > by default) and see what happens. However, given that there are no > > entries in 'All time blocked' for such, they may be something else. > Tried that, it seems to have reduced the loads a little after everything > warmed-up, but not much. > > > > How are you inserting the data? >=20 > A java client on a separate box using the datastax java driver, 48 thread= s > writing 100 records each iteration as prepared batch statements. >=20 > At 5000 records/sec, the servers just can't keep up, so the client backs = up. > That's only 5M of data/sec, which doesn't seem like much. As I mentioned= , > switching to SSDs didn't help much, so I'm assuming at this point that th= e > server overloads are what's holding up the client.