Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id DA2249251 for ; Wed, 16 May 2012 20:30:39 +0000 (UTC) Received: (qmail 95914 invoked by uid 500); 16 May 2012 20:30:37 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 95894 invoked by uid 500); 16 May 2012 20:30:37 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 95885 invoked by uid 99); 16 May 2012 20:30:37 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 16 May 2012 20:30:37 +0000 X-ASF-Spam-Status: No, hits=-0.0 required=5.0 tests=SPF_HELO_PASS,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of gcdcu-cassandra-user-1@m.gmane.org designates 80.91.229.3 as permitted sender) Received: from [80.91.229.3] (HELO plane.gmane.org) (80.91.229.3) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 16 May 2012 20:30:28 +0000 Received: from list by plane.gmane.org with local (Exim 4.69) (envelope-from ) id 1SUkrE-0002r4-Em for user@cassandra.apache.org; Wed, 16 May 2012 22:30:04 +0200 Received: from c-68-32-133-231.hsd1.nj.comcast.net ([68.32.133.231]) by main.gmane.org with esmtp (Gmexim 0.1 (Debian)) id 1AlnuQ-0007hv-00 for ; Wed, 16 May 2012 22:30:04 +0200 Received: from oleg.dulin by c-68-32-133-231.hsd1.nj.comcast.net with local (Gmexim 0.1 (Debian)) id 1AlnuQ-0007hv-00 for ; Wed, 16 May 2012 22:30:04 +0200 X-Injected-Via-Gmane: http://gmane.org/ To: user@cassandra.apache.org From: Oleg Dulin Subject: Re: how can we get (a lot) more performance from cassandra Date: Wed, 16 May 2012 16:25:23 -0400 Lines: 76 Message-ID: References: <4FB408EF.2070504@softwareprojects.com> Mime-Version: 1.0 Content-Type: text/plain; charset=iso-8859-1; format=flowed Content-Transfer-Encoding: 8bit X-Complaints-To: usenet@dough.gmane.org X-Gmane-NNTP-Posting-Host: c-68-32-133-231.hsd1.nj.comcast.net User-Agent: Unison/2.1.7 Indeed. This is how we are trying to solve this problem. Our application has a built-in cache that resembles a supercolumn or standardcolumn data structure and has API that resembles a combination of Pelops selector and mutator. You can do something like that for Hector. The cache is constrained and uses LRU to purge unused items and keep memory usage steady. It is not perfect and we have bugs still but it cuts down on 90% of cassandra reads. On 2012-05-16 20:07:11 +0000, Mike Peters said: > Hi Yiming, > > Cassandra is optimized for write-heavy environments. > > If you have a read-heavy application, you shouldn't be running your > reads through Cassandra. > > On the bright side - Cassandra read throughput will remain consistent, > regardless of your volume.� But you are going to have to "wrap" your > reads with memcache (or redis), so that the bulk of your reads can be > served from memory. > > > Thanks, > Mike Peters > > On 5/16/2012 3:59 PM, Yiming Sun wrote: > Hello, > > I asked the question as a follow-up under a different thread, so I > figure I should ask here instead in case the other one gets buried, and > besides, I have a little more information. > > "We find the lack of performance disturbing" as we are only able to get > about 3-4MB/sec read performance out of Cassandra. > > We are using cassandra as the backend for an IR repository of digital > texts. It is a read-mostly repository with�occasional�writes. �Each row > represents a book volume, and each column of a row represents a page of > the volume. �Granted the data size is small -- the average size of a > column text is 2-3KB, and each row has about 250 columns (varies quite > a bit from one volume to another). > > Currently we are running a 3-node cluster, and will soon be upgraded to > a 6-node setup. �Each node is a VM with 4 cores and 16GB of memory. > All VMs use SAN as disk storage. � > > To retrieve a volume, a slice query is used via Hector that specifies > the row key (the volume), and a list of column keys (pages), and the > consistency level is set to ONE. �It is typical to retrieve multiple > volumes per request. > > The read rate that I have been seeing is about 3-4 MB/sec, and that is > reading the raw bytes... using string serializer the rate is even > lower, about 2.2MB/sec. � > > The server log shows the GC ParNew frequently gets longer than 200ms, > often in the range of 4-5seconds. �But nowhere near 15 seconds (which > is an indication that JVM heap is being swapped out). > > Currently we have not added JNA. �From a blog post, it seems JNA is > able to increase the performance by 13%, and we are hoping to increase > the performance by something more like 1300% (3-4 MB/sec is just > disturbingly low). �And we are hesitant to disable swap entirely since > one of the nodes is running a couple other services > > Do you have any suggestions on how we may boost the performance? �Thanks! > > -- Y.