Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 67FD4924A for ; Sun, 8 Jul 2012 02:51:54 +0000 (UTC) Received: (qmail 22370 invoked by uid 500); 8 Jul 2012 02:51:51 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 22347 invoked by uid 500); 8 Jul 2012 02:51:51 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 22304 invoked by uid 99); 8 Jul 2012 02:51:49 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 08 Jul 2012 02:51:49 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=FSL_RCVD_USER,HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of tyler@datastax.com designates 209.85.217.172 as permitted sender) Received: from [209.85.217.172] (HELO mail-lb0-f172.google.com) (209.85.217.172) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 08 Jul 2012 02:51:43 +0000 Received: by lbbgo11 with SMTP id go11so16045593lbb.31 for ; Sat, 07 Jul 2012 19:51:22 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20120113; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type:x-gm-message-state; bh=TaCzTv/NTK7nJpwCDZssWn4TA16kgHiWtkHB0no6Jdg=; b=dzokUB766SleLmDb/Qql4icWr5Q8NnhCUuFKfkhGkgx0cWxgp35U10MRI7E15jmy5+ rpEcWl6/sBggn2rb35Hkr5+Se5okTUMwLkBYm4qF7gQP6tCbJ97D0Y1YYSIs8vSkdM4o gbsihSeJ5Xe5x/VMH29Qh/5eeOMgDAJUuCpyUZTHKbGcvkvZ6CpfgzKKYfuQdQUP436j 0CrYDgI4cpEikRoK1YLKBCIgVVmMt5VXy9HDPXZNUNX8A5syYzE55EK+pn8Ljig5frzZ PWngRqa4MjDZPrsVddRRebs6aSFrvZYBc3nhZa83RT23J5OaLJv3hvenlQtPi64B0y4H 0L7w== MIME-Version: 1.0 Received: by 10.152.148.1 with SMTP id to1mr35323880lab.34.1341715882387; Sat, 07 Jul 2012 19:51:22 -0700 (PDT) Received: by 10.112.59.132 with HTTP; Sat, 7 Jul 2012 19:51:22 -0700 (PDT) In-Reply-To: <4FF79AE3.9010302@syncopated.net> References: <4FF79AE3.9010302@syncopated.net> Date: Sat, 7 Jul 2012 21:51:22 -0500 Message-ID: Subject: Re: node vs node latency From: Tyler Hobbs To: user@cassandra.apache.org Content-Type: multipart/alternative; boundary=e89a8f22c34f8f288b04c448921f X-Gm-Message-State: ALoCoQkDce0oaCy382FICaChmmfW8zOGX2Dh3363xB7WSFu/H67fgFHS4PH3HYXzNAtr4cSfNPci --e89a8f22c34f8f288b04c448921f Content-Type: text/plain; charset=ISO-8859-1 Those latencies look like the difference between a couple of disk seeks and reading something that's already in the os cache. The dynamic snitch will favor nodes with lower latencies. Once a node has served enough reads, it might not have to hit disk very often, which produces lower latencies. So, if you have a hot dataset that fits into memory, the dynamic snitch starts a positive feedback loop where most reads will be served from one replica. I'm guessing the node with the low latencies is serving most of your reads. You can look at how quickly the total read count is increasing for each of the replicas to confirm this. It's not easy to do with only nodetool cfstats, but something like OpsCenter would help. On Fri, Jul 6, 2012 at 9:11 PM, Deno Vichas wrote: > all, > > what would explain a huge different (12ms vs 0.1ms) in read latency from > node to node. i've got a 4 node cluster w/ replication factor of 3 using > hector. i'm seeing these numbers with nodetool cfstats. > > > thx, > deno > -- Tyler Hobbs DataStax --e89a8f22c34f8f288b04c448921f Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable Those latencies look like the difference between a couple of disk seeks and= reading something that's already in the os cache.

The dynamic s= nitch will favor nodes with lower latencies.=A0 Once a node has served enou= gh reads, it might not have to hit disk very often, which produces lower la= tencies.=A0 So, if you have a hot dataset that fits into memory, the dynami= c snitch starts a positive feedback loop where most reads will be served fr= om one replica.

I'm guessing the node with the low latencies is serving most of you= r reads. You can look at how quickly the total read count is increasing for= each of the replicas to confirm this.=A0 It's not easy to do with only= nodetool cfstats, but something like OpsCenter would help.

On Fri, Jul 6, 2012 at 9:11 PM, Deno Vichas = <deno@syncopated.net> wrote:
all,

what would explain a huge different (12ms vs 0.1ms) in read latency from no= de to node. =A0i've got a 4 node cluster w/ replication factor of 3 usi= ng hector. =A0i'm seeing these numbers with nodetool cfstats.


thx,
deno



--
Tyler Hobbs
DataStax
<= br> --e89a8f22c34f8f288b04c448921f--