Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 456D49ED0 for ; Tue, 14 May 2013 16:11:34 +0000 (UTC) Received: (qmail 42324 invoked by uid 500); 14 May 2013 16:11:31 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 42297 invoked by uid 500); 14 May 2013 16:11:31 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 42285 invoked by uid 99); 14 May 2013 16:11:31 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 14 May 2013 16:11:31 +0000 X-ASF-Spam-Status: No, hits=2.7 required=5.0 tests=FREEMAIL_ENVFROM_END_DIGIT,FREEMAIL_REPLYTO_END_DIGIT,HTML_MESSAGE,RCVD_IN_DNSWL_NONE,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: local policy) Received: from [98.139.212.168] (HELO nm9.bullet.mail.bf1.yahoo.com) (98.139.212.168) by apache.org (qpsmtpd/0.29) with SMTP; Tue, 14 May 2013 16:10:51 +0000 Received: from [98.139.215.141] by nm9.bullet.mail.bf1.yahoo.com with NNFMP; 14 May 2013 16:10:30 -0000 Received: from [98.139.212.219] by tm12.bullet.mail.bf1.yahoo.com with NNFMP; 14 May 2013 16:10:30 -0000 Received: from [127.0.0.1] by omp1028.mail.bf1.yahoo.com with NNFMP; 14 May 2013 16:10:30 -0000 X-Yahoo-Newman-Property: ymail-3 X-Yahoo-Newman-Id: 664913.47721.bm@omp1028.mail.bf1.yahoo.com Received: (qmail 54998 invoked by uid 60001); 14 May 2013 16:10:30 -0000 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=yahoo.com; s=s1024; t=1368547830; bh=UMgqKqc9bn3ZJqW16bpZUn0wBiGwXOEjtVMBNfJN3F0=; h=X-YMail-OSG:Received:X-Mailer:Message-ID:Date:From:Reply-To:Subject:To:MIME-Version:Content-Type; b=PZqvIPU+BA3yhe/Alc5CJoBNHVNll0ZmTP4w2/nq/AqWicf37WWY99FN+EXUh6rI7wCogDNFTnWSv/5jEB+pwei/3tZ0xROmFSc+yxqCmuniHBCfSELFgF6/LtUvkn3H0UYpAqAtVb8IdGzPu02MmwdxLkMq5D9mX7Z5s6l5S0c= DomainKey-Signature: a=rsa-sha1; q=dns; c=nofws; s=s1024; d=yahoo.com; h=X-YMail-OSG:Received:X-Mailer:Message-ID:Date:From:Reply-To:Subject:To:MIME-Version:Content-Type; b=5ixy13GmL0WIcIKKnlS3ZzVoWxadi9yqu549oqR0QpcEUP0j7iUqQPhLZX+95BKMJjbu+I95lx7FtdJKc/qHNag5DZZEmgbJ2sIy51dKhw7f2fj5VxDBW3YPGIU1NhKEAPz1owFxcCfJsz7eXomCkETjMtRrS1SXjfXFCBBluoI=; X-YMail-OSG: x7_cL38VM1kTFbaJFyxBiALuKdMGrvtEzXJGzjS8txasj.x 8RHtJzFBUumI3wkWvKQ2e9EWc62.e.y5Fkf1BD5zr9wMq3V1pEu3OCVmMXD2 ZHF.5Zi8Z7IhKXue3POAWwMXvTf74HLTkwm7D9sAB1pI8lQEER3w_7.h2ygs 5C_2pgqZEtCz9nkRI9XtehslSYoRVRrRxKjXkqRywCUe1uew.9qbkQFhOheE 9k1b8YkmDeU6QJZUi4x53HD_aetJ2o2NHuCc5G0LBuoqW_WjLjKZXV2EYy5y WQGxLo8mgTR7tXTO7cIRPaw.aUrJ0Fdznt1YaEY0Ae6hTqJKSSm6uEqyobZ9 7pE8DC1Hfy1Rq_LP1Yj3Unp9anjaxhqjjXSVIvd5dte1uzCavQZXpkOfcRde gN5dJaGPZ6vJxC8eunPfKZVeJ7DeQCkJQVn9KN3buYwb1f2LBaUtdka3d1BH Z8k0_y4Fb_.yqQpOwFJ1ZicfkyhN98YJU3XDPECKIacGIPZbLXKxu0i5kmJh CyJuN_dBasMOISfOqjOCMdRbIJd.jCSesG5y3JvyZTNCJ2urfjY18AMbVBSD Aj1mYuyuMCg-- Received: from [208.185.20.30] by web160901.mail.bf1.yahoo.com via HTTP; Tue, 14 May 2013 09:10:30 PDT X-Mailer: YahooMailWebService/0.8.141.536 Message-ID: <1368547830.22785.GenericBBA@web160901.mail.bf1.yahoo.com> Date: Tue, 14 May 2013 09:10:30 -0700 (PDT) From: Wei Zhu Reply-To: Wei Zhu Subject: Re: (unofficial) Community Poll for Production Operators : Repair To: user@cassandra.apache.org MIME-Version: 1.0 Content-Type: multipart/alternative; boundary="-285128509-1093262410-1368547830=:22785" X-Virus-Checked: Checked by ClamAV on apache.org ---285128509-1093262410-1368547830=:22785 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: quoted-printable 1) 1.1.6 on 5 nodes, 24CPU, 72 RAM =0A2) local quorum (we only have one DC = though). We do delete through TTL =0A3) yes =0A4) once a week rolling repai= rs -pr using cron job =0A5) it definitely has negative impact on the perfor= mance. Our data size is around 100G per node and during repair it brings in= additional 60G - 80G data and created about 7K compaction (We use LCS with= SSTable size of 10M which was a mistake we made at the beginning). It take= s more than a day for the compaction tasks to clear and by then the next co= mpaction starts. We had to set client side (Hector) timeout to deal with it= and the SLA is still under control for now. =0ABut we had to halt go live = for another cluster due to the unanticipated "double" the space during the = repair. =0A=0APer Dean's question to simulate the slow response, someone in= the IRC mentioned a trick to start Cassandra with -f and ctrl-z and it wor= ks for our test. =0A=0A-Wei =0A----- Original Message -----=0A=0AFrom: "Dea= n Hiller" =0ATo: user@cassandra.apache.org =0ASent: = Tuesday, May 14, 2013 4:48:02 AM =0ASubject: Re: (unofficial) Community Pol= l for Production Operators : Repair =0A=0AWe had to roll out a fix in cassa= ndra as a slow node was slowing down our clients of cassandra in 1.2.2 for = some reason. Every time we had a slow node, we found out fast as performanc= e degraded. We tested this in QA and had the same issue. This means a repai= r made that node slow which made our clients slow. With this fix which I th= ink one our team is going to try to get it back into cassandra, the slow no= de does not affect our clients anymore. =0A=0AI am curious though, if someo= ne else would use the "tc" program to simulate linux packet delay on a sing= le node, does your client's response time get much slower? We simulated a 5= 00ms delay on the node to simulate the slow node=E2=80=A6.it seems the co-o= rdinator node was incorrectly waiting for BOTH responses on CL_QUOROM inste= ad of just one (as itself was one as well) or something like that. (I don't= know too much as my colleague was the one that debugged this issue) =0A=0A= Dean =0A=0AFrom: Alain RODRIGUEZ > =0AReply-To: "user@cassandra.apache.org" > =0ADate= : Tuesday, May 14, 2013 1:42 AM =0ATo: "user@cassandra.apache.org" > =0ASubject: Re: (unofficial) Community Poll for Production Ope= rators : Repair =0A=0AHi Rob, =0A=0A1) 1.2.2 on 6 to 12 EC2 m1.xlarge =0A2)= Quorum R&W . Almost no deletes (just some TTL) =0A3) Yes =0A4) On each nod= e once a week (rolling repairs using crontab) =0A5) The only behavior that = is quite odd or unexplained to me is why a repair doesn't fix a counter mis= match between 2 nodes. I mean when I read my counters with a CL.One I have = inconsistency (the counter value may change anytime I read it, depending, I= guess, on what node I read from. Reading with CL.Quorum fixes this bug, bu= t the data is still wrong on some nodes. About performance, it's quite expe= nsive to run a repair but doing it in a low charge period and in a rolling = fashion works quite well and has no impact on the service. =0A=0AHope this = will help somehow. Let me know if you need more information. =0A=0AAlain = =0A=0A=0A=0A2013/5/10 Robert Coli > =0AHi! =0A=0AI have been wondering how Repair is actually used b= y operators. If =0Apeople operating Cassandra in production could answer th= e following =0Aquestions, I would greatly appreciate it. =0A=0A1) What vers= ion of Cassandra do you run, on what hardware? =0A2) What consistency level= do you write at? Do you do DELETEs? =0A3) Do you run a regularly scheduled= repair? =0A4) If you answered "yes" to 3, what is the frequency of the rep= air? =0A5) What has been your subjective experience with the performance of= =0Arepair? (Does it work as you would expect? Does its overhead have a =0A= significant impact on the performance of your cluster?) =0A=0AThanks! =0A= =0A=3DRob =0A=0A=0A ---285128509-1093262410-1368547830=:22785 Content-Type: text/html; charset=utf-8 Content-Transfer-Encoding: quoted-printable <= div style=3D'font-family: arial,helvetica,sans-serif; font-size: 10pt; colo= r: #000000'>1) 1.1.6 on 5 nodes, 24CPU, 72 RAM
2) local quorum (we only = have one DC though). We do delete through TTL
3) yes
4) once a week r= olling repairs -pr using cron job
5) it definitely has negative impact o= n the performance. Our data size is around 100G per node and during repair = it brings in additional 60G - 80G data and created about 7K compaction (We = use LCS with SSTable size of 10M which was a mistake we made at the beginni= ng). It takes more than a day for the compaction tasks to clear and by then= the next compaction starts. We had to set client side (Hector) timeout to = deal with it and the SLA is still under control for now.
But we had to h= alt go live for another cluster due to the unanticipated "double" the space= during the repair.

Per Dean's question to simulate the slow response, someone in the IRC mentioned a trick to start Cassandra with -f = and ctrl-z and it works for our test.

-Wei

From: "Dean Hiller" <Dean.Hiller@nrel.gov>
To: us= er@cassandra.apache.org
Sent: Tuesday, May 14, 2013 4:48:02 AMSubject: Re: (unofficial) Community Poll for Production Operators := Repair

We had to roll out a fix in cassandra as a slow node was slo= wing down our clients of cassandra in 1.2.2 for some reason.  Every ti= me we had a slow node, we found out fast as performance degraded.  We = tested this in QA and had the same issue.  This means a repair made th= at node slow which made our clients slow.  With this fix which I think= one our team is going to try to get it back into cassandra, the slow node does not affect our clients anymore.

I am curious though, if s= omeone else would use the "tc" program to simulate linux packet delay on a = single node, does your client's response time get much slower?  We sim= ulated a 500ms delay on the node to simulate the slow node=E2=80=A6.it seem= s the co-ordinator node was incorrectly waiting for BOTH responses on CL_QU= OROM instead of just one (as itself was one as well) or something like that= .  (I don't know too much as my colleague was the one that debugged th= is issue)

Dean

From: Alain RODRIGUEZ <arodrime@gmail.com&l= t;mailto:arodrime@gmail.com>>
Reply-To: "user@cassandra.apache.org= <mailto:user@cassandra.apache.org>" <user@cassandra.apache.org<= mailto:user@cassandra.apache.org>>
Date: Tuesday, May 14, 2013 1:4= 2 AM
To: "user@cassandra.apache.org<mailto:user@cassandra.apache.org&= gt;" <user@cassandra.apache.org<mailto:user@cassandra.apache.org>><= br>Subject: Re: (unofficial) Community Poll for Production Operators : Repa= ir

Hi Rob,

1) 1.2.2 on 6 to 12 EC2 m1.xlarge
2) Quorum R&a= mp;W . Almost no deletes (just some TTL)
3) Yes
4) On each node once = a week (rolling repairs using crontab)
5) The only behavior that is quit= e odd or unexplained to me is why a repair doesn't fix a counter mismatch b= etween 2 nodes. I mean when I read my counters with a CL.One I have inconsi= stency (the counter value may change anytime I read it, depending, I guess,= on what node I read from. Reading with CL.Quorum fixes this bug, but the d= ata is still wrong on some nodes. About performance, it's quite expensive t= o run a repair but doing it in a low charge period and in a rolling fashion= works quite well and has no impact on the service.

Hope this will h= elp somehow. Let me know if you need more information.

Alain



2013/5/10 Robert Coli <rcoli@e= ventbrite.com<mailto:rcoli@eventbrite.com>>
Hi!

I have b= een wondering how Repair is actually used by operators. If
people operat= ing Cassandra in production could answer the following
questions, I woul= d greatly appreciate it.

1) What version of Cassandra do you run, on= what hardware?
2) What consistency level do you write at? Do you do DEL= ETEs?
3) Do you run a regularly scheduled repair?
4) If you answered = "yes" to 3, what is the frequency of the repair?
5) What has been your s= ubjective experience with the performance of
repair? (Does it work as yo= u would expect? Does its overhead have a
significant impact on the perfo= rmance of your cluster?)

Thanks!

=3DRob


---285128509-1093262410-1368547830=:22785--