Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 53FFD200C4F for ; Sat, 1 Apr 2017 21:14:18 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id 526F5160B9D; Sat, 1 Apr 2017 19:14:18 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 46CE3160B78 for ; Sat, 1 Apr 2017 21:14:17 +0200 (CEST) Received: (qmail 43676 invoked by uid 500); 1 Apr 2017 19:14:15 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 43666 invoked by uid 99); 1 Apr 2017 19:14:15 -0000 Received: from pnap-us-west-generic-nat.apache.org (HELO spamd1-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 01 Apr 2017 19:14:15 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd1-us-west.apache.org (ASF Mail Server at spamd1-us-west.apache.org) with ESMTP id 510A8C016F for ; Sat, 1 Apr 2017 19:14:15 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd1-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: 2.63 X-Spam-Level: ** X-Spam-Status: No, score=2.63 tagged_above=-999 required=6.31 tests=[DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, FREEMAIL_ENVFROM_END_DIGIT=0.25, HTML_MESSAGE=2, RCVD_IN_DNSWL_NONE=-0.0001, RCVD_IN_MSPIKE_H3=-0.01, RCVD_IN_MSPIKE_WL=-0.01, RCVD_IN_SORBS_SPAM=0.5, SPF_PASS=-0.001, URIBL_BLOCKED=0.001] autolearn=disabled Authentication-Results: spamd1-us-west.apache.org (amavisd-new); dkim=pass (2048-bit key) header.d=gmail.com Received: from mx1-lw-us.apache.org ([10.40.0.8]) by localhost (spamd1-us-west.apache.org [10.40.0.7]) (amavisd-new, port 10024) with ESMTP id 4dOVDYxPDKMS for ; Sat, 1 Apr 2017 19:14:14 +0000 (UTC) Received: from mail-qk0-f173.google.com (mail-qk0-f173.google.com [209.85.220.173]) by mx1-lw-us.apache.org (ASF Mail Server at mx1-lw-us.apache.org) with ESMTPS id 2305F5F340 for ; Sat, 1 Apr 2017 19:14:14 +0000 (UTC) Received: by mail-qk0-f173.google.com with SMTP id p22so89579865qka.3 for ; Sat, 01 Apr 2017 12:14:14 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=mime-version:in-reply-to:references:from:date:message-id:subject:to; bh=T5saoBmMCA/U2Rp0a5JT1odmxyutS7vdUzdhZY6i9Ek=; b=TCYv2DejvYWwmVYeJTcUqBXCEIARSPOOcn0pBVWRTnqhvoXppf32FiR4Gsk9sB2vHd +aHc182HjGBZOq/7zvEHPcfaKykSEdzDEXiCIzeTfRNZENZPICagOd8nvZa1EUrPJIf/ 0Y+nY3RWUiO/QOh6IAJC7JH5TZyxLXsT08CeZWWI8dcAlokjrcrg5BMmgdBGGTHsXBHS dzDDreCIvU5NxMv2AkX4IlMJYRPyurZl26evDJebDSZmMx9L/7EO86GJ2ekp1W2Br9Yx 78PjMXGdZnBkPlU9YfESZvkNlGc3MC8nJmL7UNgs7Q1WNZwZmyfMSyw95v3T36sNz8UZ LWIQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:in-reply-to:references:from:date :message-id:subject:to; bh=T5saoBmMCA/U2Rp0a5JT1odmxyutS7vdUzdhZY6i9Ek=; b=WcN7pqUIgjlBUuuuEEMevx5AUMktiurQLyH6tzANxG32XU9EW/wDkRFIied7IJ6BtO YHcxD1VDwlkKlplKqmLRCwysBrw0jvMg50i3CLppvjuAvetJp9HbtyEn5jJZqOE2FvsW EYgTX+4QHv6/8WGpPYUKb1V1SjZnZ1Ee5Wnc9FNKnvoFrjb4SLooUyf2WhxDLaccRGTV o1g/kvXWIY4u6dGP6ole1NP1LCs+Fnf4AO9G7UfnyvXYoJXl5ZhR1cHgybkpyMXNB+U1 ribzR75HPlzFlG5qajU48HkhZvXe+yp1g9n3n/6Q7uejcrcOB0ce//rT5NmNe6xSRqks 8czg== X-Gm-Message-State: AFeK/H1aJ95h+3civIupT/ZKfwgFsaY7NMXUuR8cm3Ju87wByKBwGlKZW6jSOUi2JW+yRB9Gg4m55m2gCT+yyA== X-Received: by 10.233.223.198 with SMTP id t189mr8432923qkf.292.1491074048607; Sat, 01 Apr 2017 12:14:08 -0700 (PDT) MIME-Version: 1.0 Received: by 10.140.107.197 with HTTP; Sat, 1 Apr 2017 12:14:08 -0700 (PDT) In-Reply-To: <1490862146.8615.4.camel@willhaben.at> References: <1490862146.8615.4.camel@willhaben.at> From: Chris Lohfink Date: Sat, 1 Apr 2017 12:14:08 -0700 Message-ID: Subject: Re: nodes are always out of sync To: user@cassandra.apache.org Content-Type: multipart/alternative; boundary=94eb2c04448e004bee054c1fbbf5 archived-at: Sat, 01 Apr 2017 19:14:18 -0000 --94eb2c04448e004bee054c1fbbf5 Content-Type: text/plain; charset=UTF-8 Repairs do not have an ability to instantly build a perfect view of its data between your 3 nodes at an exact time. When a piece of data is written there is a delay between when they applied between the nodes, even if its just 500ms. So if a request to read the data and build the merkle tree of the data occurs and it finishes on node1 at 12:01 while node2 finishes at 12:02 the 1 minute or so delta (even if a few seconds, or if using snapshot repairs) between the partition/range hashes in the merkle tree can be different. On a moving data set its almost impossible to have the clusters perfectly in sync for a repair. I wouldnt worry about that log message. If you are worried about consistency between your read/writes use each or local quorum for both. Chris On Thu, Mar 30, 2017 at 1:22 AM, Roland Otta wrote: > hi, > > we see the following behaviour in our environment: > > cluster consists of 6 nodes (cassandra version 3.0.7). keyspace has a > replication factor 3. > clients are writing data to the keyspace with consistency one. > > we are doing parallel, incremental repairs with cassandra reaper. > > even if a repair just finished and we are starting a new one > immediately, we can see the following entries in our logs: > > INFO [RepairJobTask:1] 2017-03-30 10:14:00,782 SyncTask.java:73 - > [repair #d0f651f6-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.188 > and /192.168.0.191 have 1 range(s) out of sync for ad_event_history > INFO [RepairJobTask:2] 2017-03-30 10:14:00,782 SyncTask.java:73 - > [repair #d0f651f6-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.188 > and /192.168.0.189 have 1 range(s) out of sync for ad_event_history > INFO [RepairJobTask:4] 2017-03-30 10:14:00,782 SyncTask.java:73 - > [repair #d0f651f6-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.189 > and /192.168.0.191 have 1 range(s) out of sync for ad_event_history > INFO [RepairJobTask:2] 2017-03-30 10:14:03,997 SyncTask.java:73 - > [repair #d0fa70a1-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.26 > and /192.168.0.189 have 2 range(s) out of sync for ad_event_history > INFO [RepairJobTask:1] 2017-03-30 10:14:03,997 SyncTask.java:73 - > [repair #d0fa70a1-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.26 > and /192.168.0.191 have 2 range(s) out of sync for ad_event_history > INFO [RepairJobTask:4] 2017-03-30 10:14:03,997 SyncTask.java:73 - > [repair #d0fa70a1-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.189 > and /192.168.0.191 have 2 range(s) out of sync for ad_event_history > INFO [RepairJobTask:1] 2017-03-30 10:14:05,375 SyncTask.java:73 - > [repair #d0fbd033-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.189 > and /192.168.0.191 have 1 range(s) out of sync for ad_event_history > INFO [RepairJobTask:2] 2017-03-30 10:14:05,375 SyncTask.java:73 - > [repair #d0fbd033-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.189 > and /192.168.0.190 have 1 range(s) out of sync for ad_event_history > INFO [RepairJobTask:4] 2017-03-30 10:14:05,375 SyncTask.java:73 - > [repair #d0fbd033-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.190 > and /192.168.0.191 have 1 range(s) out of sync for ad_event_history > > we cant see any hints on the systems ... so we thought everything is > running smoothly with the writes. > > do we have to be concerned about the nodes always being out of sync or > is this a normal behaviour in a write intensive table (as the tables > will never be 100% in sync for the latest inserts)? > > bg, > roland > > > --94eb2c04448e004bee054c1fbbf5 Content-Type: text/html; charset=UTF-8 Content-Transfer-Encoding: quoted-printable
Repairs do not have an ability to instantly build a perfec= t view of its data between your 3 nodes at an exact time. When a piece of d= ata is written there is a delay between when they applied between the nodes= , even if its just 500ms. So if a request to read the data and build the me= rkle tree of the data occurs and it finishes on node1 at 12:01 while node2 = finishes at 12:02 the 1 minute or so delta (even if a few seconds, or if us= ing snapshot repairs) between the partition/range hashes in the merkle tree= can be different. On a moving data set its almost impossible to have the c= lusters perfectly in sync for a repair. I wouldnt worry about that log mess= age. If you are worried about consistency between your read/writes use each= or local quorum for both.

Chris

On Thu, Mar 30, 2017 at 1:2= 2 AM, Roland Otta <Roland.Otta@willhaben.at> wrote:
hi,

we see the following behaviour in our environment:

cluster consists of 6 nodes (cassandra version 3.0.7). keyspace has a
replication factor 3.
clients are writing data to the keyspace with consistency one.

we are doing parallel, incremental repairs with cassandra reaper.

even if a repair just finished and we are starting a new one
immediately, we can see the following entries in our logs:

INFO=C2=A0=C2=A0[RepairJobTask:1] 2017-03-30 10:14:00,782 SyncTask.java:73 = -
[repair #d0f651f6-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.188=
and /= 192.168.0.191 have 1 range(s) out of sync for ad_event_history
INFO=C2=A0=C2=A0[RepairJobTask:2] 2017-03-30 10:14:00,782 SyncTask.java:73 = -
[repair #d0f651f6-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.188=
and /= 192.168.0.189 have 1 range(s) out of sync for ad_event_history
INFO=C2=A0=C2=A0[RepairJobTask:4] 2017-03-30 10:14:00,782 SyncTask.java:73 = -
[repair #d0f651f6-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.189=
and /= 192.168.0.191 have 1 range(s) out of sync for ad_event_history
INFO=C2=A0=C2=A0[RepairJobTask:2] 2017-03-30 10:14:03,997 SyncTask.java:73 = -
[repair #d0fa70a1-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.26 and /= 192.168.0.189 have 2 range(s) out of sync for ad_event_history
INFO=C2=A0=C2=A0[RepairJobTask:1] 2017-03-30 10:14:03,997 SyncTask.java:73 = -
[repair #d0fa70a1-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.26 and /= 192.168.0.191 have 2 range(s) out of sync for ad_event_history
INFO=C2=A0=C2=A0[RepairJobTask:4] 2017-03-30 10:14:03,997 SyncTask.java:73 = -
[repair #d0fa70a1-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.189=
and /= 192.168.0.191 have 2 range(s) out of sync for ad_event_history
INFO=C2=A0=C2=A0[RepairJobTask:1] 2017-03-30 10:14:05,375 SyncTask.java:73 = -
[repair #d0fbd033-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.189=
and /= 192.168.0.191 have 1 range(s) out of sync for ad_event_history
INFO=C2=A0=C2=A0[RepairJobTask:2] 2017-03-30 10:14:05,375 SyncTask.java:73 = -
[repair #d0fbd033-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.189=
and /= 192.168.0.190 have 1 range(s) out of sync for ad_event_history
INFO=C2=A0=C2=A0[RepairJobTask:4] 2017-03-30 10:14:05,375 SyncTask.java:73 = -
[repair #d0fbd033-1520-11e7-a443-d9f5b942818e] Endpoints /192.168.0.190=
and /= 192.168.0.191 have 1 range(s) out of sync for ad_event_history

we cant see any hints on the systems ... so we thought everything is
running smoothly with the writes.

do we have to be concerned about the nodes always being out of sync or
is this a normal behaviour in a write intensive table (as the tables
will never be 100% in sync for the latest inserts)?

bg,
roland



--94eb2c04448e004bee054c1fbbf5--