Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 60D9B200C7E for ; Tue, 23 May 2017 18:12:42 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id 5C07E160BC3; Tue, 23 May 2017 16:12:42 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 7C1D4160BB6 for ; Tue, 23 May 2017 18:12:41 +0200 (CEST) Received: (qmail 34547 invoked by uid 500); 23 May 2017 16:12:39 -0000 Mailing-List: contact solr-user-help@lucene.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: solr-user@lucene.apache.org Delivered-To: mailing list solr-user@lucene.apache.org Received: (qmail 34535 invoked by uid 99); 23 May 2017 16:12:39 -0000 Received: from pnap-us-west-generic-nat.apache.org (HELO spamd1-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 23 May 2017 16:12:39 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd1-us-west.apache.org (ASF Mail Server at spamd1-us-west.apache.org) with ESMTP id 243C8C072F for ; Tue, 23 May 2017 16:12:39 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd1-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: 2.13 X-Spam-Level: ** X-Spam-Status: No, score=2.13 tagged_above=-999 required=6.31 tests=[DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, FREEMAIL_ENVFROM_END_DIGIT=0.25, HTML_MESSAGE=2, RCVD_IN_DNSWL_NONE=-0.0001, RCVD_IN_MSPIKE_H3=-0.01, RCVD_IN_MSPIKE_WL=-0.01, SPF_PASS=-0.001, URIBL_BLOCKED=0.001] autolearn=disabled Authentication-Results: spamd1-us-west.apache.org (amavisd-new); dkim=pass (2048-bit key) header.d=gmail.com Received: from mx1-lw-eu.apache.org ([10.40.0.8]) by localhost (spamd1-us-west.apache.org [10.40.0.7]) (amavisd-new, port 10024) with ESMTP id BEcGk9jdtmVr for ; Tue, 23 May 2017 16:12:37 +0000 (UTC) Received: from mail-wm0-f47.google.com (mail-wm0-f47.google.com [74.125.82.47]) by mx1-lw-eu.apache.org (ASF Mail Server at mx1-lw-eu.apache.org) with ESMTPS id 39A8A5FC57 for ; Tue, 23 May 2017 16:12:37 +0000 (UTC) Received: by mail-wm0-f47.google.com with SMTP id e127so35901993wmg.1 for ; Tue, 23 May 2017 09:12:37 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=mime-version:in-reply-to:references:from:date:message-id:subject:to; bh=i1IllfdV5U/twuALVJ/ZBIz5j2PzRDYuAL1y5ox8qpY=; b=jCmCXoplM3FhCLVKUIXDw0hULrALScsy5sjAXzMMxYOnDCOW9+eybID5XLbyoHbkCV +tzApQ4hhXQ0pmE/Cz/1b1Ug9xURk5p5aN2g+u8fl/x/e8N76q/ObQ0fizsnTPNcFw23 mJR+RdzW59S9l0JJejuskraACIjX+H18EJ9qjWo67nxS1JlClU/jm5wMuibuVqI+njXV OA7SsxBKFps26e8H9jVZKvA4PggyWo5hFGDQ4c3swKdlOJ4PVHCNG49/dc59sWRlu8sb NEW1cA9BXXZuffudy/MGxxOJc+45tFwM1Ibt5aylydpSrUe1gtogs9jkZFUilLt3hK+W jDMw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:in-reply-to:references:from:date :message-id:subject:to; bh=i1IllfdV5U/twuALVJ/ZBIz5j2PzRDYuAL1y5ox8qpY=; b=ia9wdUatrpbzXeGMOYX9jiOpWV5qyJl9mPtX0aCqkHygkHUUZFGBo9wcy7onHhffmZ STPtSfgVu+W9SXIVXRA3vLrkE//B9w15pWQlbRbeYUX4LFyjDRosAevyeA6tyFem8Yp3 bSkZ/dmkl6hJB7d20BVrzGhuk6CQxpp36hlswOvx/5fJxwLt7ulE61z71xjpBbnZFv40 E4y+MPBAyWmnZvgKjnwxiA5wB9n3rDACp7bAIHsDYKPzeOQ17yuBPLkl/hdycXerQHMR oifs/ZnuzmBcDusBY2V6rnvly7+jNnaEX6qnrqTY642asuH5X1XxeIBnGgIlvl3P3EC/ deaA== X-Gm-Message-State: AODbwcB9BKthtuyD3wj82DAftGOPPdWl91ivgn6VSbnCGPIDVt63dKio kyAU6tHiE3qufS+2koPhgvD6wfkaODdJ X-Received: by 10.80.167.163 with SMTP id i32mr22699436edc.101.1495555956848; Tue, 23 May 2017 09:12:36 -0700 (PDT) MIME-Version: 1.0 Received: by 10.80.167.227 with HTTP; Tue, 23 May 2017 09:12:36 -0700 (PDT) In-Reply-To: References: From: Susheel Kumar Date: Tue, 23 May 2017 12:12:36 -0400 Message-ID: Subject: Re: Spread SolrCloud across two locations To: solr-user@lucene.apache.org Content-Type: multipart/alternative; boundary="94eb2c1947548cbd0d055033413b" archived-at: Tue, 23 May 2017 16:12:42 -0000 --94eb2c1947548cbd0d055033413b Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable Hi Jan, FYI - Since last year, I have been running a Solr 6.0 cluster in one of lower env with 6 shards/replica in dc1 & 6 shard/replica in dc2 (eac= h shard replicated cross data center) with 3 ZK in dc1 and 2 ZK in dc2. (I didn't have the availability of 3rd data center for ZK so went with only 2 data center with above configuration) and so far no issues. Its been running fine, indexing, replicating data, serving queries etc. So in my test, setting up single cluster across two zones/data center works without any issue when there is no or very minimal latency (in my case around 30ms one way) Thanks, Susheel On Tue, May 23, 2017 at 9:20 AM, Jan H=C3=B8ydahl w= rote: > I.e. tell the customer that in order to have automatic failover and > recovery in a 2-location setup we require at least one ZK instance in a > separate third location. Kind of a tough requirement but necessary to > safe-guard against split brain during network partition. > > If a third location is not an option, how would you setup ZK for manual > reconfiguration? > Two ZK in DC1 and one in DC2 would give you automatic recovery in case DC= 2 > falls out, but if DC1 falls out, WRITE would be disabled and to resume > write in DC2 only, one would need to stop Solr + ZK, reconfigure ZK in DC= 2 > as standalone (or setup two more) and then start Solr again with only one > ZK. > > -- > Jan H=C3=B8ydahl, search solution architect > Cominvent AS - www.cominvent.com > > > 23. mai 2017 kl. 11.14 skrev Markus Jelsma = : > > > > I would probably start by renting a VM at a third location to run > Zookeeper. > > > > Markus > > > > -----Original message----- > >> From:Jan H=C3=B8ydahl > >> Sent: Tuesday 23rd May 2017 11:09 > >> To: solr-user > >> Subject: Spread SolrCloud across two locations > >> > >> Hi, > >> > >> A customer has two locations (a few km apart) with super-fast > networking in-between, so for day-to-day operation they view all VMs in > both locations as a pool of servers. They typically spin up redundant > servers for various services in each zone and if a zone should fail (they > are a few km apart), the other will just continue working. > >> > >> How can we best support such a setup with Cloud and Zookeeper? > >> They do not need (or want) CDCR since latency and bandwidth is no > problem, and CDCR is active-passive only so it anyway requires manual > intervention to catch up if indexing is switched to the passive DC > temporarily. > >> If it was not for ZK I would setup one Cloud cluster and make sure eac= h > shard was replicated cross zones and all would be fine. > >> But ZK really requires a third location in order to tolerate loss of a= n > entire location/zone. > >> All solutions I can think of involves manual intervention, > re-configuring of ZK followed by a restart of the surviving Solr nodes in > order to point to the =E2=80=9Cnew=E2=80=9D ZK. > >> > >> How have you guys solved such setups? > >> > >> -- > >> Jan H=C3=B8ydahl, search solution architect > >> Cominvent AS - www.cominvent.com > >> > >> > > --94eb2c1947548cbd0d055033413b--