Return-Path: X-Original-To: apmail-lucene-solr-user-archive@minotaur.apache.org Delivered-To: apmail-lucene-solr-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 286E710DE4 for ; Thu, 6 Mar 2014 23:16:22 +0000 (UTC) Received: (qmail 95887 invoked by uid 500); 6 Mar 2014 23:16:17 -0000 Delivered-To: apmail-lucene-solr-user-archive@lucene.apache.org Received: (qmail 95766 invoked by uid 500); 6 Mar 2014 23:16:17 -0000 Mailing-List: contact solr-user-help@lucene.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: solr-user@lucene.apache.org Delivered-To: mailing list solr-user@lucene.apache.org Received: (qmail 95758 invoked by uid 99); 6 Mar 2014 23:16:17 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 06 Mar 2014 23:16:17 +0000 X-ASF-Spam-Status: No, hits=3.2 required=5.0 tests=FORGED_YAHOO_RCVD,HTML_MESSAGE,RCVD_IN_DNSWL_NONE,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: local policy) Received: from [98.136.218.147] (HELO nm3-vm4.bullet.mail.gq1.yahoo.com) (98.136.218.147) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 06 Mar 2014 23:16:08 +0000 Received: from [216.39.60.180] by nm3.bullet.mail.gq1.yahoo.com with NNFMP; 06 Mar 2014 23:15:47 -0000 Received: from [98.136.164.70] by tm16.bullet.mail.gq1.yahoo.com with NNFMP; 06 Mar 2014 23:15:47 -0000 Received: from [127.0.0.1] by smtp232.mail.gq1.yahoo.com with NNFMP; 06 Mar 2014 23:15:47 -0000 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=yahoo.com; s=s1024; t=1394147747; bh=tw9cVBHAVHoH0FS7AuweJkFURSarCMFsTq6/lSBhtPg=; h=X-Yahoo-Newman-Id:X-Yahoo-Newman-Property:X-YMail-OSG:X-Yahoo-SMTP:X-Rocket-Received:From:To:Subject:Date:Message-ID:MIME-Version:Content-Type:X-Mailer:Thread-Index:Content-Language; b=lHQGzz5z+JxDGqT759phj6voecr+5Qe4WdNMxxT6GyrlcFNNsEN85sI2BAJ7FQuCZpzKc2KqHyclrqFxskGNrFNC9dgXppmTcpYIVto8F8OfqAL0iqCRgP2SO7TYjBaBkiS5go3JxT94eLPDddFvJBeg9bEyLl4GTXsUpTjqffE= X-Yahoo-Newman-Id: 47325.41629.bm@smtp232.mail.gq1.yahoo.com X-Yahoo-Newman-Property: ymail-3 X-YMail-OSG: N0S_t_QVM1mPKxIXDBvq9UlfWsNb.IpqKVOvfD4i9rY28na Bp93qA3Kca48bPSucN573KJafefA3wYvSLvfhlOriVzVwK0__VDT2wwjtAp6 fynPreB7fuZdVjKWkNmxNrg.W4JkOghz9hCqZ3SlLiLAUgOHuI6_.p6maPOU evqN70lyXCbcF0ZYgBn7z0.FCEGEjTmzSFay4vnpZXTZkXJ.AMQioJ5jcrAu oZj1.LPi83mimFxk9rfwPDtBSa8pgObkU.SmvWtRDN1hYUEwBxEa253rZK5l e1wwoyd2CZ3DxXE25QPp9MANMnEbQ6GRUKq4A_imcaM8oQZB7m0kw06iwaes cVV_RUCv2lFzBJm8lmHnQbZMr4ArRUDUQov3_E7KpsGIyfo75LYklsL6NuP7 W0YYVB63nm_GADrv42H6QSc5ww.sHomm1jWLCQ2tRZk5fhXpYlRp9cTr6sAr MVlJKU3ugYwA4xRSp01Ab76KO0F5W454Y9DsYjzG.GibE8fvnSBCYLRZillz 61miwWDQPPfr3C8WBCKFabhtbS5PzwJfee.ajGY4qs.NLKSiDQK_HJAy0P0c vGuSrb_E- X-Yahoo-SMTP: Qt9TDmeswBCDMR2PlOkJr8u3iJ3u X-Rocket-Received: from redcastlePC (nazikhuq@173.48.137.133 with plain [98.138.105.21]) by smtp232.mail.gq1.yahoo.com with SMTP; 06 Mar 2014 23:15:47 +0000 UTC From: "Nazik Huq" To: Subject: SolrCloud recovery after nodes are rebooted in rapid succession Date: Thu, 6 Mar 2014 18:15:43 -0500 Message-ID: <007e01cf3992$00641080$012c3180$@yahoo.com> MIME-Version: 1.0 Content-Type: multipart/alternative; boundary="----=_NextPart_000_007F_01CF3968.178FB630" X-Mailer: Microsoft Outlook 14.0 Thread-Index: Ac85kbw0o9PEbmVPSkCIEUCfdwL0uw== Content-Language: en-us X-Virus-Checked: Checked by ClamAV on apache.org ------=_NextPart_000_007F_01CF3968.178FB630 Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit Hello, I have a question from a colleague who's managing a 3-node(VMs) SolrCloud cluster with a separate 3-node Zookeeper ensemble. Periodically the data center underneath the SolrCloud decides to upgrade the SolrCloud instance infrastructure in a "rolling upgrade" fashion. So after the 1st instance of the SolrCloud is shut down and while it is in the process of rebooting, the 2nd instance starts to shut down and so on. Eventually all three Solr instances are rebooted and up and running but the cluster in now inoperable. Meaning clients can't query or ingest data. My colleague is trying to ascertain if this problem is due to Solr's inability to recover from a rapid succession of reboots of the nodes or from the data center upgrade that is triggering a "situation" making SolrCloud inoperable. My question is, can a SolrCloud cluster become inoperable after its nodes are rebooted in rapid succession as described above? Is there an edge case similar to this? Thanks, Nazik Huq ------=_NextPart_000_007F_01CF3968.178FB630--