cassandra-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Robert Stupp (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (CASSANDRA-8917) Upgrading from 2.0.9 to 2.1.3 with 3 nodes, CL = quorum causes exceptions
Date Sun, 08 Mar 2015 14:21:38 GMT

    [ https://issues.apache.org/jira/browse/CASSANDRA-8917?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14352059#comment-14352059
] 

Robert Stupp commented on CASSANDRA-8917:
-----------------------------------------

What I've seen from your logs:
* node3 went down at approx 19:01
* node3 was started at approx 19:11
* node3 went down at approx 19:14
* node3 was started at approx 19:41 with this neat message: {{Node localhost/127.0.0.1 state
jump to normal}} - looks like a configuration mistake (I assume you don't want 127.0.0.1).

Another thing I've noticed is a lot of {{HintedHandoffMetrics.java (line 79) /10.6.71.198
has 119756 dropped hints, because node is down past configured hint window.}} messages appearing
every some minutes for example on node1. Did you reconfigure the {{max_hint_window_in_ms}}
parameter in {{cassandra.yaml}} - maybe too low?

> Upgrading from 2.0.9 to 2.1.3 with 3 nodes, CL = quorum causes exceptions
> -------------------------------------------------------------------------
>
>                 Key: CASSANDRA-8917
>                 URL: https://issues.apache.org/jira/browse/CASSANDRA-8917
>             Project: Cassandra
>          Issue Type: Bug
>         Environment: C* 2.0.9, Centos 6.5, Java 1.7.0_72, spring data cassandra 1.1.1,
cassandra java driver 2.0.9
>            Reporter: Gary Ogden
>         Attachments: b_output.log, jersey_error.log, node1-system.log, node2-system.log,
node3-system.log
>
>
> We have java apps running on glassfish that read/write to our 3 node cluster running
on 2.0.9. 
> we have the CL set to quorum for all reads and writes.
> When we started to upgrade the first node and did the sstable upgrade on that node, we
started getting this error on reads and writes:
> com.datastax.driver.core.exceptions.UnavailableException: Not enough replica available
for query at consistency QUORUM (2 required but only 1 alive)
> How is that possible when we have 3 nodes total, and there was 2 that were up and it's
saying we can't get the required CL?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Mime
View raw message