incubator-cassandra-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Jim Cistaro <>
Subject Re: removing old nodes
Date Fri, 22 Mar 2013 03:01:59 GMT
I do not recall what the "50" means, but IIRC, the 1364152145790 is the unix timestamp (in
millisecs rather than secs) of the expire time when they _should_ go away completely.

perl -e 'print scalar(gmtime(1364152145))'
Sun Mar 24 19:09:05 2013

From: Ben Chobot <<>>
Reply-To: "<>" <<>>
Date: Thursday, March 21, 2013 3:21 PM
To: "<>" <<>>
Subject: Re: removing old nodes

Ah, well I'll check back in a week then. But for the record, what I meant was that nodetool
gossipinfo now has entries like:


Where is shows "50" is where the token used to be, and where it still is on all my live nodes.
So it appears to me as if all my assassinated nodes now have a token of 50. Either way, they
don't seem to be bugging the rest of the cluster anymore, so thanks again.

On Mar 21, 2013, at 3:05 PM, Alain RODRIGUEZ wrote:

"(And now all sharing token 50? I dunno where that came from.)"

Not sure about what you mean.

"nodetool gossipinfo still shows all the old nodes there"

They must appear with a "left" or "remove" status. Off the top of my head, this information
will remains 7 days. Not sure about it.

2013/3/21 Ben Chobot <<>>
Thanks Alain, this seems to have stopped the log messages, even though nodetool gossipinfo
still shows all the old nodes there. (And now all sharing token 50? I dunno where that came
from.) Will they eventually fall away from the cluster, or are they there for good?

On Mar 21, 2013, at 11:53 AM, Alain RODRIGUEZ wrote:

Using the unsafeAssassinateEndpoint function with old IPs from JMX should do the trick.

This was already discussed in this mailing list, search using "unsafeAssassinateEndpoint"
as keyword to know all that you need to know about it.

Hope you'll be ok after that.


2013/3/21 Ben Chobot <<>>
I've got a 1.1.5 cluster, and a few weeks ago I removed some nodes from it. (I was trying
to upgrade nodes from AWS' large to xlarge, and for some reason that made sense at the time,
it seemed better to double my nodes and then decommission the smaller ones, rather than to
simply rebuild the existing nodes serially.)

Now the remaining nodes are all frequently logging that the old, decommissioned nodes are
dead and that their old token is being removed.... which is great, I guess, but why does my
cluster know about them at all? Doing a nodetool removetoken doesn't work, as the dead nodes
don't display in the ring. Is this expected behavior after a nodetool decommission? Is maybe
something cached that I can safely uncache?

View raw message