cassandra-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Igor <i...@4friends.od.ua>
Subject repair strange behavior
Date Sat, 21 Apr 2012 18:43:06 GMT
Hi

I can't understand the repair behavior in my case. I have 12 nodes ring 
(all 1.0.7):

10.254.237.2    LA          ADS-LA-1    Up     Normal  50.92 GB        
0.00%   0
10.254.238.2    TX          TX-24-RACK  Up     Normal  33.29 GB        
0.00%   1
10.254.236.2    VA          ADS-VA-1    Up     Normal  50.07 GB        
0.00%   2
10.254.93.2     IL          R1          Up     Normal  49.29 GB        
0.00%   3
10.253.4.2      AZ          R1          Up     Normal  37.83 GB        
0.00%   5
10.254.180.2    GB          GB-1        Up     Normal  42.86 GB        
50.00%  85070591730234615865843651857942052863
10.254.191.2    LA          ADS-LA-1    Up     Normal  47.64 GB        
0.00%   85070591730234615865843651857942052864
10.254.221.2    TX          TX-24-RACK  Up     Normal  43.42 GB        
0.00%   85070591730234615865843651857942052865
10.254.217.2    VA          ADS-VA-1    Up     Normal  38.44 GB        
0.00%   85070591730234615865843651857942052866
10.254.94.2     IL          R1          Up     Normal  49.31 GB        
0.00%   85070591730234615865843651857942052867
10.253.5.2      AZ          R1          Up     Normal  49.01 GB        
0.00%   85070591730234615865843651857942052869
10.254.179.2    GB          GB-1        Up     Normal  27.08 GB        
50.00%  170141183460469231731687303715884105727

I have single keyspace 'meter' and two column families (one 'ids' is 
small, and second is bigger). The strange thing happened today when I 
try to run
"nodetool -h 10.254.180.2 -pr meter ids"
two times one after another. First repair finished successfully

  INFO 16:33:02,492 [repair #db582370-8bba-11e1-0000-5b777f708bff] ids 
is fully synced
  INFO 16:33:02,526 [repair #db582370-8bba-11e1-0000-5b777f708bff] 
session completed successfully

after moving near 50G of data, and I started second session one hour later:

INFO 17:44:37,842 [repair #aa415d00-8bd9-11e1-0000-5b777f708bff] new 
session: will sync localhost/1
0.254.180.2, /10.254.221.2, /10.254.191.2, /10.254.217.2, /10.253.5.2, 
/10.254.94.2 on range (5,8507
0591730234615865843651857942052863] for meter.[ids]

What is strange - when streams for the second repair starts they have 
the same or even bigger total volume, and I expected that second run 
will move less data (or even no data at all).

Is it OK? Or should I fix something?

Thanks!


Mime
View raw message