Hi

I can't understand the repair behavior in my case. I have 12 nodes ring (all 1.0.7):

10.254.237.2 LA ADS-LA-1 Up Normal 50.92 GB 0.00% 0 10.254.238.2 TX TX-24-RACK Up Normal 33.29 GB 0.00% 1 10.254.236.2 VA ADS-VA-1 Up Normal 50.07 GB 0.00% 2 10.254.93.2 IL R1 Up Normal 49.29 GB 0.00% 3 10.253.4.2 AZ R1 Up Normal 37.83 GB 0.00% 5 10.254.180.2 GB GB-1 Up Normal 42.86 GB 50.00% 85070591730234615865843651857942052863 10.254.191.2 LA ADS-LA-1 Up Normal 47.64 GB 0.00% 85070591730234615865843651857942052864 10.254.221.2 TX TX-24-RACK Up Normal 43.42 GB 0.00% 85070591730234615865843651857942052865 10.254.217.2 VA ADS-VA-1 Up Normal 38.44 GB 0.00% 85070591730234615865843651857942052866 10.254.94.2 IL R1 Up Normal 49.31 GB 0.00% 85070591730234615865843651857942052867 10.253.5.2 AZ R1 Up Normal 49.01 GB 0.00% 85070591730234615865843651857942052869 10.254.179.2 GB GB-1 Up Normal 27.08 GB 50.00% 170141183460469231731687303715884105727

I have single keyspace 'meter' and two column families (one 'ids' is small, and second is bigger). The strange thing happened today when I try to run
"nodetool -h 10.254.180.2 -pr meter ids"
two times one after another. First repair finished successfully

INFO 16:33:02,492 [repair #db582370-8bba-11e1-0000-5b777f708bff] ids is fully synced INFO 16:33:02,526 [repair #db582370-8bba-11e1-0000-5b777f708bff] session completed successfully

after moving near 50G of data, and I started second session one hour later:

INFO 17:44:37,842 [repair #aa415d00-8bd9-11e1-0000-5b777f708bff] new session: will sync localhost/1 0.254.180.2, /10.254.221.2, /10.254.191.2, /10.254.217.2, /10.253.5.2, /10.254.94.2 on range (5,8507
0591730234615865843651857942052863] for meter.[ids]

What is strange - when streams for the second repair starts they have the same or even bigger total volume, and I expected that second run will move less data (or even no data at all).

Is it OK? Or should I fix something?

Thanks!

Reply via email to