Just saw a very similar question from Laxmikanth (laxmikanth...@gmail.com)
on an other thread, with the same logs.

Would you mind to avoid splitting multiple threads, to gather up
informations so we can better help you from this mailing list?

C*heers,

2016-09-28 14:28 GMT+02:00 Alain RODRIGUEZ <arodr...@gmail.com>:

> Hi,
>
> It looks like streams are failing right away when trying to rebuild.
>
>
>    - Could you please share with us the command you used?
>
>
> It should be ran from DC3 servers, after altering keyspace to add
> keyspaces to the new datacenter. Is this the way you're doing it?
>
>    - Are all the nodes using the same version ('nodetool version')?
>    - What does 'nodetool status keyspace_name1' output?
>    - Are you sure to be using Network Topology Strategy on '*keyspace_name1'?
>    *Have you modified this schema to add replications on DC3
>
> My guess is something could be wrong with the configuration.
>
> I checked with our network operations team , they have confirmed network
>> is stable and no network hiccups.
>> I have set 'streaming_socket_timeout_in_ms: 86400000' (24 hours) as
>> suggested in datastax blog  - https://support.datastax.com
>> /hc/en-us/articles/206502913-FAQ-How-to-reduce-the-impact-
>> of-streaming-errors-or-failures and ran 'nodetool rebuild' one node at a
>> time but was of NO USE . Still we are getting above exception.
>>
>
> This look correct to me, good you added this information, thanks.
>
> An other thought is I believe you need all the nodes to be up to have
> those streams working on the origin DC you use for your 'nodetool rebuild
> <origin_dc>' command.
>
> This look a bit weird, good luck.
>
> C*heers,
> -----------------------
> Alain Rodriguez - @arodream - al...@thelastpickle.com
> France
>
> The Last Pickle - Apache Cassandra Consulting
> http://www.thelastpickle.com
>
>
> 2016-09-27 18:54 GMT+02:00 techpyaasa . <techpya...@gmail.com>:
>
>> Hi,
>>
>> I'm trying to add new data center - DC3 to existing c*-2.0.17 cluster
>> with 2 data centers DC1, DC2 with replication DC1:3 , DC2:3 , DC3:3.
>>
>>  I'm getting following exception repeatedly on new nodes after I run
>> 'nodetool rebuild'.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> *DEBUG [ScheduledTasks:1] 2016-09-27 04:24:00,416 GCInspector.java (line
>> 118) GC for ParNew: 20 ms for 1 collections, 9837479688 used; max is
>> 16760438784DEBUG [ScheduledTasks:1] 2016-09-27 04:24:03,417
>> GCInspector.java (line 118) GC for ParNew: 20 ms for 1 collections,
>> 9871193904 used; max is 16760438784DEBUG [ScheduledTasks:1] 2016-09-27
>> 04:24:06,418 GCInspector.java (line 118) GC for ParNew: 20 ms for 1
>> collections, 9950298136 used; max is 16760438784DEBUG [ScheduledTasks:1]
>> 2016-09-27 04:24:09,419 GCInspector.java (line 118) GC for ParNew: 19 ms
>> for 1 collections, 9941119568 used; max is 16760438784DEBUG
>> [ScheduledTasks:1] 2016-09-27 04:24:12,421 GCInspector.java (line 118) GC
>> for ParNew: 20 ms for 1 collections, 9864185024 used; max is
>> 16760438784DEBUG [ScheduledTasks:1] 2016-09-27 04:24:15,422
>> GCInspector.java (line 118) GC for ParNew: 60 ms for 2 collections,
>> 9730374352 used; max is 16760438784DEBUG [ScheduledTasks:1] 2016-09-27
>> 04:24:18,423 GCInspector.java (line 118) GC for ParNew: 18 ms for 1
>> collections, 9775448168 used; max is 16760438784DEBUG [ScheduledTasks:1]
>> 2016-09-27 04:24:21,424 GCInspector.java (line 118) GC for ParNew: 22 ms
>> for 1 collections, 9850794272 used; max is 16760438784DEBUG
>> [ScheduledTasks:1] 2016-09-27 04:24:24,425 GCInspector.java (line 118) GC
>> for ParNew: 20 ms for 1 collections, 9729992448 <9729992448> used; max is
>> 16760438784DEBUG [ScheduledTasks:1] 2016-09-27 04:24:27,426
>> GCInspector.java (line 118) GC for ParNew: 22 ms for 1 collections,
>> 9699783920 used; max is 16760438784DEBUG [ScheduledTasks:1] 2016-09-27
>> 04:24:30,427 GCInspector.java (line 118) GC for ParNew: 21 ms for 1
>> collections, 9696523920 used; max is 16760438784DEBUG [ScheduledTasks:1]
>> 2016-09-27 04:24:33,429 GCInspector.java (line 118) GC for ParNew: 20 ms
>> for 1 collections, 9560497904 used; max is 16760438784DEBUG
>> [ScheduledTasks:1] 2016-09-27 04:24:36,430 GCInspector.java (line 118) GC
>> for ParNew: 19 ms for 1 collections, 9568718352 <9568718352> used; max is
>> 16760438784DEBUG [ScheduledTasks:1] 2016-09-27 04:24:39,431
>> GCInspector.java (line 118) GC for ParNew: 22 ms for 1 collections,
>> 9496991384 <9496991384> used; max is 16760438784DEBUG [ScheduledTasks:1]
>> 2016-09-27 04:24:42,432 GCInspector.java (line 118) GC for ParNew: 19 ms
>> for 1 collections, 9486433840 used; max is 16760438784DEBUG
>> [ScheduledTasks:1] 2016-09-27 04:24:45,434 GCInspector.java (line 118) GC
>> for ParNew: 19 ms for 1 collections, 9442642688 used; max is
>> 16760438784DEBUG [ScheduledTasks:1] 2016-09-27 04:24:48,435
>> GCInspector.java (line 118) GC for ParNew: 20 ms for 1 collections,
>> 9548532008 <9548532008> used; max is 16760438784DEBUG
>> [STREAM-IN-/xxx.xxx.98.168] 2016-09-27 04:24:49,756 ConnectionHandler.java
>> (line 244) [Stream #5e1b7f40-8496-11e6-8847-1b88665e430d] Received File
>> (Header (cfId: bf446a90-71c5-3552-a2e5-b1b94dbf86e3, #0, version: jb,
>> estimated keys: 252928, transfer size: 5496759656, compressed?: true),
>> file:
>> /home/cassandra/data_directories/data/keyspace_name1/columnfamily_1/keyspace_name1-columnfamily_1-tmp-jb-54-Data.db)DEBUG
>> [STREAM-OUT-/xxx.xxx.98.168] 2016-09-27 04:24:49,757 ConnectionHandler.java
>> (line 310) [Stream #5e1b7f40-8496-11e6-8847-1b88665e430d] Sending Received
>> (bf446a90-71c5-3552-a2e5-b1b94dbf86e3, #0)ERROR
>> [STREAM-OUT-/xxx.xxx.98.168] 2016-09-27 04:24:49,759 StreamSession.java
>> (line 461) [Stream #5e1b7f40-8496-11e6-8847-1b88665e430d] Streaming error
>> occurredjava.io.IOException: Connection timed out    at
>> sun.nio.ch.FileDispatcherImpl.write0(Native Method)    at
>> sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:47)    at
>> sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:93)    at
>> sun.nio.ch.IOUtil.write(IOUtil.java:65)    at
>> sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:487)    at
>> org.apache.cassandra.streaming.messages.StreamMessage.serialize(StreamMessage.java:44)
>> at
>> org.apache.cassandra.streaming.ConnectionHandler$OutgoingMessageHandler.sendMessage(ConnectionHandler.java:339)
>> at
>> org.apache.cassandra.streaming.ConnectionHandler$OutgoingMessageHandler.run(ConnectionHandler.java:311)
>> at java.lang.Thread.run(Thread.java:745)DEBUG [STREAM-OUT-/xxx.xxx.98.168]
>> 2016-09-27 04:24:49,764 ConnectionHandler.java (line 104) [Stream
>> #5e1b7f40-8496-11e6-8847-1b88665e430d] Closing stream connection handler on
>> /xxx.xxx.98.168 INFO [STREAM-OUT-/xxx.xxx.98.168] 2016-09-27 04:24:49,764
>> StreamResultFuture.java (line 186) [Stream
>> #5e1b7f40-8496-11e6-8847-1b88665e430d] Session with /xxx.xxx.98.168 is
>> completeERROR [STREAM-OUT-/xxx.xxx.98.168] 2016-09-27 04:24:49,764
>> StreamSession.java (line 461) [Stream
>> #5e1b7f40-8496-11e6-8847-1b88665e430d] Streaming error
>> occurredjava.io.IOException: Broken pipe    at
>> sun.nio.ch.FileDispatcherImpl.write0(Native Method)    at
>> sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:47)    at
>> sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:93)    at
>> sun.nio.ch.IOUtil.write(IOUtil.java:65)    at
>> sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:487)    at
>> org.apache.cassandra.streaming.messages.StreamMessage.serialize(StreamMessage.java:44)
>> at
>> org.apache.cassandra.streaming.ConnectionHandler$OutgoingMessageHandler.sendMessage(ConnectionHandler.java:339)
>> at
>> org.apache.cassandra.streaming.ConnectionHandler$OutgoingMessageHandler.run(ConnectionHandler.java:319)
>> at java.lang.Thread.run(Thread.java:745)DEBUG [STREAM-IN-/xxx.xxx.98.168]
>> 2016-09-27 04:24:49,909 ConnectionHandler.java (line 244) [Stream
>> #5e1b7f40-8496-11e6-8847-1b88665e430d] Received File (Header (cfId:
>> 68af9ee0-96f8-3b1d-a418-e5ae844f2cc2, #3, version: jb, estimated keys:
>> 4736, transfer size: 2306880, compressed?: true), file:
>> /home/cassandra/data_directories/data/keyspace_name1/archiving_metadata/keyspace_name1-archiving_metadata-tmp-jb-27-Data.db)ERROR
>> [STREAM-IN-/xxx.xxx.98.168] 2016-09-27 04:24:49,909 StreamSession.java
>> (line 461) [Stream #5e1b7f40-8496-11e6-8847-1b88665e430d] Streaming error
>> occurredjava.lang.RuntimeException: Outgoing stream handler has been
>> closed    at
>> org.apache.cassandra.streaming.ConnectionHandler.sendMessage(ConnectionHandler.java:126)
>> at
>> org.apache.cassandra.streaming.StreamSession.receive(StreamSession.java:524)
>> at
>> org.apache.cassandra.streaming.StreamSession.messageReceived(StreamSession.java:413)
>> at
>> org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:245)
>> at java.lang.Thread.run(Thread.java:745)*
>>
>>
>> I checked with our network operations team , they have confirmed network
>> is stable and no network hiccups.
>> I have set 'streaming_socket_timeout_in_ms: 86400000' (24 hours) as
>> suggested in datastax blog  - https://support.datastax.com/h
>> c/en-us/articles/206502913-FAQ-How-to-reduce-the-impact-of-s
>> treaming-errors-or-failures and ran 'nodetool rebuild' one node at a
>> time but was of NO USE . Still we are getting above exception.
>>
>> Can someone please help me in debugging and fixing this.
>>
>>
>> Thanks,
>> techpyaasa
>>
>>
>>
>>
>

Reply via email to