[jira] [Commented] (CASSANDRA-11044) query under certain partition key takes much more time than expected
[ https://issues.apache.org/jira/browse/CASSANDRA-11044?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15112076#comment-15112076 ] xiaost commented on CASSANDRA-11044: seems duplicated of CASSANDRA-8546 > query under certain partition key takes much more time than expected > - > > Key: CASSANDRA-11044 > URL: https://issues.apache.org/jira/browse/CASSANDRA-11044 > Project: Cassandra > Issue Type: Bug > Components: Local Write-Read Paths >Reporter: xiaost > Attachments: tracing.log > > > In my cluster, all the nodes is under low workload, > but query under certain partition key (we found one) takes much more time > than expected. > we write & updates about 3 times per row in one day, > reads are much more than writes. > HARDWARD: > 6*nodes(E5-2630, 1*ssd with 5GB data) > TABLE DESCRIBE: > {noformat} > CREATE TABLE album.user_updates ( > user_id bigint, > time_uuid bigint, > key ascii, > PRIMARY KEY (user_id, time_uuid) > ) WITH CLUSTERING ORDER BY (time_uuid ASC) > AND bloom_filter_fp_chance = 0.01 > AND caching = '{"keys":"ALL", "rows_per_partition":"NONE"}' > AND comment = '' > AND compaction = {'class': > 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy'} > AND compression = {'sstable_compression': > 'org.apache.cassandra.io.compress.LZ4Compressor'} > AND dclocal_read_repair_chance = 0.1 > AND default_time_to_live = 0 > AND gc_grace_seconds = 864000 > AND max_index_interval = 2048 > AND memtable_flush_period_in_ms = 0 > AND min_index_interval = 128 > AND read_repair_chance = 0.0 > AND speculative_retry = '99.0PERCENTILE'; > {noformat} > QUERYs: > {noformat} > select * from user_updates where user_id = 1432138730701829 limit 100; > select count(1) from user_updates where user_id = 1432138730701829; > {noformat} > RESULT: (takes about 3.5 minutes) > {noformat} > count > --- > 1058 > (1 rows) > {noformat} > check attachments for the tracing log -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (CASSANDRA-11044) query under certain partition key takes much more time than expected
[ https://issues.apache.org/jira/browse/CASSANDRA-11044?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15110251#comment-15110251 ] xiaost commented on CASSANDRA-11044: {noformat} "SharedPool-Worker-3" #371 daemon prio=5 os_prio=0 tid=0x7fd3412ad000 nid=0x3bad runnable [0x7fd21f687000] java.lang.Thread.State: RUNNABLE at org.apache.cassandra.db.RangeTombstoneList.insertFrom(RangeTombstoneList.java:649) at org.apache.cassandra.db.RangeTombstoneList.add(RangeTombstoneList.java:170) at org.apache.cassandra.db.RangeTombstoneList.add(RangeTombstoneList.java:142) at org.apache.cassandra.db.DeletionInfo.add(DeletionInfo.java:239) at org.apache.cassandra.db.ArrayBackedSortedColumns.delete(ArrayBackedSortedColumns.java:490) at org.apache.cassandra.db.ColumnFamily.addAtom(ColumnFamily.java:153) at org.apache.cassandra.db.filter.QueryFilter$2.getNext(QueryFilter.java:189) at org.apache.cassandra.db.filter.QueryFilter$2.hasNext(QueryFilter.java:161) at org.apache.cassandra.utils.MergeIterator$Candidate.advance(MergeIterator.java:146) at org.apache.cassandra.utils.MergeIterator$ManyToOne.(MergeIterator.java:89) at org.apache.cassandra.utils.MergeIterator.get(MergeIterator.java:48) at org.apache.cassandra.db.filter.QueryFilter.collateColumns(QueryFilter.java:107) at org.apache.cassandra.db.filter.QueryFilter.collateOnDiskAtom(QueryFilter.java:83) at org.apache.cassandra.db.filter.QueryFilter.collateOnDiskAtom(QueryFilter.java:70) at org.apache.cassandra.db.CollationController.collectAllData(CollationController.java:317) at org.apache.cassandra.db.CollationController.getTopLevelColumns(CollationController.java:61) at org.apache.cassandra.db.ColumnFamilyStore.getTopLevelColumns(ColumnFamilyStore.java:2011) at org.apache.cassandra.db.ColumnFamilyStore.getColumnFamily(ColumnFamilyStore.java:1815) at org.apache.cassandra.db.Keyspace.getRow(Keyspace.java:360) at org.apache.cassandra.db.SliceFromReadCommand.getRow(SliceFromReadCommand.java:85) at org.apache.cassandra.db.ReadVerbHandler.doVerb(ReadVerbHandler.java:38) at org.apache.cassandra.net.MessageDeliveryTask.run(MessageDeliveryTask.java:67) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at org.apache.cassandra.concurrent.AbstractLocalAwareExecutorService$FutureTask.run(AbstractLocalAwareExecutorService.java:164) at org.apache.cassandra.concurrent.AbstractLocalAwareExecutorService$LocalSessionFutureTask.run(AbstractLocalAwareExecutorService.java:136) at org.apache.cassandra.concurrent.SEPWorker.run(SEPWorker.java:105) at java.lang.Thread.run(Thread.java:745) {noformat} here is the stack. I investigated into the code and found that {{gatherTombstones.getNext}} sometimes iters 9000+ times and takes 5s in {{returnCF.addAtom(atom)}} I am trying to set gc_grace_seconds with a small value and compact and table. > query under certain partition key takes much more time than expected > - > > Key: CASSANDRA-11044 > URL: https://issues.apache.org/jira/browse/CASSANDRA-11044 > Project: Cassandra > Issue Type: Bug > Components: Local Write-Read Paths >Reporter: xiaost > Attachments: tracing.log > > > In my cluster, all the nodes is under low workload, > but query under certain partition key (we found one) takes much more time > than expected. > we write & updates about 3 times per row in one day, > reads are much more than writes. > HARDWARD: > 6*nodes(E5-2630, 1*ssd with 5GB data) > TABLE DESCRIBE: > {noformat} > CREATE TABLE album.user_updates ( > user_id bigint, > time_uuid bigint, > key ascii, > PRIMARY KEY (user_id, time_uuid) > ) WITH CLUSTERING ORDER BY (time_uuid ASC) > AND bloom_filter_fp_chance = 0.01 > AND caching = '{"keys":"ALL", "rows_per_partition":"NONE"}' > AND comment = '' > AND compaction = {'class': > 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy'} > AND compression = {'sstable_compression': > 'org.apache.cassandra.io.compress.LZ4Compressor'} > AND dclocal_read_repair_chance = 0.1 > AND default_time_to_live = 0 > AND gc_grace_seconds = 864000 > AND max_index_interval = 2048 > AND memtable_flush_period_in_ms = 0 > AND min_index_interval = 128 > AND read_repair_chance = 0.0 > AND speculative_retry = '99.0PERCENTILE'; > {noformat} > QUERYs: > {noformat} > select * from user_updates where user_id = 1432138730701829 limit 100; > select count(1) from user_updates where user_id = 1432138730701829; > {noformat} > RESULT: (takes about 3.5 minutes) > {noformat} > count > --- > 1058
[jira] [Comment Edited] (CASSANDRA-11044) query under certain partition key takes much more time than expected
[ https://issues.apache.org/jira/browse/CASSANDRA-11044?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15110251#comment-15110251 ] xiaost edited comment on CASSANDRA-11044 at 1/21/16 8:06 AM: - {noformat} "SharedPool-Worker-3" #371 daemon prio=5 os_prio=0 tid=0x7fd3412ad000 nid=0x3bad runnable [0x7fd21f687000] java.lang.Thread.State: RUNNABLE at org.apache.cassandra.db.RangeTombstoneList.insertFrom(RangeTombstoneList.java:649) at org.apache.cassandra.db.RangeTombstoneList.add(RangeTombstoneList.java:170) at org.apache.cassandra.db.RangeTombstoneList.add(RangeTombstoneList.java:142) at org.apache.cassandra.db.DeletionInfo.add(DeletionInfo.java:239) at org.apache.cassandra.db.ArrayBackedSortedColumns.delete(ArrayBackedSortedColumns.java:490) at org.apache.cassandra.db.ColumnFamily.addAtom(ColumnFamily.java:153) at org.apache.cassandra.db.filter.QueryFilter$2.getNext(QueryFilter.java:189) at org.apache.cassandra.db.filter.QueryFilter$2.hasNext(QueryFilter.java:161) at org.apache.cassandra.utils.MergeIterator$Candidate.advance(MergeIterator.java:146) at org.apache.cassandra.utils.MergeIterator$ManyToOne.(MergeIterator.java:89) at org.apache.cassandra.utils.MergeIterator.get(MergeIterator.java:48) at org.apache.cassandra.db.filter.QueryFilter.collateColumns(QueryFilter.java:107) at org.apache.cassandra.db.filter.QueryFilter.collateOnDiskAtom(QueryFilter.java:83) at org.apache.cassandra.db.filter.QueryFilter.collateOnDiskAtom(QueryFilter.java:70) at org.apache.cassandra.db.CollationController.collectAllData(CollationController.java:317) at org.apache.cassandra.db.CollationController.getTopLevelColumns(CollationController.java:61) at org.apache.cassandra.db.ColumnFamilyStore.getTopLevelColumns(ColumnFamilyStore.java:2011) at org.apache.cassandra.db.ColumnFamilyStore.getColumnFamily(ColumnFamilyStore.java:1815) at org.apache.cassandra.db.Keyspace.getRow(Keyspace.java:360) at org.apache.cassandra.db.SliceFromReadCommand.getRow(SliceFromReadCommand.java:85) at org.apache.cassandra.db.ReadVerbHandler.doVerb(ReadVerbHandler.java:38) at org.apache.cassandra.net.MessageDeliveryTask.run(MessageDeliveryTask.java:67) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at org.apache.cassandra.concurrent.AbstractLocalAwareExecutorService$FutureTask.run(AbstractLocalAwareExecutorService.java:164) at org.apache.cassandra.concurrent.AbstractLocalAwareExecutorService$LocalSessionFutureTask.run(AbstractLocalAwareExecutorService.java:136) at org.apache.cassandra.concurrent.SEPWorker.run(SEPWorker.java:105) at java.lang.Thread.run(Thread.java:745) {noformat} here is the stack. I investigated into the code and found that {{gatherTombstones.getNext}} sometimes iters 9000+ times and takes 5s in {{returnCF.addAtom(atom)}} I am trying to set {{gc_grace_seconds}} with a small value and compact and table. was (Author: xiaost): {noformat} "SharedPool-Worker-3" #371 daemon prio=5 os_prio=0 tid=0x7fd3412ad000 nid=0x3bad runnable [0x7fd21f687000] java.lang.Thread.State: RUNNABLE at org.apache.cassandra.db.RangeTombstoneList.insertFrom(RangeTombstoneList.java:649) at org.apache.cassandra.db.RangeTombstoneList.add(RangeTombstoneList.java:170) at org.apache.cassandra.db.RangeTombstoneList.add(RangeTombstoneList.java:142) at org.apache.cassandra.db.DeletionInfo.add(DeletionInfo.java:239) at org.apache.cassandra.db.ArrayBackedSortedColumns.delete(ArrayBackedSortedColumns.java:490) at org.apache.cassandra.db.ColumnFamily.addAtom(ColumnFamily.java:153) at org.apache.cassandra.db.filter.QueryFilter$2.getNext(QueryFilter.java:189) at org.apache.cassandra.db.filter.QueryFilter$2.hasNext(QueryFilter.java:161) at org.apache.cassandra.utils.MergeIterator$Candidate.advance(MergeIterator.java:146) at org.apache.cassandra.utils.MergeIterator$ManyToOne.(MergeIterator.java:89) at org.apache.cassandra.utils.MergeIterator.get(MergeIterator.java:48) at org.apache.cassandra.db.filter.QueryFilter.collateColumns(QueryFilter.java:107) at org.apache.cassandra.db.filter.QueryFilter.collateOnDiskAtom(QueryFilter.java:83) at org.apache.cassandra.db.filter.QueryFilter.collateOnDiskAtom(QueryFilter.java:70) at org.apache.cassandra.db.CollationController.collectAllData(CollationController.java:317) at org.apache.cassandra.db.CollationController.getTopLevelColumns(CollationController.java:61) at org.apache.cassandra.db.ColumnFamilyStore.getTopLevelColumns(ColumnFamilyStore.java:2011) at
[jira] [Created] (CASSANDRA-11044) query under certain partition key takes much more time than expected
xiaost created CASSANDRA-11044: -- Summary: query under certain partition key takes much more time than expected Key: CASSANDRA-11044 URL: https://issues.apache.org/jira/browse/CASSANDRA-11044 Project: Cassandra Issue Type: Bug Components: Local Write-Read Paths Reporter: xiaost Attachments: tracing.log In my cluster, all the nodes is under low workload, but query under certain partition key (we found one) takes much more time than expected. we write & updates about 3 times per row in one day, reads are much more than writes. HARDWARD: 6*nodes(E5-2630, 1*ssd with 5GB data) TABLE DESCRIBE: {noformat} CREATE TABLE album.user_updates ( user_id bigint, time_uuid bigint, key ascii, PRIMARY KEY (user_id, time_uuid) ) WITH CLUSTERING ORDER BY (time_uuid ASC) AND bloom_filter_fp_chance = 0.01 AND caching = '{"keys":"ALL", "rows_per_partition":"NONE"}' AND comment = '' AND compaction = {'class': 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy'} AND compression = {'sstable_compression': 'org.apache.cassandra.io.compress.LZ4Compressor'} AND dclocal_read_repair_chance = 0.1 AND default_time_to_live = 0 AND gc_grace_seconds = 864000 AND max_index_interval = 2048 AND memtable_flush_period_in_ms = 0 AND min_index_interval = 128 AND read_repair_chance = 0.0 AND speculative_retry = '99.0PERCENTILE'; {noformat} QUERYs: {noformat} select * from user_updates where user_id = 1432138730701829 limit 100; select count(1) from user_updates where user_id = 1432138730701829; {noformat} RESULT: (takes about 3.5 minutes) {noformat} count --- 1058 (1 rows) {noformat} check attachments for the tracing log -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15088654#comment-15088654 ] xiaost commented on CASSANDRA-10961: wonderful! I will test it ASAP. > Not enough bytes error when add nodes to cluster > > > Key: CASSANDRA-10961 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging >Reporter: xiaost >Assignee: Paulo Motta > Attachments: apache-cassandra-2.2.4-SNAPSHOT.jar, debug.1.log, > debug.logs.zip, netstats.1.log > > > we got the same problem all the time when we add nodes to cluster. > netstats: > on HostA > {noformat} > /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB > {noformat} > on HostB > {noformat} > tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from > idx:0/HostA > {noformat} > After a while, Error on HostB > {noformat} > WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - > [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error > java.lang.IllegalArgumentException: Not enough bytes > at > org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) > ~[guava-16.0.jar:na] > at > com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > [apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] > ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - > Exception in thread Thread[Thread-28,5,main] > java.lang.RuntimeException: java.lang.InterruptedException > at com.google.common.base.Throwables.propagate(Throwables.java:160) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] > Caused by: java.lang.InterruptedException: null > at > java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) > ~[na:1.8.0_66-internal] > at > org.apache.cassandra.streaming.compress.CompressedInputStream$Reader.runMayThrow(CompressedInputStream.java:176) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) > ~[apache-cassandra-2.2.4.jar:2.2.4] > ... 1 common
[jira] [Commented] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15088801#comment-15088801 ] xiaost commented on CASSANDRA-10961: It works! over 500GB streaming data is received without any error. hope release in 2.2.5 soon thanks [~pauloricardomg] > Not enough bytes error when add nodes to cluster > > > Key: CASSANDRA-10961 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging >Reporter: xiaost >Assignee: Paulo Motta > Attachments: apache-cassandra-2.2.4-SNAPSHOT.jar, debug.1.log, > debug.logs.zip, netstats.1.log > > > we got the same problem all the time when we add nodes to cluster. > netstats: > on HostA > {noformat} > /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB > {noformat} > on HostB > {noformat} > tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from > idx:0/HostA > {noformat} > After a while, Error on HostB > {noformat} > WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - > [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error > java.lang.IllegalArgumentException: Not enough bytes > at > org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) > ~[guava-16.0.jar:na] > at > com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > [apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] > ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - > Exception in thread Thread[Thread-28,5,main] > java.lang.RuntimeException: java.lang.InterruptedException > at com.google.common.base.Throwables.propagate(Throwables.java:160) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] > Caused by: java.lang.InterruptedException: null > at > java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) > ~[na:1.8.0_66-internal] > at > org.apache.cassandra.streaming.compress.CompressedInputStream$Reader.runMayThrow(CompressedInputStream.java:176) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at >
[jira] [Commented] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15083423#comment-15083423 ] xiaost commented on CASSANDRA-10961: [^debug.logs.zip] all in one > Not enough bytes error when add nodes to cluster > > > Key: CASSANDRA-10961 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging >Reporter: xiaost >Assignee: Paulo Motta > Attachments: debug.1.log, debug.logs.zip, netstats.1.log > > > we got the same problem all the time when we add nodes to cluster. > netstats: > on HostA > {noformat} > /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB > {noformat} > on HostB > {noformat} > tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from > idx:0/HostA > {noformat} > After a while, Error on HostB > {noformat} > WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - > [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error > java.lang.IllegalArgumentException: Not enough bytes > at > org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) > ~[guava-16.0.jar:na] > at > com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > [apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] > ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - > Exception in thread Thread[Thread-28,5,main] > java.lang.RuntimeException: java.lang.InterruptedException > at com.google.common.base.Throwables.propagate(Throwables.java:160) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] > Caused by: java.lang.InterruptedException: null > at > java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) > ~[na:1.8.0_66-internal] > at > org.apache.cassandra.streaming.compress.CompressedInputStream$Reader.runMayThrow(CompressedInputStream.java:176) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) > ~[apache-cassandra-2.2.4.jar:2.2.4] > ... 1 common frames omitted > ERROR [STREAM-IN-/HostA]
[jira] [Updated] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] xiaost updated CASSANDRA-10961: --- Attachment: debug.logs.zip > Not enough bytes error when add nodes to cluster > > > Key: CASSANDRA-10961 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging >Reporter: xiaost >Assignee: Paulo Motta > Attachments: debug.1.log, debug.logs.zip, netstats.1.log > > > we got the same problem all the time when we add nodes to cluster. > netstats: > on HostA > {noformat} > /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB > {noformat} > on HostB > {noformat} > tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from > idx:0/HostA > {noformat} > After a while, Error on HostB > {noformat} > WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - > [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error > java.lang.IllegalArgumentException: Not enough bytes > at > org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) > ~[guava-16.0.jar:na] > at > com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > [apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] > ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - > Exception in thread Thread[Thread-28,5,main] > java.lang.RuntimeException: java.lang.InterruptedException > at com.google.common.base.Throwables.propagate(Throwables.java:160) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] > Caused by: java.lang.InterruptedException: null > at > java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) > ~[na:1.8.0_66-internal] > at > org.apache.cassandra.streaming.compress.CompressedInputStream$Reader.runMayThrow(CompressedInputStream.java:176) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) > ~[apache-cassandra-2.2.4.jar:2.2.4] > ... 1 common frames omitted > ERROR [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:524
[jira] [Commented] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15083459#comment-15083459 ] xiaost commented on CASSANDRA-10961: BTW, I had ran {{nodetool scrub}} on every nodes before debugging. > Not enough bytes error when add nodes to cluster > > > Key: CASSANDRA-10961 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging >Reporter: xiaost >Assignee: Paulo Motta > Attachments: debug.1.log, debug.logs.zip, netstats.1.log > > > we got the same problem all the time when we add nodes to cluster. > netstats: > on HostA > {noformat} > /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB > {noformat} > on HostB > {noformat} > tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from > idx:0/HostA > {noformat} > After a while, Error on HostB > {noformat} > WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - > [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error > java.lang.IllegalArgumentException: Not enough bytes > at > org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) > ~[guava-16.0.jar:na] > at > com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > [apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] > ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - > Exception in thread Thread[Thread-28,5,main] > java.lang.RuntimeException: java.lang.InterruptedException > at com.google.common.base.Throwables.propagate(Throwables.java:160) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] > Caused by: java.lang.InterruptedException: null > at > java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) > ~[na:1.8.0_66-internal] > at > org.apache.cassandra.streaming.compress.CompressedInputStream$Reader.runMayThrow(CompressedInputStream.java:176) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) > ~[apache-cassandra-2.2.4.jar:2.2.4] > ... 1 common frames
[jira] [Commented] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15080995#comment-15080995 ] xiaost commented on CASSANDRA-10961: [~slebresne] the node receiving data is the added one. we only have one table in the cluster :-( I tried to add node to the cluster but failed with the errors above and I kill the process. the progress seems always sucks in streaming file which larger than 10GB (after abort 4~5 times?) I use {code}nodetool bootstrap resume{code} to resume the bootstrap, but it failed and logs like the other issue: https://issues.apache.org/jira/browse/CASSANDRA-10448 the cluster now have 3 nodes(UN status) with abort 600GB*3 data on HDD*3, and one node(UJ status) with ~1GB load > Not enough bytes error when add nodes to cluster > > > Key: CASSANDRA-10961 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging >Reporter: xiaost > > we got the same problem all the time when we add nodes to cluster. > netstats: > on HostA > {noformat} > /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB > {noformat} > on HostB > {noformat} > tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from > idx:0/HostA > {noformat} > After a while, Error on HostB > {noformat} > WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - > [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error > java.lang.IllegalArgumentException: Not enough bytes > at > org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) > ~[guava-16.0.jar:na] > at > com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > [apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] > ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - > Exception in thread Thread[Thread-28,5,main] > java.lang.RuntimeException: java.lang.InterruptedException > at com.google.common.base.Throwables.propagate(Throwables.java:160) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] > Caused by: java.lang.InterruptedException: null > at > java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) > ~[na:1.8.0_66-internal] > at >
[jira] [Commented] (CASSANDRA-10448) "Unknown type 0" Stream failure on Repair
[ https://issues.apache.org/jira/browse/CASSANDRA-10448?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15080888#comment-15080888 ] xiaost commented on CASSANDRA-10448: +1, in 2.2.4 3 nodes cluster while adding new one. > "Unknown type 0" Stream failure on Repair > - > > Key: CASSANDRA-10448 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10448 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging > Environment: Cassandra 2.2.2 > 5 Nodes in Google Compute Engine > Java 1.8.0_60 >Reporter: Omri Iluz > Fix For: 2.2.x > > Attachments: apache-cassandra-2.2.4-SNAPSHOT.jar, casslogs.txt, > receiversystem.log, sendersystem.log > > > While running repair after upgrading to 2.2.2 I am getting many stream fail > errors: > {noformat} > [2015-10-05 23:52:30,353] Repair session 4c181051-6bbb-11e5-acdb-d9a8bbd39330 > for range (59694553044959221,86389982480621619] failed with error [repair > #4c181051-6bbb-11e5-acdb-d9a8bbd39330 on px/acti > vities, (59694553044959221,86389982480621619]] Sync failed between > /10.240.81.104 and /10.240.134.221 (progress: 4%) > {noformat} > Logs from both sides of the stream: > Sides 1 - > {noformat} > INFO [STREAM-INIT-/10.240.81.104:52722] 2015-10-05 23:52:30,063 > StreamResultFuture.java:111 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550 > ID#0] Creating new streaming plan for Repair > INFO [STREAM-INIT-/10.240.81.104:52722] 2015-10-05 23:52:30,063 > StreamResultFuture.java:118 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550, > ID#0] Received streaming plan for Repair > INFO [STREAM-INIT-/10.240.81.104:52723] 2015-10-05 23:52:30,063 > StreamResultFuture.java:118 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550, > ID#0] Received streaming plan for Repair > INFO [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,098 > StreamResultFuture.java:168 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550 > ID#0] Prepare completed. Receiving 13 files(517391317 bytes), sending 10 > files(469491729 bytes) > ERROR [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,234 > StreamSession.java:524 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] > Streaming error occurred > java.lang.IllegalArgumentException: Unknown type 0 > at > org.apache.cassandra.streaming.messages.StreamMessage$Type.get(StreamMessage.java:96) > ~[apache-cassandra-2.2.2.jar:2.2.2] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:57) > ~[apache-cassandra-2.2.2.jar:2.2.2] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > ~[apache-cassandra-2.2.2.jar:2.2.2] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_60] > INFO [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,302 > StreamResultFuture.java:182 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] > Session with /10.240.81.104 is complete > WARN [STREAM-IN-/10.240.81.104] 2015-10-05 23:52:30,302 > StreamResultFuture.java:209 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] > Stream failed > {noformat} > Side 2 - > {noformat} > INFO [AntiEntropyStage:1] 2015-10-05 23:52:30,060 StreamResultFuture.java:86 > - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] Executing streaming plan for > Repair > INFO [StreamConnectionEstablisher:6] 2015-10-05 23:52:30,061 > StreamSession.java:232 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] > Starting streaming to /10.240.134.221 > INFO [StreamConnectionEstablisher:6] 2015-10-05 23:52:30,063 > StreamCoordinator.java:213 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550, > ID#0] Beginning stream session with /10.240.134.221 > INFO [STREAM-IN-/10.240.134.221] 2015-10-05 23:52:30,098 > StreamResultFuture.java:168 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550 > ID#0] Prepare completed. Receiving 10 files(469491729 bytes), sending 13 > files(517391317 bytes) > INFO [STREAM-IN-/10.240.134.221] 2015-10-05 23:52:30,349 > StreamResultFuture.java:182 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] > Session with /10.240.134.221 is complete > ERROR [STREAM-OUT-/10.240.134.221] 2015-10-05 23:52:30,349 > StreamSession.java:524 - [Stream #239d8e60-6bbc-11e5-93ac-31bdef2dc550] > Streaming error occurred > org.apache.cassandra.io.FSReadError: java.io.IOException: Broken pipe > at > org.apache.cassandra.io.util.ChannelProxy.transferTo(ChannelProxy.java:144) > ~[apache-cassandra-2.2.2.jar:2.2.2] > at > org.apache.cassandra.streaming.compress.CompressedStreamWriter$1.apply(CompressedStreamWriter.java:79) > ~[apache-cassandra-2.2.2.jar:2.2.2] > at > org.apache.cassandra.streaming.compress.CompressedStreamWriter$1.apply(CompressedStreamWriter.java:76) > ~[apache-cassandra-2.2.2.jar:2.2.2] > at >
[jira] [Updated] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] xiaost updated CASSANDRA-10961: --- Attachment: netstats.log jstack.log debug.log progress is stuck now :-( I will give more informations later > Not enough bytes error when add nodes to cluster > > > Key: CASSANDRA-10961 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging >Reporter: xiaost >Assignee: Paulo Motta > Attachments: debug.log, jstack.log, netstats.log > > > we got the same problem all the time when we add nodes to cluster. > netstats: > on HostA > {noformat} > /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB > {noformat} > on HostB > {noformat} > tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from > idx:0/HostA > {noformat} > After a while, Error on HostB > {noformat} > WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - > [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error > java.lang.IllegalArgumentException: Not enough bytes > at > org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) > ~[guava-16.0.jar:na] > at > com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > [apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] > ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - > Exception in thread Thread[Thread-28,5,main] > java.lang.RuntimeException: java.lang.InterruptedException > at com.google.common.base.Throwables.propagate(Throwables.java:160) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] > Caused by: java.lang.InterruptedException: null > at > java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) > ~[na:1.8.0_66-internal] > at > org.apache.cassandra.streaming.compress.CompressedInputStream$Reader.runMayThrow(CompressedInputStream.java:176) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) > ~[apache-cassandra-2.2.4.jar:2.2.4] >
[jira] [Updated] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] xiaost updated CASSANDRA-10961: --- Attachment: netstats.1.log debug.1.log reproduced. [~pauloricardomg] > Not enough bytes error when add nodes to cluster > > > Key: CASSANDRA-10961 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging >Reporter: xiaost >Assignee: Paulo Motta > Attachments: debug.1.log, debug.log, jstack.log, netstats.1.log, > netstats.log > > > we got the same problem all the time when we add nodes to cluster. > netstats: > on HostA > {noformat} > /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB > {noformat} > on HostB > {noformat} > tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from > idx:0/HostA > {noformat} > After a while, Error on HostB > {noformat} > WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - > [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error > java.lang.IllegalArgumentException: Not enough bytes > at > org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) > ~[guava-16.0.jar:na] > at > com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > [apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] > ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - > Exception in thread Thread[Thread-28,5,main] > java.lang.RuntimeException: java.lang.InterruptedException > at com.google.common.base.Throwables.propagate(Throwables.java:160) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] > Caused by: java.lang.InterruptedException: null > at > java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) > ~[na:1.8.0_66-internal] > at > org.apache.cassandra.streaming.compress.CompressedInputStream$Reader.runMayThrow(CompressedInputStream.java:176) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) > ~[apache-cassandra-2.2.4.jar:2.2.4] > ... 1 common
[jira] [Comment Edited] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15081315#comment-15081315 ] xiaost edited comment on CASSANDRA-10961 at 1/4/16 4:54 PM: I'm trying to fix the deadlock issue with this patch. [^0001-Fix-streaming-deadlock-when-interrupted.patch] I don't known whether it works. It is running, and will figure out. :-) was (Author: xiaost): I'm trying to fix the deadlock issue with this patch. I don't known whether it works. It is running, and will figure out. :-) > Not enough bytes error when add nodes to cluster > > > Key: CASSANDRA-10961 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging >Reporter: xiaost >Assignee: Paulo Motta > Attachments: 0001-Fix-streaming-deadlock-when-interrupted.patch, > debug.1.log, netstats.1.log > > > we got the same problem all the time when we add nodes to cluster. > netstats: > on HostA > {noformat} > /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB > {noformat} > on HostB > {noformat} > tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from > idx:0/HostA > {noformat} > After a while, Error on HostB > {noformat} > WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - > [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error > java.lang.IllegalArgumentException: Not enough bytes > at > org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) > ~[guava-16.0.jar:na] > at > com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > [apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] > ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - > Exception in thread Thread[Thread-28,5,main] > java.lang.RuntimeException: java.lang.InterruptedException > at com.google.common.base.Throwables.propagate(Throwables.java:160) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] > Caused by: java.lang.InterruptedException: null > at > java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) >
[jira] [Updated] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] xiaost updated CASSANDRA-10961: --- Attachment: 0001-Fix-streaming-deadlock-when-interrupted.patch I'm trying to fix the deadlock issue with this patch. I don't known whether it works. It is running, and will figure out. :-) > Not enough bytes error when add nodes to cluster > > > Key: CASSANDRA-10961 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging >Reporter: xiaost >Assignee: Paulo Motta > Attachments: 0001-Fix-streaming-deadlock-when-interrupted.patch, > debug.1.log, debug.log, jstack.log, netstats.1.log, netstats.log > > > we got the same problem all the time when we add nodes to cluster. > netstats: > on HostA > {noformat} > /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB > {noformat} > on HostB > {noformat} > tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from > idx:0/HostA > {noformat} > After a while, Error on HostB > {noformat} > WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - > [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error > java.lang.IllegalArgumentException: Not enough bytes > at > org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) > ~[guava-16.0.jar:na] > at > com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > [apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] > ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - > Exception in thread Thread[Thread-28,5,main] > java.lang.RuntimeException: java.lang.InterruptedException > at com.google.common.base.Throwables.propagate(Throwables.java:160) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] > Caused by: java.lang.InterruptedException: null > at > java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) > ~[na:1.8.0_66-internal] > at > org.apache.cassandra.streaming.compress.CompressedInputStream$Reader.runMayThrow(CompressedInputStream.java:176) >
[jira] [Comment Edited] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15081243#comment-15081243 ] xiaost edited comment on CASSANDRA-10961 at 1/4/16 4:55 PM: reproduced. [~pauloricardomg] [^debug.1.log] [^netstats.1.log] was (Author: xiaost): reproduced. [~pauloricardomg] > Not enough bytes error when add nodes to cluster > > > Key: CASSANDRA-10961 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging >Reporter: xiaost >Assignee: Paulo Motta > Attachments: 0001-Fix-streaming-deadlock-when-interrupted.patch, > debug.1.log, netstats.1.log > > > we got the same problem all the time when we add nodes to cluster. > netstats: > on HostA > {noformat} > /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB > {noformat} > on HostB > {noformat} > tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from > idx:0/HostA > {noformat} > After a while, Error on HostB > {noformat} > WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - > [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error > java.lang.IllegalArgumentException: Not enough bytes > at > org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) > ~[guava-16.0.jar:na] > at > com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > [apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] > ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - > Exception in thread Thread[Thread-28,5,main] > java.lang.RuntimeException: java.lang.InterruptedException > at com.google.common.base.Throwables.propagate(Throwables.java:160) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] > Caused by: java.lang.InterruptedException: null > at > java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) > ~[na:1.8.0_66-internal] > at > org.apache.cassandra.streaming.compress.CompressedInputStream$Reader.runMayThrow(CompressedInputStream.java:176) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at >
[jira] [Updated] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] xiaost updated CASSANDRA-10961: --- Attachment: (was: 0001-Fix-streaming-deadlock-when-interrupted.patch) > Not enough bytes error when add nodes to cluster > > > Key: CASSANDRA-10961 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging >Reporter: xiaost >Assignee: Paulo Motta > Attachments: debug.1.log, netstats.1.log > > > we got the same problem all the time when we add nodes to cluster. > netstats: > on HostA > {noformat} > /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB > {noformat} > on HostB > {noformat} > tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from > idx:0/HostA > {noformat} > After a while, Error on HostB > {noformat} > WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - > [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error > java.lang.IllegalArgumentException: Not enough bytes > at > org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) > ~[guava-16.0.jar:na] > at > com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > [apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] > ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - > Exception in thread Thread[Thread-28,5,main] > java.lang.RuntimeException: java.lang.InterruptedException > at com.google.common.base.Throwables.propagate(Throwables.java:160) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] > Caused by: java.lang.InterruptedException: null > at > java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) > ~[na:1.8.0_66-internal] > at > org.apache.cassandra.streaming.compress.CompressedInputStream$Reader.runMayThrow(CompressedInputStream.java:176) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) > ~[apache-cassandra-2.2.4.jar:2.2.4] > ... 1 common frames omitted > ERROR [STREAM-IN-/HostA] 2016-01-02
[jira] [Issue Comment Deleted] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] xiaost updated CASSANDRA-10961: --- Comment: was deleted (was: I'm trying to fix the deadlock issue with this patch. [^0001-Fix-streaming-deadlock-when-interrupted.patch] I don't known whether it works. It is running, and will figure out. :-)) > Not enough bytes error when add nodes to cluster > > > Key: CASSANDRA-10961 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging >Reporter: xiaost >Assignee: Paulo Motta > Attachments: 0001-Fix-streaming-deadlock-when-interrupted.patch, > debug.1.log, netstats.1.log > > > we got the same problem all the time when we add nodes to cluster. > netstats: > on HostA > {noformat} > /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB > {noformat} > on HostB > {noformat} > tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from > idx:0/HostA > {noformat} > After a while, Error on HostB > {noformat} > WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - > [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error > java.lang.IllegalArgumentException: Not enough bytes > at > org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) > ~[guava-16.0.jar:na] > at > com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > [apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] > ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - > Exception in thread Thread[Thread-28,5,main] > java.lang.RuntimeException: java.lang.InterruptedException > at com.google.common.base.Throwables.propagate(Throwables.java:160) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] > Caused by: java.lang.InterruptedException: null > at > java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) > ~[na:1.8.0_66-internal] > at > org.apache.cassandra.streaming.compress.CompressedInputStream$Reader.runMayThrow(CompressedInputStream.java:176) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at >
[jira] [Updated] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] xiaost updated CASSANDRA-10961: --- Attachment: (was: debug.log) > Not enough bytes error when add nodes to cluster > > > Key: CASSANDRA-10961 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging >Reporter: xiaost >Assignee: Paulo Motta > Attachments: 0001-Fix-streaming-deadlock-when-interrupted.patch, > debug.1.log, netstats.1.log > > > we got the same problem all the time when we add nodes to cluster. > netstats: > on HostA > {noformat} > /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB > {noformat} > on HostB > {noformat} > tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from > idx:0/HostA > {noformat} > After a while, Error on HostB > {noformat} > WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - > [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error > java.lang.IllegalArgumentException: Not enough bytes > at > org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) > ~[guava-16.0.jar:na] > at > com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > [apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] > ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - > Exception in thread Thread[Thread-28,5,main] > java.lang.RuntimeException: java.lang.InterruptedException > at com.google.common.base.Throwables.propagate(Throwables.java:160) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] > Caused by: java.lang.InterruptedException: null > at > java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) > ~[na:1.8.0_66-internal] > at > org.apache.cassandra.streaming.compress.CompressedInputStream$Reader.runMayThrow(CompressedInputStream.java:176) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) > ~[apache-cassandra-2.2.4.jar:2.2.4] > ... 1 common frames omitted > ERROR [STREAM-IN-/HostA]
[jira] [Updated] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] xiaost updated CASSANDRA-10961: --- Attachment: (was: jstack.log) > Not enough bytes error when add nodes to cluster > > > Key: CASSANDRA-10961 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging >Reporter: xiaost >Assignee: Paulo Motta > Attachments: 0001-Fix-streaming-deadlock-when-interrupted.patch, > debug.1.log, netstats.1.log > > > we got the same problem all the time when we add nodes to cluster. > netstats: > on HostA > {noformat} > /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB > {noformat} > on HostB > {noformat} > tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from > idx:0/HostA > {noformat} > After a while, Error on HostB > {noformat} > WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - > [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error > java.lang.IllegalArgumentException: Not enough bytes > at > org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) > ~[guava-16.0.jar:na] > at > com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > [apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] > ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - > Exception in thread Thread[Thread-28,5,main] > java.lang.RuntimeException: java.lang.InterruptedException > at com.google.common.base.Throwables.propagate(Throwables.java:160) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] > Caused by: java.lang.InterruptedException: null > at > java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) > ~[na:1.8.0_66-internal] > at > org.apache.cassandra.streaming.compress.CompressedInputStream$Reader.runMayThrow(CompressedInputStream.java:176) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) > ~[apache-cassandra-2.2.4.jar:2.2.4] > ... 1 common frames omitted > ERROR [STREAM-IN-/HostA]
[jira] [Updated] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] xiaost updated CASSANDRA-10961: --- Attachment: (was: netstats.log) > Not enough bytes error when add nodes to cluster > > > Key: CASSANDRA-10961 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 > Project: Cassandra > Issue Type: Bug > Components: Streaming and Messaging >Reporter: xiaost >Assignee: Paulo Motta > Attachments: 0001-Fix-streaming-deadlock-when-interrupted.patch, > debug.1.log, netstats.1.log > > > we got the same problem all the time when we add nodes to cluster. > netstats: > on HostA > {noformat} > /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB > {noformat} > on HostB > {noformat} > tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from > idx:0/HostA > {noformat} > After a while, Error on HostB > {noformat} > WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - > [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error > java.lang.IllegalArgumentException: Not enough bytes > at > org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) > ~[guava-16.0.jar:na] > at > com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) > [apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) > [apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] > ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - > Exception in thread Thread[Thread-28,5,main] > java.lang.RuntimeException: java.lang.InterruptedException > at com.google.common.base.Throwables.propagate(Throwables.java:160) > ~[guava-16.0.jar:na] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] > Caused by: java.lang.InterruptedException: null > at > java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) > ~[na:1.8.0_66-internal] > at > java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) > ~[na:1.8.0_66-internal] > at > org.apache.cassandra.streaming.compress.CompressedInputStream$Reader.runMayThrow(CompressedInputStream.java:176) > ~[apache-cassandra-2.2.4.jar:2.2.4] > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) > ~[apache-cassandra-2.2.4.jar:2.2.4] > ... 1 common frames omitted > ERROR
[jira] [Updated] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
[ https://issues.apache.org/jira/browse/CASSANDRA-10961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] xiaost updated CASSANDRA-10961: --- Description: we got the same problem all the time when we add nodes to cluster. netstats: on HostA {noformat} /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB {noformat} on HostB {noformat} tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from idx:0/HostA {noformat} After a while, Error on HostB {noformat} WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error java.lang.IllegalArgumentException: Not enough bytes at org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) ~[apache-cassandra-2.2.4.jar:2.2.4] at com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) ~[guava-16.0.jar:na] at com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) ~[guava-16.0.jar:na] at org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) [apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) [apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) [apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) [apache-cassandra-2.2.4.jar:2.2.4] at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - Exception in thread Thread[Thread-28,5,main] java.lang.RuntimeException: java.lang.InterruptedException at com.google.common.base.Throwables.propagate(Throwables.java:160) ~[guava-16.0.jar:na] at org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) ~[apache-cassandra-2.2.4.jar:2.2.4] at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] Caused by: java.lang.InterruptedException: null at java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) ~[na:1.8.0_66-internal] at java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) ~[na:1.8.0_66-internal] at java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) ~[na:1.8.0_66-internal] at org.apache.cassandra.streaming.compress.CompressedInputStream$Reader.runMayThrow(CompressedInputStream.java:176) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) ~[apache-cassandra-2.2.4.jar:2.2.4] ... 1 common frames omitted ERROR [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:524 - [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Streaming error occurred java.nio.channels.ClosedChannelException: null at sun.nio.ch.SocketChannelImpl.ensureReadOpen(SocketChannelImpl.java:257) ~[na:1.8.0_66-internal] at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:300) ~[na:1.8.0_66-internal] at org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:53) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) ~[apache-cassandra-2.2.4.jar:2.2.4] at
[jira] [Created] (CASSANDRA-10961) Not enough bytes error when add nodes to cluster
xiaost created CASSANDRA-10961: -- Summary: Not enough bytes error when add nodes to cluster Key: CASSANDRA-10961 URL: https://issues.apache.org/jira/browse/CASSANDRA-10961 Project: Cassandra Issue Type: Bug Components: Streaming and Messaging Reporter: xiaost we got the same problem all the time when we add nodes to cluster. netstats: on HostA {noformat} /la-38395-big-Data.db 14792091851/14792091851 bytes(100%) sent to idx:0/HostB {noformat} on HostB {noformat} tmp-la-4-big-Data.db 2667087450/14792091851 bytes(18%) received from idx:0/HostA {noformat} After a while, Error on HostB {noformat} WARN [STREAM-IN-/HostA] 2016-01-02 12:08:14,737 StreamSession.java:644 - [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Retrying for following error java.lang.IllegalArgumentException: Not enough bytes at org.apache.cassandra.db.composites.AbstractCType.checkRemaining(AbstractCType.java:362) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.db.composites.AbstractCompoundCellNameType.fromByteBuffer(AbstractCompoundCellNameType.java:98) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:381) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.db.composites.AbstractCType$Serializer.deserialize(AbstractCType.java:365) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:75) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:52) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.db.AbstractCell$1.computeNext(AbstractCell.java:46) ~[apache-cassandra-2.2.4.jar:2.2.4] at com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) ~[guava-16.0.jar:na] at com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) ~[guava-16.0.jar:na] at org.apache.cassandra.io.sstable.format.big.BigTableWriter.appendFromStream(BigTableWriter.java:243) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.streaming.StreamReader.writeRow(StreamReader.java:173) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.streaming.compress.CompressedStreamReader.read(CompressedStreamReader.java:95) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:49) [apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.streaming.messages.IncomingFileMessage$1.deserialize(IncomingFileMessage.java:38) [apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:58) [apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.streaming.ConnectionHandler$IncomingMessageHandler.run(ConnectionHandler.java:261) [apache-cassandra-2.2.4.jar:2.2.4] at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal] ERROR [Thread-28] 2016-01-02 12:08:14,737 CassandraDaemon.java:185 - Exception in thread Thread[Thread-28,5,main] java.lang.RuntimeException: java.lang.InterruptedException at com.google.common.base.Throwables.propagate(Throwables.java:160) ~[guava-16.0.jar:na] at org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) ~[apache-cassandra-2.2.4.jar:2.2.4] at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_66-internal] Caused by: java.lang.InterruptedException: null at java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) ~[na:1.8.0_66-internal] at java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) ~[na:1.8.0_66-internal] at java.util.concurrent.ArrayBlockingQueue.put(ArrayBlockingQueue.java:350) ~[na:1.8.0_66-internal] at org.apache.cassandra.streaming.compress.CompressedInputStream$Reader.runMayThrow(CompressedInputStream.java:176) ~[apache-cassandra-2.2.4.jar:2.2.4] at org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) ~[apache-cassandra-2.2.4.jar:2.2.4] ... 1 common frames omitted ERROR [STREAM-IN-/10.6.96.12] 2016-01-02 12:08:14,737 StreamSession.java:524 - [Stream #b91a4e90-b105-11e5-bd57-dd0cc3b4634c] Streaming error occurred java.nio.channels.ClosedChannelException: null at sun.nio.ch.SocketChannelImpl.ensureReadOpen(SocketChannelImpl.java:257) ~[na:1.8.0_66-internal] at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:300) ~[na:1.8.0_66-internal] at org.apache.cassandra.streaming.messages.StreamMessage.deserialize(StreamMessage.java:53) ~[apache-cassandra-2.2.4.jar:2.2.4] at