[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-12-05 Thread Michael Semb Wever (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Michael Semb Wever updated CASSANDRA-18963:
---
Fix Version/s: (was: 5.0-beta)

> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Legacy/Local Write-Read Paths
>Reporter: Berenguer Blasi
>Assignee: Jacek Lewandowski
>Priority: Normal
> Fix For: 4.0.12, 4.1.4, 5.0-beta1, 5.0, 5.1
>
>
> Seems flaky 
> [https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
>  * 
> h4. test_abortedcompaction
> sstableutil_test.TestSSTableUtil 
>  
> {code:java}
> self =  def 
> test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
> CASSANDRA-11497 Check that we can cleanup temporary files after a compaction 
> is aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
> cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
> self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
> tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
> len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
> len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, 
> TableName, filename=log_file_name, delay=2) t.start() try: 
> logger.debug("Compacting...") node.compact() except ToolError: pass # 
> expected to fail t.join() finalfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
> type='final')) tmpfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='tmp')) 
> # In most cases we should end up with some temporary files to clean up, but 
> it may happen # that no temporary files are created if compaction finishes 
> too early or starts too late # see CASSANDRA-11497 logger.debug("Got {} final 
> files and {} tmp files after compaction was interrupted" 
> .format(len(finalfiles), len(tmpfiles))) 
> self._invoke_sstableutil(KeyspaceName, TableName, cleanup=True) 
> self._check_files(node, KeyspaceName, TableName, finalfiles, []) # restart to 
> make sure not data is lost logger.debug("Restarting node...") > 
> node.start(wait_for_binary_proto=True) sstableutil_test.py:97: _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
> self.wait_for_binary_interface(from_mark=self.mark) 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
> wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
> clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
> in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> self =  def 
> raise_node_error_if_cassandra_process_is_terminated(self): if not 
> self._is_pid_running(): msg = "C* process with {pid} is 
> terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
> ccmlib.node.NodeError: C* process with 19530 is terminated 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}
> {code:java}
> failed on teardown with "Unexpected error found in node logs (see stdout for 
> full details). Errors: [[node1] 'ERROR [SSTableBatchOpen:1] 2023-10-19 
> 22:25:07,449 DefaultFSErrorHandler.java:129 - Exiting forcefully due to file 
> system exception on startup, disk failure policy 
> "stop"\norg.apache.cassandra.io.sstable.CorruptSSTableException: Corrupted: 
> /tmp/dtest-ahz16wrh/test/node1/data0/system/prepared_statements-18a9c2576a0c3841ba718cd529849fef/nc-4-big\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReaderLoadingBuilder.build(SSTableReaderLoadingBuilder.java:111)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:397)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:353)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.lambda$openAll$4(SSTableReader.java:414)\n\tat
>  org.apache.cassandra.concurrent.FutureTask$1.call(FutureTask.java:96)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.call(FutureTask.java:61)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.run(FutureTask.java:71)\n\tat 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\n\tat
>  
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\n\tat
>  
> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)\n\tat
>  

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-12-05 Thread Michael Semb Wever (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Michael Semb Wever updated CASSANDRA-18963:
---
Fix Version/s: 5.0
   5.0-beta1

> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Legacy/Local Write-Read Paths
>Reporter: Berenguer Blasi
>Assignee: Jacek Lewandowski
>Priority: Normal
> Fix For: 4.0.12, 4.1.4, 5.0-beta1, 5.0-beta, 5.0, 5.1
>
>
> Seems flaky 
> [https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
>  * 
> h4. test_abortedcompaction
> sstableutil_test.TestSSTableUtil 
>  
> {code:java}
> self =  def 
> test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
> CASSANDRA-11497 Check that we can cleanup temporary files after a compaction 
> is aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
> cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
> self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
> tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
> len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
> len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, 
> TableName, filename=log_file_name, delay=2) t.start() try: 
> logger.debug("Compacting...") node.compact() except ToolError: pass # 
> expected to fail t.join() finalfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
> type='final')) tmpfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='tmp')) 
> # In most cases we should end up with some temporary files to clean up, but 
> it may happen # that no temporary files are created if compaction finishes 
> too early or starts too late # see CASSANDRA-11497 logger.debug("Got {} final 
> files and {} tmp files after compaction was interrupted" 
> .format(len(finalfiles), len(tmpfiles))) 
> self._invoke_sstableutil(KeyspaceName, TableName, cleanup=True) 
> self._check_files(node, KeyspaceName, TableName, finalfiles, []) # restart to 
> make sure not data is lost logger.debug("Restarting node...") > 
> node.start(wait_for_binary_proto=True) sstableutil_test.py:97: _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
> self.wait_for_binary_interface(from_mark=self.mark) 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
> wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
> clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
> in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> self =  def 
> raise_node_error_if_cassandra_process_is_terminated(self): if not 
> self._is_pid_running(): msg = "C* process with {pid} is 
> terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
> ccmlib.node.NodeError: C* process with 19530 is terminated 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}
> {code:java}
> failed on teardown with "Unexpected error found in node logs (see stdout for 
> full details). Errors: [[node1] 'ERROR [SSTableBatchOpen:1] 2023-10-19 
> 22:25:07,449 DefaultFSErrorHandler.java:129 - Exiting forcefully due to file 
> system exception on startup, disk failure policy 
> "stop"\norg.apache.cassandra.io.sstable.CorruptSSTableException: Corrupted: 
> /tmp/dtest-ahz16wrh/test/node1/data0/system/prepared_statements-18a9c2576a0c3841ba718cd529849fef/nc-4-big\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReaderLoadingBuilder.build(SSTableReaderLoadingBuilder.java:111)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:397)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:353)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.lambda$openAll$4(SSTableReader.java:414)\n\tat
>  org.apache.cassandra.concurrent.FutureTask$1.call(FutureTask.java:96)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.call(FutureTask.java:61)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.run(FutureTask.java:71)\n\tat 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\n\tat
>  
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\n\tat
>  
> 

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-11-28 Thread Jacek Lewandowski (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jacek Lewandowski updated CASSANDRA-18963:
--
Reviewers: Benjamin Lerer  (was: Benjamin Lerer, Jacek Lewandowski)

> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Legacy/Local Write-Read Paths
>Reporter: Berenguer Blasi
>Assignee: Jacek Lewandowski
>Priority: Normal
> Fix For: 4.0.12, 4.1.4, 5.0-beta, 5.1
>
>
> Seems flaky 
> [https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
>  * 
> h4. test_abortedcompaction
> sstableutil_test.TestSSTableUtil 
>  
> {code:java}
> self =  def 
> test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
> CASSANDRA-11497 Check that we can cleanup temporary files after a compaction 
> is aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
> cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
> self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
> tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
> len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
> len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, 
> TableName, filename=log_file_name, delay=2) t.start() try: 
> logger.debug("Compacting...") node.compact() except ToolError: pass # 
> expected to fail t.join() finalfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
> type='final')) tmpfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='tmp')) 
> # In most cases we should end up with some temporary files to clean up, but 
> it may happen # that no temporary files are created if compaction finishes 
> too early or starts too late # see CASSANDRA-11497 logger.debug("Got {} final 
> files and {} tmp files after compaction was interrupted" 
> .format(len(finalfiles), len(tmpfiles))) 
> self._invoke_sstableutil(KeyspaceName, TableName, cleanup=True) 
> self._check_files(node, KeyspaceName, TableName, finalfiles, []) # restart to 
> make sure not data is lost logger.debug("Restarting node...") > 
> node.start(wait_for_binary_proto=True) sstableutil_test.py:97: _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
> self.wait_for_binary_interface(from_mark=self.mark) 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
> wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
> clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
> in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> self =  def 
> raise_node_error_if_cassandra_process_is_terminated(self): if not 
> self._is_pid_running(): msg = "C* process with {pid} is 
> terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
> ccmlib.node.NodeError: C* process with 19530 is terminated 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}
> {code:java}
> failed on teardown with "Unexpected error found in node logs (see stdout for 
> full details). Errors: [[node1] 'ERROR [SSTableBatchOpen:1] 2023-10-19 
> 22:25:07,449 DefaultFSErrorHandler.java:129 - Exiting forcefully due to file 
> system exception on startup, disk failure policy 
> "stop"\norg.apache.cassandra.io.sstable.CorruptSSTableException: Corrupted: 
> /tmp/dtest-ahz16wrh/test/node1/data0/system/prepared_statements-18a9c2576a0c3841ba718cd529849fef/nc-4-big\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReaderLoadingBuilder.build(SSTableReaderLoadingBuilder.java:111)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:397)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:353)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.lambda$openAll$4(SSTableReader.java:414)\n\tat
>  org.apache.cassandra.concurrent.FutureTask$1.call(FutureTask.java:96)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.call(FutureTask.java:61)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.run(FutureTask.java:71)\n\tat 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\n\tat
>  
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\n\tat
>  
> 

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-11-28 Thread Jacek Lewandowski (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jacek Lewandowski updated CASSANDRA-18963:
--
  Fix Version/s: 4.0.12
 4.1.4
 5.0-beta
 5.1
 (was: 5.x)
 (was: 4.0.x)
 (was: 4.1.x)
 (was: 5.0-rc)
  Since Version: 4.0.0
Source Control Link: 
https://github.com/apache/cassandra/commit/dece96f21dbdbbb3176d6544d72bef3d44571dc6
 Resolution: Fixed
 Status: Resolved  (was: Ready to Commit)

> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Legacy/Local Write-Read Paths
>Reporter: Berenguer Blasi
>Assignee: Jacek Lewandowski
>Priority: Normal
> Fix For: 4.0.12, 4.1.4, 5.0-beta, 5.1
>
>
> Seems flaky 
> [https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
>  * 
> h4. test_abortedcompaction
> sstableutil_test.TestSSTableUtil 
>  
> {code:java}
> self =  def 
> test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
> CASSANDRA-11497 Check that we can cleanup temporary files after a compaction 
> is aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
> cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
> self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
> tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
> len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
> len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, 
> TableName, filename=log_file_name, delay=2) t.start() try: 
> logger.debug("Compacting...") node.compact() except ToolError: pass # 
> expected to fail t.join() finalfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
> type='final')) tmpfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='tmp')) 
> # In most cases we should end up with some temporary files to clean up, but 
> it may happen # that no temporary files are created if compaction finishes 
> too early or starts too late # see CASSANDRA-11497 logger.debug("Got {} final 
> files and {} tmp files after compaction was interrupted" 
> .format(len(finalfiles), len(tmpfiles))) 
> self._invoke_sstableutil(KeyspaceName, TableName, cleanup=True) 
> self._check_files(node, KeyspaceName, TableName, finalfiles, []) # restart to 
> make sure not data is lost logger.debug("Restarting node...") > 
> node.start(wait_for_binary_proto=True) sstableutil_test.py:97: _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
> self.wait_for_binary_interface(from_mark=self.mark) 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
> wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
> clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
> in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> self =  def 
> raise_node_error_if_cassandra_process_is_terminated(self): if not 
> self._is_pid_running(): msg = "C* process with {pid} is 
> terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
> ccmlib.node.NodeError: C* process with 19530 is terminated 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}
> {code:java}
> failed on teardown with "Unexpected error found in node logs (see stdout for 
> full details). Errors: [[node1] 'ERROR [SSTableBatchOpen:1] 2023-10-19 
> 22:25:07,449 DefaultFSErrorHandler.java:129 - Exiting forcefully due to file 
> system exception on startup, disk failure policy 
> "stop"\norg.apache.cassandra.io.sstable.CorruptSSTableException: Corrupted: 
> /tmp/dtest-ahz16wrh/test/node1/data0/system/prepared_statements-18a9c2576a0c3841ba718cd529849fef/nc-4-big\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReaderLoadingBuilder.build(SSTableReaderLoadingBuilder.java:111)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:397)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:353)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.lambda$openAll$4(SSTableReader.java:414)\n\tat
>  

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-11-28 Thread Jacek Lewandowski (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jacek Lewandowski updated CASSANDRA-18963:
--
Status: Ready to Commit  (was: Review In Progress)

> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Legacy/Local Write-Read Paths
>Reporter: Berenguer Blasi
>Assignee: Jacek Lewandowski
>Priority: Normal
> Fix For: 4.0.x, 4.1.x, 5.0-rc, 5.x
>
>
> Seems flaky 
> [https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
>  * 
> h4. test_abortedcompaction
> sstableutil_test.TestSSTableUtil 
>  
> {code:java}
> self =  def 
> test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
> CASSANDRA-11497 Check that we can cleanup temporary files after a compaction 
> is aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
> cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
> self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
> tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
> len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
> len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, 
> TableName, filename=log_file_name, delay=2) t.start() try: 
> logger.debug("Compacting...") node.compact() except ToolError: pass # 
> expected to fail t.join() finalfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
> type='final')) tmpfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='tmp')) 
> # In most cases we should end up with some temporary files to clean up, but 
> it may happen # that no temporary files are created if compaction finishes 
> too early or starts too late # see CASSANDRA-11497 logger.debug("Got {} final 
> files and {} tmp files after compaction was interrupted" 
> .format(len(finalfiles), len(tmpfiles))) 
> self._invoke_sstableutil(KeyspaceName, TableName, cleanup=True) 
> self._check_files(node, KeyspaceName, TableName, finalfiles, []) # restart to 
> make sure not data is lost logger.debug("Restarting node...") > 
> node.start(wait_for_binary_proto=True) sstableutil_test.py:97: _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
> self.wait_for_binary_interface(from_mark=self.mark) 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
> wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
> clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
> in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> self =  def 
> raise_node_error_if_cassandra_process_is_terminated(self): if not 
> self._is_pid_running(): msg = "C* process with {pid} is 
> terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
> ccmlib.node.NodeError: C* process with 19530 is terminated 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}
> {code:java}
> failed on teardown with "Unexpected error found in node logs (see stdout for 
> full details). Errors: [[node1] 'ERROR [SSTableBatchOpen:1] 2023-10-19 
> 22:25:07,449 DefaultFSErrorHandler.java:129 - Exiting forcefully due to file 
> system exception on startup, disk failure policy 
> "stop"\norg.apache.cassandra.io.sstable.CorruptSSTableException: Corrupted: 
> /tmp/dtest-ahz16wrh/test/node1/data0/system/prepared_statements-18a9c2576a0c3841ba718cd529849fef/nc-4-big\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReaderLoadingBuilder.build(SSTableReaderLoadingBuilder.java:111)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:397)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:353)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.lambda$openAll$4(SSTableReader.java:414)\n\tat
>  org.apache.cassandra.concurrent.FutureTask$1.call(FutureTask.java:96)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.call(FutureTask.java:61)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.run(FutureTask.java:71)\n\tat 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\n\tat
>  
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\n\tat
>  
> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)\n\tat
>  

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-11-24 Thread Michael Semb Wever (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Michael Semb Wever updated CASSANDRA-18963:
---
Fix Version/s: 5.x
   (was: 5.1)

> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Legacy/Local Write-Read Paths
>Reporter: Berenguer Blasi
>Assignee: Jacek Lewandowski
>Priority: Normal
> Fix For: 4.0.x, 4.1.x, 5.0-rc, 5.x
>
>
> Seems flaky 
> [https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
>  * 
> h4. test_abortedcompaction
> sstableutil_test.TestSSTableUtil 
>  
> {code:java}
> self =  def 
> test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
> CASSANDRA-11497 Check that we can cleanup temporary files after a compaction 
> is aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
> cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
> self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
> tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
> len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
> len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, 
> TableName, filename=log_file_name, delay=2) t.start() try: 
> logger.debug("Compacting...") node.compact() except ToolError: pass # 
> expected to fail t.join() finalfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
> type='final')) tmpfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='tmp')) 
> # In most cases we should end up with some temporary files to clean up, but 
> it may happen # that no temporary files are created if compaction finishes 
> too early or starts too late # see CASSANDRA-11497 logger.debug("Got {} final 
> files and {} tmp files after compaction was interrupted" 
> .format(len(finalfiles), len(tmpfiles))) 
> self._invoke_sstableutil(KeyspaceName, TableName, cleanup=True) 
> self._check_files(node, KeyspaceName, TableName, finalfiles, []) # restart to 
> make sure not data is lost logger.debug("Restarting node...") > 
> node.start(wait_for_binary_proto=True) sstableutil_test.py:97: _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
> self.wait_for_binary_interface(from_mark=self.mark) 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
> wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
> clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
> in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> self =  def 
> raise_node_error_if_cassandra_process_is_terminated(self): if not 
> self._is_pid_running(): msg = "C* process with {pid} is 
> terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
> ccmlib.node.NodeError: C* process with 19530 is terminated 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}
> {code:java}
> failed on teardown with "Unexpected error found in node logs (see stdout for 
> full details). Errors: [[node1] 'ERROR [SSTableBatchOpen:1] 2023-10-19 
> 22:25:07,449 DefaultFSErrorHandler.java:129 - Exiting forcefully due to file 
> system exception on startup, disk failure policy 
> "stop"\norg.apache.cassandra.io.sstable.CorruptSSTableException: Corrupted: 
> /tmp/dtest-ahz16wrh/test/node1/data0/system/prepared_statements-18a9c2576a0c3841ba718cd529849fef/nc-4-big\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReaderLoadingBuilder.build(SSTableReaderLoadingBuilder.java:111)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:397)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:353)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.lambda$openAll$4(SSTableReader.java:414)\n\tat
>  org.apache.cassandra.concurrent.FutureTask$1.call(FutureTask.java:96)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.call(FutureTask.java:61)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.run(FutureTask.java:71)\n\tat 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\n\tat
>  
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\n\tat
>  
> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)\n\tat
>  

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-11-22 Thread Jacek Lewandowski (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jacek Lewandowski updated CASSANDRA-18963:
--
Reviewers: Benjamin Lerer, Jacek Lewandowski
   Status: Review In Progress  (was: Patch Available)

> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Legacy/Local Write-Read Paths
>Reporter: Berenguer Blasi
>Assignee: Jacek Lewandowski
>Priority: Normal
> Fix For: 4.0.x, 4.1.x, 5.0-rc, 5.1
>
>
> Seems flaky 
> [https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
>  * 
> h4. test_abortedcompaction
> sstableutil_test.TestSSTableUtil 
>  
> {code:java}
> self =  def 
> test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
> CASSANDRA-11497 Check that we can cleanup temporary files after a compaction 
> is aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
> cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
> self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
> tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
> len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
> len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, 
> TableName, filename=log_file_name, delay=2) t.start() try: 
> logger.debug("Compacting...") node.compact() except ToolError: pass # 
> expected to fail t.join() finalfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
> type='final')) tmpfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='tmp')) 
> # In most cases we should end up with some temporary files to clean up, but 
> it may happen # that no temporary files are created if compaction finishes 
> too early or starts too late # see CASSANDRA-11497 logger.debug("Got {} final 
> files and {} tmp files after compaction was interrupted" 
> .format(len(finalfiles), len(tmpfiles))) 
> self._invoke_sstableutil(KeyspaceName, TableName, cleanup=True) 
> self._check_files(node, KeyspaceName, TableName, finalfiles, []) # restart to 
> make sure not data is lost logger.debug("Restarting node...") > 
> node.start(wait_for_binary_proto=True) sstableutil_test.py:97: _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
> self.wait_for_binary_interface(from_mark=self.mark) 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
> wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
> clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
> in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> self =  def 
> raise_node_error_if_cassandra_process_is_terminated(self): if not 
> self._is_pid_running(): msg = "C* process with {pid} is 
> terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
> ccmlib.node.NodeError: C* process with 19530 is terminated 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}
> {code:java}
> failed on teardown with "Unexpected error found in node logs (see stdout for 
> full details). Errors: [[node1] 'ERROR [SSTableBatchOpen:1] 2023-10-19 
> 22:25:07,449 DefaultFSErrorHandler.java:129 - Exiting forcefully due to file 
> system exception on startup, disk failure policy 
> "stop"\norg.apache.cassandra.io.sstable.CorruptSSTableException: Corrupted: 
> /tmp/dtest-ahz16wrh/test/node1/data0/system/prepared_statements-18a9c2576a0c3841ba718cd529849fef/nc-4-big\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReaderLoadingBuilder.build(SSTableReaderLoadingBuilder.java:111)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:397)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:353)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.lambda$openAll$4(SSTableReader.java:414)\n\tat
>  org.apache.cassandra.concurrent.FutureTask$1.call(FutureTask.java:96)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.call(FutureTask.java:61)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.run(FutureTask.java:71)\n\tat 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\n\tat
>  
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\n\tat
>  
> 

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-11-22 Thread Jacek Lewandowski (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jacek Lewandowski updated CASSANDRA-18963:
--
Test and Documentation Plan: run regressions and new test
 Status: Patch Available  (was: In Progress)

> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Legacy/Local Write-Read Paths
>Reporter: Berenguer Blasi
>Assignee: Jacek Lewandowski
>Priority: Normal
> Fix For: 4.0.x, 4.1.x, 5.0-rc, 5.1
>
>
> Seems flaky 
> [https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
>  * 
> h4. test_abortedcompaction
> sstableutil_test.TestSSTableUtil 
>  
> {code:java}
> self =  def 
> test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
> CASSANDRA-11497 Check that we can cleanup temporary files after a compaction 
> is aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
> cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
> self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
> tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
> len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
> len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, 
> TableName, filename=log_file_name, delay=2) t.start() try: 
> logger.debug("Compacting...") node.compact() except ToolError: pass # 
> expected to fail t.join() finalfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
> type='final')) tmpfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='tmp')) 
> # In most cases we should end up with some temporary files to clean up, but 
> it may happen # that no temporary files are created if compaction finishes 
> too early or starts too late # see CASSANDRA-11497 logger.debug("Got {} final 
> files and {} tmp files after compaction was interrupted" 
> .format(len(finalfiles), len(tmpfiles))) 
> self._invoke_sstableutil(KeyspaceName, TableName, cleanup=True) 
> self._check_files(node, KeyspaceName, TableName, finalfiles, []) # restart to 
> make sure not data is lost logger.debug("Restarting node...") > 
> node.start(wait_for_binary_proto=True) sstableutil_test.py:97: _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
> self.wait_for_binary_interface(from_mark=self.mark) 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
> wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
> clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
> in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> self =  def 
> raise_node_error_if_cassandra_process_is_terminated(self): if not 
> self._is_pid_running(): msg = "C* process with {pid} is 
> terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
> ccmlib.node.NodeError: C* process with 19530 is terminated 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}
> {code:java}
> failed on teardown with "Unexpected error found in node logs (see stdout for 
> full details). Errors: [[node1] 'ERROR [SSTableBatchOpen:1] 2023-10-19 
> 22:25:07,449 DefaultFSErrorHandler.java:129 - Exiting forcefully due to file 
> system exception on startup, disk failure policy 
> "stop"\norg.apache.cassandra.io.sstable.CorruptSSTableException: Corrupted: 
> /tmp/dtest-ahz16wrh/test/node1/data0/system/prepared_statements-18a9c2576a0c3841ba718cd529849fef/nc-4-big\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReaderLoadingBuilder.build(SSTableReaderLoadingBuilder.java:111)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:397)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:353)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.lambda$openAll$4(SSTableReader.java:414)\n\tat
>  org.apache.cassandra.concurrent.FutureTask$1.call(FutureTask.java:96)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.call(FutureTask.java:61)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.run(FutureTask.java:71)\n\tat 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\n\tat
>  
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\n\tat
>  
> 

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-11-22 Thread Jacek Lewandowski (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jacek Lewandowski updated CASSANDRA-18963:
--
Fix Version/s: 4.0.x
   5.1

> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Legacy/Local Write-Read Paths
>Reporter: Berenguer Blasi
>Assignee: Jacek Lewandowski
>Priority: Normal
> Fix For: 4.0.x, 4.1.x, 5.0-rc, 5.1
>
>
> Seems flaky 
> [https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
>  * 
> h4. test_abortedcompaction
> sstableutil_test.TestSSTableUtil 
>  
> {code:java}
> self =  def 
> test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
> CASSANDRA-11497 Check that we can cleanup temporary files after a compaction 
> is aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
> cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
> self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
> tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
> len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
> len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, 
> TableName, filename=log_file_name, delay=2) t.start() try: 
> logger.debug("Compacting...") node.compact() except ToolError: pass # 
> expected to fail t.join() finalfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
> type='final')) tmpfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='tmp')) 
> # In most cases we should end up with some temporary files to clean up, but 
> it may happen # that no temporary files are created if compaction finishes 
> too early or starts too late # see CASSANDRA-11497 logger.debug("Got {} final 
> files and {} tmp files after compaction was interrupted" 
> .format(len(finalfiles), len(tmpfiles))) 
> self._invoke_sstableutil(KeyspaceName, TableName, cleanup=True) 
> self._check_files(node, KeyspaceName, TableName, finalfiles, []) # restart to 
> make sure not data is lost logger.debug("Restarting node...") > 
> node.start(wait_for_binary_proto=True) sstableutil_test.py:97: _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
> self.wait_for_binary_interface(from_mark=self.mark) 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
> wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
> clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
> in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> self =  def 
> raise_node_error_if_cassandra_process_is_terminated(self): if not 
> self._is_pid_running(): msg = "C* process with {pid} is 
> terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
> ccmlib.node.NodeError: C* process with 19530 is terminated 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}
> {code:java}
> failed on teardown with "Unexpected error found in node logs (see stdout for 
> full details). Errors: [[node1] 'ERROR [SSTableBatchOpen:1] 2023-10-19 
> 22:25:07,449 DefaultFSErrorHandler.java:129 - Exiting forcefully due to file 
> system exception on startup, disk failure policy 
> "stop"\norg.apache.cassandra.io.sstable.CorruptSSTableException: Corrupted: 
> /tmp/dtest-ahz16wrh/test/node1/data0/system/prepared_statements-18a9c2576a0c3841ba718cd529849fef/nc-4-big\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReaderLoadingBuilder.build(SSTableReaderLoadingBuilder.java:111)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:397)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:353)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.lambda$openAll$4(SSTableReader.java:414)\n\tat
>  org.apache.cassandra.concurrent.FutureTask$1.call(FutureTask.java:96)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.call(FutureTask.java:61)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.run(FutureTask.java:71)\n\tat 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\n\tat
>  
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\n\tat
>  
> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)\n\tat
>  

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-11-22 Thread Jacek Lewandowski (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jacek Lewandowski updated CASSANDRA-18963:
--
Component/s: Legacy/Local Write-Read Paths
 (was: Test/unit)

> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Legacy/Local Write-Read Paths
>Reporter: Berenguer Blasi
>Assignee: Jacek Lewandowski
>Priority: Normal
> Fix For: 4.1.x, 5.0-rc
>
>
> Seems flaky 
> [https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
>  * 
> h4. test_abortedcompaction
> sstableutil_test.TestSSTableUtil 
>  
> {code:java}
> self =  def 
> test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
> CASSANDRA-11497 Check that we can cleanup temporary files after a compaction 
> is aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
> cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
> self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
> tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
> len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
> len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, 
> TableName, filename=log_file_name, delay=2) t.start() try: 
> logger.debug("Compacting...") node.compact() except ToolError: pass # 
> expected to fail t.join() finalfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
> type='final')) tmpfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='tmp')) 
> # In most cases we should end up with some temporary files to clean up, but 
> it may happen # that no temporary files are created if compaction finishes 
> too early or starts too late # see CASSANDRA-11497 logger.debug("Got {} final 
> files and {} tmp files after compaction was interrupted" 
> .format(len(finalfiles), len(tmpfiles))) 
> self._invoke_sstableutil(KeyspaceName, TableName, cleanup=True) 
> self._check_files(node, KeyspaceName, TableName, finalfiles, []) # restart to 
> make sure not data is lost logger.debug("Restarting node...") > 
> node.start(wait_for_binary_proto=True) sstableutil_test.py:97: _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
> self.wait_for_binary_interface(from_mark=self.mark) 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
> wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
> clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
> in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> self =  def 
> raise_node_error_if_cassandra_process_is_terminated(self): if not 
> self._is_pid_running(): msg = "C* process with {pid} is 
> terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
> ccmlib.node.NodeError: C* process with 19530 is terminated 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}
> {code:java}
> failed on teardown with "Unexpected error found in node logs (see stdout for 
> full details). Errors: [[node1] 'ERROR [SSTableBatchOpen:1] 2023-10-19 
> 22:25:07,449 DefaultFSErrorHandler.java:129 - Exiting forcefully due to file 
> system exception on startup, disk failure policy 
> "stop"\norg.apache.cassandra.io.sstable.CorruptSSTableException: Corrupted: 
> /tmp/dtest-ahz16wrh/test/node1/data0/system/prepared_statements-18a9c2576a0c3841ba718cd529849fef/nc-4-big\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReaderLoadingBuilder.build(SSTableReaderLoadingBuilder.java:111)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:397)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:353)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.lambda$openAll$4(SSTableReader.java:414)\n\tat
>  org.apache.cassandra.concurrent.FutureTask$1.call(FutureTask.java:96)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.call(FutureTask.java:61)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.run(FutureTask.java:71)\n\tat 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\n\tat
>  
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\n\tat
>  
> 

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-11-21 Thread Jacek Lewandowski (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jacek Lewandowski updated CASSANDRA-18963:
--
Fix Version/s: 5.0-rc
   (was: 5.0-beta)

> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Test/unit
>Reporter: Berenguer Blasi
>Assignee: Jacek Lewandowski
>Priority: Normal
> Fix For: 4.1.x, 5.0-rc
>
>
> Seems flaky 
> [https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
>  * 
> h4. test_abortedcompaction
> sstableutil_test.TestSSTableUtil 
>  
> {code:java}
> self =  def 
> test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
> CASSANDRA-11497 Check that we can cleanup temporary files after a compaction 
> is aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
> cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
> self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
> tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
> len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
> len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, 
> TableName, filename=log_file_name, delay=2) t.start() try: 
> logger.debug("Compacting...") node.compact() except ToolError: pass # 
> expected to fail t.join() finalfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
> type='final')) tmpfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='tmp')) 
> # In most cases we should end up with some temporary files to clean up, but 
> it may happen # that no temporary files are created if compaction finishes 
> too early or starts too late # see CASSANDRA-11497 logger.debug("Got {} final 
> files and {} tmp files after compaction was interrupted" 
> .format(len(finalfiles), len(tmpfiles))) 
> self._invoke_sstableutil(KeyspaceName, TableName, cleanup=True) 
> self._check_files(node, KeyspaceName, TableName, finalfiles, []) # restart to 
> make sure not data is lost logger.debug("Restarting node...") > 
> node.start(wait_for_binary_proto=True) sstableutil_test.py:97: _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
> self.wait_for_binary_interface(from_mark=self.mark) 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
> wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
> clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
> in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> self =  def 
> raise_node_error_if_cassandra_process_is_terminated(self): if not 
> self._is_pid_running(): msg = "C* process with {pid} is 
> terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
> ccmlib.node.NodeError: C* process with 19530 is terminated 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}
> {code:java}
> failed on teardown with "Unexpected error found in node logs (see stdout for 
> full details). Errors: [[node1] 'ERROR [SSTableBatchOpen:1] 2023-10-19 
> 22:25:07,449 DefaultFSErrorHandler.java:129 - Exiting forcefully due to file 
> system exception on startup, disk failure policy 
> "stop"\norg.apache.cassandra.io.sstable.CorruptSSTableException: Corrupted: 
> /tmp/dtest-ahz16wrh/test/node1/data0/system/prepared_statements-18a9c2576a0c3841ba718cd529849fef/nc-4-big\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReaderLoadingBuilder.build(SSTableReaderLoadingBuilder.java:111)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:397)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:353)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.lambda$openAll$4(SSTableReader.java:414)\n\tat
>  org.apache.cassandra.concurrent.FutureTask$1.call(FutureTask.java:96)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.call(FutureTask.java:61)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.run(FutureTask.java:71)\n\tat 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\n\tat
>  
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\n\tat
>  
> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)\n\tat
>  

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-11-09 Thread Ekaterina Dimitrova (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ekaterina Dimitrova updated CASSANDRA-18963:

Fix Version/s: 4.1.x

> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Test/unit
>Reporter: Berenguer Blasi
>Assignee: Brandon Williams
>Priority: Normal
> Fix For: 4.1.x, 5.0-beta
>
>
> Seems flaky 
> [https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
>  * 
> h4. test_abortedcompaction
> sstableutil_test.TestSSTableUtil 
>  
> {code:java}
> self =  def 
> test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
> CASSANDRA-11497 Check that we can cleanup temporary files after a compaction 
> is aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
> cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
> self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
> tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
> len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
> len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, 
> TableName, filename=log_file_name, delay=2) t.start() try: 
> logger.debug("Compacting...") node.compact() except ToolError: pass # 
> expected to fail t.join() finalfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
> type='final')) tmpfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='tmp')) 
> # In most cases we should end up with some temporary files to clean up, but 
> it may happen # that no temporary files are created if compaction finishes 
> too early or starts too late # see CASSANDRA-11497 logger.debug("Got {} final 
> files and {} tmp files after compaction was interrupted" 
> .format(len(finalfiles), len(tmpfiles))) 
> self._invoke_sstableutil(KeyspaceName, TableName, cleanup=True) 
> self._check_files(node, KeyspaceName, TableName, finalfiles, []) # restart to 
> make sure not data is lost logger.debug("Restarting node...") > 
> node.start(wait_for_binary_proto=True) sstableutil_test.py:97: _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
> self.wait_for_binary_interface(from_mark=self.mark) 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
> wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
> clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
> in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> self =  def 
> raise_node_error_if_cassandra_process_is_terminated(self): if not 
> self._is_pid_running(): msg = "C* process with {pid} is 
> terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
> ccmlib.node.NodeError: C* process with 19530 is terminated 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}
> {code:java}
> failed on teardown with "Unexpected error found in node logs (see stdout for 
> full details). Errors: [[node1] 'ERROR [SSTableBatchOpen:1] 2023-10-19 
> 22:25:07,449 DefaultFSErrorHandler.java:129 - Exiting forcefully due to file 
> system exception on startup, disk failure policy 
> "stop"\norg.apache.cassandra.io.sstable.CorruptSSTableException: Corrupted: 
> /tmp/dtest-ahz16wrh/test/node1/data0/system/prepared_statements-18a9c2576a0c3841ba718cd529849fef/nc-4-big\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReaderLoadingBuilder.build(SSTableReaderLoadingBuilder.java:111)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:397)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:353)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.lambda$openAll$4(SSTableReader.java:414)\n\tat
>  org.apache.cassandra.concurrent.FutureTask$1.call(FutureTask.java:96)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.call(FutureTask.java:61)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.run(FutureTask.java:71)\n\tat 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\n\tat
>  
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\n\tat
>  
> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)\n\tat
>  java.base/java.lang.Thread.run(Thread.java:829)\nCaused by: 
> 

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-11-02 Thread Benjamin Lerer (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benjamin Lerer updated CASSANDRA-18963:
---
Fix Version/s: 5.0-beta
   (was: 5.0-rc)

> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Test/unit
>Reporter: Berenguer Blasi
>Priority: Normal
> Fix For: 5.0-beta
>
>
> Seems flaky 
> [https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
>  * 
> h4. test_abortedcompaction
> sstableutil_test.TestSSTableUtil 
>  
> {code:java}
> self =  def 
> test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
> CASSANDRA-11497 Check that we can cleanup temporary files after a compaction 
> is aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
> cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
> self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
> tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
> len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
> len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, 
> TableName, filename=log_file_name, delay=2) t.start() try: 
> logger.debug("Compacting...") node.compact() except ToolError: pass # 
> expected to fail t.join() finalfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
> type='final')) tmpfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='tmp')) 
> # In most cases we should end up with some temporary files to clean up, but 
> it may happen # that no temporary files are created if compaction finishes 
> too early or starts too late # see CASSANDRA-11497 logger.debug("Got {} final 
> files and {} tmp files after compaction was interrupted" 
> .format(len(finalfiles), len(tmpfiles))) 
> self._invoke_sstableutil(KeyspaceName, TableName, cleanup=True) 
> self._check_files(node, KeyspaceName, TableName, finalfiles, []) # restart to 
> make sure not data is lost logger.debug("Restarting node...") > 
> node.start(wait_for_binary_proto=True) sstableutil_test.py:97: _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
> self.wait_for_binary_interface(from_mark=self.mark) 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
> wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
> clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
> in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> self =  def 
> raise_node_error_if_cassandra_process_is_terminated(self): if not 
> self._is_pid_running(): msg = "C* process with {pid} is 
> terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
> ccmlib.node.NodeError: C* process with 19530 is terminated 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}
> {code:java}
> failed on teardown with "Unexpected error found in node logs (see stdout for 
> full details). Errors: [[node1] 'ERROR [SSTableBatchOpen:1] 2023-10-19 
> 22:25:07,449 DefaultFSErrorHandler.java:129 - Exiting forcefully due to file 
> system exception on startup, disk failure policy 
> "stop"\norg.apache.cassandra.io.sstable.CorruptSSTableException: Corrupted: 
> /tmp/dtest-ahz16wrh/test/node1/data0/system/prepared_statements-18a9c2576a0c3841ba718cd529849fef/nc-4-big\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReaderLoadingBuilder.build(SSTableReaderLoadingBuilder.java:111)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:397)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:353)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.lambda$openAll$4(SSTableReader.java:414)\n\tat
>  org.apache.cassandra.concurrent.FutureTask$1.call(FutureTask.java:96)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.call(FutureTask.java:61)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.run(FutureTask.java:71)\n\tat 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\n\tat
>  
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\n\tat
>  
> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)\n\tat
>  java.base/java.lang.Thread.run(Thread.java:829)\nCaused by: 
> 

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-11-02 Thread Benjamin Lerer (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benjamin Lerer updated CASSANDRA-18963:
---
Fix Version/s: 5.0-rc
   (was: 5.0.x)

> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Test/unit
>Reporter: Berenguer Blasi
>Priority: Normal
> Fix For: 5.0-rc
>
>
> Seems flaky 
> [https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
>  * 
> h4. test_abortedcompaction
> sstableutil_test.TestSSTableUtil 
>  
> {code:java}
> self =  def 
> test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
> CASSANDRA-11497 Check that we can cleanup temporary files after a compaction 
> is aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
> cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
> self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
> tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
> len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
> len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, 
> TableName, filename=log_file_name, delay=2) t.start() try: 
> logger.debug("Compacting...") node.compact() except ToolError: pass # 
> expected to fail t.join() finalfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
> type='final')) tmpfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='tmp')) 
> # In most cases we should end up with some temporary files to clean up, but 
> it may happen # that no temporary files are created if compaction finishes 
> too early or starts too late # see CASSANDRA-11497 logger.debug("Got {} final 
> files and {} tmp files after compaction was interrupted" 
> .format(len(finalfiles), len(tmpfiles))) 
> self._invoke_sstableutil(KeyspaceName, TableName, cleanup=True) 
> self._check_files(node, KeyspaceName, TableName, finalfiles, []) # restart to 
> make sure not data is lost logger.debug("Restarting node...") > 
> node.start(wait_for_binary_proto=True) sstableutil_test.py:97: _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
> self.wait_for_binary_interface(from_mark=self.mark) 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
> wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
> clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
> in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> self =  def 
> raise_node_error_if_cassandra_process_is_terminated(self): if not 
> self._is_pid_running(): msg = "C* process with {pid} is 
> terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
> ccmlib.node.NodeError: C* process with 19530 is terminated 
> ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}
> {code:java}
> failed on teardown with "Unexpected error found in node logs (see stdout for 
> full details). Errors: [[node1] 'ERROR [SSTableBatchOpen:1] 2023-10-19 
> 22:25:07,449 DefaultFSErrorHandler.java:129 - Exiting forcefully due to file 
> system exception on startup, disk failure policy 
> "stop"\norg.apache.cassandra.io.sstable.CorruptSSTableException: Corrupted: 
> /tmp/dtest-ahz16wrh/test/node1/data0/system/prepared_statements-18a9c2576a0c3841ba718cd529849fef/nc-4-big\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReaderLoadingBuilder.build(SSTableReaderLoadingBuilder.java:111)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:397)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:353)\n\tat
>  
> org.apache.cassandra.io.sstable.format.SSTableReader.lambda$openAll$4(SSTableReader.java:414)\n\tat
>  org.apache.cassandra.concurrent.FutureTask$1.call(FutureTask.java:96)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.call(FutureTask.java:61)\n\tat 
> org.apache.cassandra.concurrent.FutureTask.run(FutureTask.java:71)\n\tat 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\n\tat
>  
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\n\tat
>  
> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)\n\tat
>  java.base/java.lang.Thread.run(Thread.java:829)\nCaused by: 
> 

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-10-31 Thread Ekaterina Dimitrova (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ekaterina Dimitrova updated CASSANDRA-18963:

Description: 
Seems flaky 
[https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
 * 
h4. test_abortedcompaction

sstableutil_test.TestSSTableUtil 
 
{code:java}
self =  def 
test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
CASSANDRA-11497 Check that we can cleanup temporary files after a compaction is 
aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, TableName, 
filename=log_file_name, delay=2) t.start() try: logger.debug("Compacting...") 
node.compact() except ToolError: pass # expected to fail t.join() finalfiles = 
_normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='final')) 
tmpfiles = _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
type='tmp')) # In most cases we should end up with some temporary files to 
clean up, but it may happen # that no temporary files are created if compaction 
finishes too early or starts too late # see CASSANDRA-11497 logger.debug("Got 
{} final files and {} tmp files after compaction was interrupted" 
.format(len(finalfiles), len(tmpfiles))) self._invoke_sstableutil(KeyspaceName, 
TableName, cleanup=True) self._check_files(node, KeyspaceName, TableName, 
finalfiles, []) # restart to make sure not data is lost 
logger.debug("Restarting node...") > node.start(wait_for_binary_proto=True) 
sstableutil_test.py:97: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
_ _ _ _ _ _ _ _ _ _ _ _ 
../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
self.wait_for_binary_interface(from_mark=self.mark) 
../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ _ 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
self =  def 
raise_node_error_if_cassandra_process_is_terminated(self): if not 
self._is_pid_running(): msg = "C* process with {pid} is 
terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
ccmlib.node.NodeError: C* process with 19530 is terminated 
../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}
{code:java}
failed on teardown with "Unexpected error found in node logs (see stdout for 
full details). Errors: [[node1] 'ERROR [SSTableBatchOpen:1] 2023-10-19 
22:25:07,449 DefaultFSErrorHandler.java:129 - Exiting forcefully due to file 
system exception on startup, disk failure policy 
"stop"\norg.apache.cassandra.io.sstable.CorruptSSTableException: Corrupted: 
/tmp/dtest-ahz16wrh/test/node1/data0/system/prepared_statements-18a9c2576a0c3841ba718cd529849fef/nc-4-big\n\tat
 
org.apache.cassandra.io.sstable.format.SSTableReaderLoadingBuilder.build(SSTableReaderLoadingBuilder.java:111)\n\tat
 
org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:397)\n\tat
 
org.apache.cassandra.io.sstable.format.SSTableReader.open(SSTableReader.java:353)\n\tat
 
org.apache.cassandra.io.sstable.format.SSTableReader.lambda$openAll$4(SSTableReader.java:414)\n\tat
 org.apache.cassandra.concurrent.FutureTask$1.call(FutureTask.java:96)\n\tat 
org.apache.cassandra.concurrent.FutureTask.call(FutureTask.java:61)\n\tat 
org.apache.cassandra.concurrent.FutureTask.run(FutureTask.java:71)\n\tat 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\n\tat
 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\n\tat
 
io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)\n\tat
 java.base/java.lang.Thread.run(Thread.java:829)\nCaused by: 
java.lang.IllegalArgumentException: null\n\tat 
com.google.common.base.Preconditions.checkArgument(Preconditions.java:129)\n\tat
 
org.apache.cassandra.io.sstable.format.big.BigSSTableReaderLoadingBuilder.openComponents(BigSSTableReaderLoadingBuilder.java:79)\n\tat
 
org.apache.cassandra.io.sstable.format.big.BigSSTableReaderLoadingBuilder.openComponents(BigSSTableReaderLoadingBuilder.java:58)\n\tat
 
org.apache.cassandra.io.sstable.format.SSTableReaderLoadingBuilder.build(SSTableReaderLoadingBuilder.java:92)\n\t...
 10 common frames omitted', [node1] 'ERROR [SSTableBatchOpen:4] 2023-10-19 
22:25:07,449 

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-10-31 Thread Ekaterina Dimitrova (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ekaterina Dimitrova updated CASSANDRA-18963:

Description: 
Seems flaky 
[https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
 * 
h4. test_abortedcompaction


sstableutil_test.TestSSTableUtil 
 
{code:java}
self =  def 
test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
CASSANDRA-11497 Check that we can cleanup temporary files after a compaction is 
aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, TableName, 
filename=log_file_name, delay=2) t.start() try: logger.debug("Compacting...") 
node.compact() except ToolError: pass # expected to fail t.join() finalfiles = 
_normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='final')) 
tmpfiles = _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
type='tmp')) # In most cases we should end up with some temporary files to 
clean up, but it may happen # that no temporary files are created if compaction 
finishes too early or starts too late # see CASSANDRA-11497 logger.debug("Got 
{} final files and {} tmp files after compaction was interrupted" 
.format(len(finalfiles), len(tmpfiles))) self._invoke_sstableutil(KeyspaceName, 
TableName, cleanup=True) self._check_files(node, KeyspaceName, TableName, 
finalfiles, []) # restart to make sure not data is lost 
logger.debug("Restarting node...") > node.start(wait_for_binary_proto=True) 
sstableutil_test.py:97: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
_ _ _ _ _ _ _ _ _ _ _ _ 
../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
self.wait_for_binary_interface(from_mark=self.mark) 
../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ _ 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
self =  def 
raise_node_error_if_cassandra_process_is_terminated(self): if not 
self._is_pid_running(): msg = "C* process with {pid} is 
terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
ccmlib.node.NodeError: C* process with 19530 is terminated 
../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}

  was:
Seems flaky 
[https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]


 * 
h4. test_abortedcompaction
 
sstableutil_test.TestSSTableUtil
 
 
{code:java}
self =  def 
test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
CASSANDRA-11497 Check that we can cleanup temporary files after a compaction is 
aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, TableName, 
filename=log_file_name, delay=2) t.start() try: logger.debug("Compacting...") 
node.compact() except ToolError: pass # expected to fail t.join() finalfiles = 
_normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='final')) 
tmpfiles = _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
type='tmp')) # In most cases we should end up with some temporary files to 
clean up, but it may happen # that no temporary files are created if compaction 
finishes too early or starts too late # see CASSANDRA-11497 logger.debug("Got 
{} final files and {} tmp files after compaction was interrupted" 
.format(len(finalfiles), len(tmpfiles))) self._invoke_sstableutil(KeyspaceName, 
TableName, cleanup=True) self._check_files(node, KeyspaceName, TableName, 
finalfiles, []) # restart to make sure not data is lost 
logger.debug("Restarting node...") > node.start(wait_for_binary_proto=True) 
sstableutil_test.py:97: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
_ _ _ _ _ _ _ _ _ _ _ _ 
../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
self.wait_for_binary_interface(from_mark=self.mark) 
../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
wait_for_binary_interface 

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-10-31 Thread Ekaterina Dimitrova (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ekaterina Dimitrova updated CASSANDRA-18963:

Description: 
Seems flaky 
[https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]


 * 
h4. test_abortedcompaction
 
sstableutil_test.TestSSTableUtil
 
 
{code:java}
self =  def 
test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
CASSANDRA-11497 Check that we can cleanup temporary files after a compaction is 
aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, TableName, 
filename=log_file_name, delay=2) t.start() try: logger.debug("Compacting...") 
node.compact() except ToolError: pass # expected to fail t.join() finalfiles = 
_normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='final')) 
tmpfiles = _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
type='tmp')) # In most cases we should end up with some temporary files to 
clean up, but it may happen # that no temporary files are created if compaction 
finishes too early or starts too late # see CASSANDRA-11497 logger.debug("Got 
{} final files and {} tmp files after compaction was interrupted" 
.format(len(finalfiles), len(tmpfiles))) self._invoke_sstableutil(KeyspaceName, 
TableName, cleanup=True) self._check_files(node, KeyspaceName, TableName, 
finalfiles, []) # restart to make sure not data is lost 
logger.debug("Restarting node...") > node.start(wait_for_binary_proto=True) 
sstableutil_test.py:97: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
_ _ _ _ _ _ _ _ _ _ _ _ 
../env3.6/lib/python3.6/site-packages/ccmlib/node.py:914: in start 
self.wait_for_binary_interface(from_mark=self.mark) 
../env3.6/lib/python3.6/site-packages/ccmlib/node.py:702: in 
wait_for_binary_interface self.watch_log_for("Starting listening for CQL 
clients", **kwargs) ../env3.6/lib/python3.6/site-packages/ccmlib/node.py:599: 
in watch_log_for self.raise_node_error_if_cassandra_process_is_terminated() _ _ 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
self =  def 
raise_node_error_if_cassandra_process_is_terminated(self): if not 
self._is_pid_running(): msg = "C* process with {pid} is 
terminated".format(pid=self.pid) common.debug(msg) > raise NodeError(msg) E 
ccmlib.node.NodeError: C* process with 19530 is terminated 
../env3.6/lib/python3.6/site-packages/ccmlib/node.py:683: NodeError{code}

  was:Seems flaky 
https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests


> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Test/unit
>Reporter: Berenguer Blasi
>Priority: Normal
> Fix For: 5.0.x
>
>
> Seems flaky 
> [https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests]
>  * 
> h4. test_abortedcompaction
>  
> sstableutil_test.TestSSTableUtil
>  
>  
> {code:java}
> self =  def 
> test_abortedcompaction(self): """ @jira_ticket CASSANDRA-7066 @jira_ticket 
> CASSANDRA-11497 Check that we can cleanup temporary files after a compaction 
> is aborted. """ log_file_name = 'debug.log' cluster = self.cluster 
> cluster.populate(1).start() node = cluster.nodelist()[0] numrecords = 25 
> self._create_data(node, KeyspaceName, TableName, numrecords) finalfiles, 
> tmpfiles = self._check_files(node, KeyspaceName, TableName) assert 
> len(finalfiles) > 0, "Expected to find some final files" assert 0 == 
> len(tmpfiles), "Expected no tmp files" t = InterruptCompaction(node, 
> TableName, filename=log_file_name, delay=2) t.start() try: 
> logger.debug("Compacting...") node.compact() except ToolError: pass # 
> expected to fail t.join() finalfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, 
> type='final')) tmpfiles = 
> _normcase_all(self._invoke_sstableutil(KeyspaceName, TableName, type='tmp')) 
> # In most cases we should end up with some temporary files to clean up, but 
> it may happen # that no temporary files are created if compaction finishes 
> too early or starts too late # see CASSANDRA-11497 logger.debug("Got {} final 
> files and {} tmp files after compaction was interrupted" 

[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-10-26 Thread Brandon Williams (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Williams updated CASSANDRA-18963:
-
Fix Version/s: 5.0.x
   (was: 5.0)

> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Test/unit
>Reporter: Berenguer Blasi
>Priority: Normal
> Fix For: 5.0.x
>
>
> Seems flaky 
> https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: commits-unsubscr...@cassandra.apache.org
For additional commands, e-mail: commits-h...@cassandra.apache.org



[jira] [Updated] (CASSANDRA-18963) Test Failure: sstableutil_test.TestSSTableUtil

2023-10-26 Thread Berenguer Blasi (Jira)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-18963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Berenguer Blasi updated CASSANDRA-18963:

 Bug Category: Parent values: Correctness(12982)
   Complexity: Normal
  Component/s: Test/unit
Discovered By: Unit Test
Fix Version/s: 5.0
 Severity: Normal
   Status: Open  (was: Triage Needed)

> Test Failure: sstableutil_test.TestSSTableUtil
> --
>
> Key: CASSANDRA-18963
> URL: https://issues.apache.org/jira/browse/CASSANDRA-18963
> Project: Cassandra
>  Issue Type: Bug
>  Components: Test/unit
>Reporter: Berenguer Blasi
>Priority: Normal
> Fix For: 5.0
>
>
> Seems flaky 
> https://app.circleci.com/pipelines/github/driftx/cassandra/1342/workflows/9554c7b5-dd60-4f08-8db5-c954febc8ad6/jobs/58957/tests



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: commits-unsubscr...@cassandra.apache.org
For additional commands, e-mail: commits-h...@cassandra.apache.org