Baunsgaard commented on PR #1961:
URL: https://github.com/apache/systemds/pull/1961#issuecomment-1865566384

   > > The change is fine, and most likely all good. But we need to test it 
with an actual Hadoop instance. The test is what makes this upgrade harder than 
it seems on the surface. It is the same for the Spark update.
   > 
   > @Baunsgaard What would testing entail? Is it as simple as running all 
unit/integration tests on a Hadoop cluster?
   
   Good question. 
   
   The unit tests we have does use HDFS per default but you would not see if it 
works or not when running the tests unless some of them suddenly crash. 
   
   The way I did it before was to have simple scripts that write and read 
matrices and frames from an HDFS cluster (our normal execution should default 
to it if HDFS is detected). This does not give a full guarantee everything 
works but does sufficiently verify support. 
   
   
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: dev-unsubscr...@systemds.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org

Reply via email to