Baunsgaard commented on PR #1961: URL: https://github.com/apache/systemds/pull/1961#issuecomment-1865566384
> > The change is fine, and most likely all good. But we need to test it with an actual Hadoop instance. The test is what makes this upgrade harder than it seems on the surface. It is the same for the Spark update. > > @Baunsgaard What would testing entail? Is it as simple as running all unit/integration tests on a Hadoop cluster? Good question. The unit tests we have does use HDFS per default but you would not see if it works or not when running the tests unless some of them suddenly crash. The way I did it before was to have simple scripts that write and read matrices and frames from an HDFS cluster (our normal execution should default to it if HDFS is detected). This does not give a full guarantee everything works but does sufficiently verify support. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: dev-unsubscr...@systemds.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org