Hello, I am trying to run the PyDML script below using the Spark ML Context.
import systemml as smlimport numpy as npsml.setSparkContext(sc)m1 = sml.matrix(np.ones((3,3)) + 2)m2 = sml.matrix(np.ones((3,3)) + 3)m2 = m1 * (m2 + m1)m4 = 1.0 - m2m4.sum(axis=1).toNumPyArray() I start Spark Shell and create ML context successfully. Then I load the script from a file using the following command val s4 = ScriptFactory.pydmlFromFile("test.pydml") Finally, I execute the script using ml.execute(s4) The imports are not recognized. I suppose that the first import and setting the Spark context are not required, since we set up a MLContext after starting Spark Shell, but what about numpy? I am a bit confused as to what changes I need to make to run this example. Thank you in advance for your help, Nantia