hmm, I just tested it again and it runs fine on a setup spark cluster and that's how I tested the artifact. Regarding forced single-node computations (as used in bin/systemds or the python API) it runs into issues. So yes, let's cut, vote, and release a fix that moves the code change that pulls in the Spark dependencies into CommonThreadPool.

Regards,
Matthias

On 11/9/2021 1:35 PM, Baunsgaard, Sebastian wrote:
Hi Devs,

If you install the official release and follow our instructions it fails 
because somehow not all the required java packages is included.
This error is not found in our automatic tests, since we don't have any 
verification tests of our binary release assets.

python release is fine, since it include all the java packages.

I suggest that we make a new patch release 2.2.1 (forked from the official 
release commit), where all required java packages are included in the bin 
release.

Best regards
Sebastian

steps to reproduce:

   1.  Download the release <https://systemds.apache.org/download> 
https://www.apache.org/dyn/closer.lua/systemds/2.2.0/systemds-2.2.0-bin.zip
   2.  Setup environment ... (like our docs says)
      *   export SYSTEMDS_ROOT= "path/to/unziped/bin"
      *   export PATH=$SYSTEMDS_ROOT/bin:$PATH
   3.  run any script 'systemds test.dml'

The error looks as follows:


Me:~/temp$ systemds test.dml
###############################################################################
#  SYSTEMDS_ROOT= ../systemds/systemds-2.2.0-bin
#  SYSTEMDS_JAR_FILE= ../systemds/systemds-2.2.0-bin/lib/systemds-2.2.0.jar
#  SYSDS_EXEC_MODE= singlenode
#  CONFIG_FILE=
#  LOG4JPROP= 
-Dlog4j.configuration=file:/home/baunsgaard/systemds/systemds-2.2.0-bin//conf/log4j.properties
#  CLASSPATH= 
../systemds/systemds-2.2.0-bin/lib/systemds-2.2.0.jar:../systemds/systemds-2.2.0-bin/lib/*:../systemds/systemds-2.2.0-bin/target/lib/*
#  HADOOP_HOME= /home/baunsgaard/systemds/systemds-2.2.0-bin/lib/hadoop
#
#  Running script test.dml locally with opts:
###############################################################################
Executing command:     java       -Xmx4g      -Xms4g      -Xmn400m    -cp 
../systemds/systemds-2.2.0-bin/lib/systemds-2.2.0.jar:../systemds/systemds-2.2.0-bin/lib/*:../systemds/systemds-2.2.0-bin/target/lib/*
   
-Dlog4j.configuration=file:/home/baunsgaard/systemds/systemds-2.2.0-bin//conf/log4j.properties
   org.apache.sysds.api.DMLScript   -f test.dml   -exec singlenode

Exception in thread "main" java.lang.NoClassDefFoundError: scala/Function0
     at org.apache.sysds.lops.Checkpoint.<clinit>(Checkpoint.java:43)
     at 
org.apache.sysds.runtime.instructions.spark.utils.SparkUtils.<clinit>(SparkUtils.java:69)
     at 
org.apache.sysds.api.DMLScript.cleanupHadoopExecution(DMLScript.java:522)
     at org.apache.sysds.api.DMLScript.initHadoopExecution(DMLScript.java:494)
     at org.apache.sysds.api.DMLScript.execute(DMLScript.java:402)
     at org.apache.sysds.api.DMLScript.executeScript(DMLScript.java:274)
     at org.apache.sysds.api.DMLScript.main(DMLScript.java:169)
Caused by: java.lang.ClassNotFoundException: scala.Function0
     at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
     at java.lang.ClassLoader.loadClass(ClassLoader.java:418)
     at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352)
     at java.lang.ClassLoader.loadClass(ClassLoader.java:351)
     ... 7 more





Reply via email to