hmm, I just tested it again and it runs fine on a setup spark cluster
and that's how I tested the artifact. Regarding forced single-node
computations (as used in bin/systemds or the python API) it runs into
issues. So yes, let's cut, vote, and release a fix that moves the code
change that pulls in the Spark dependencies into CommonThreadPool.
Regards,
Matthias
On 11/9/2021 1:35 PM, Baunsgaard, Sebastian wrote:
Hi Devs,
If you install the official release and follow our instructions it fails
because somehow not all the required java packages is included.
This error is not found in our automatic tests, since we don't have any
verification tests of our binary release assets.
python release is fine, since it include all the java packages.
I suggest that we make a new patch release 2.2.1 (forked from the official
release commit), where all required java packages are included in the bin
release.
Best regards
Sebastian
steps to reproduce:
1. Download the release <https://systemds.apache.org/download>
https://www.apache.org/dyn/closer.lua/systemds/2.2.0/systemds-2.2.0-bin.zip
2. Setup environment ... (like our docs says)
* export SYSTEMDS_ROOT= "path/to/unziped/bin"
* export PATH=$SYSTEMDS_ROOT/bin:$PATH
3. run any script 'systemds test.dml'
The error looks as follows:
Me:~/temp$ systemds test.dml
###############################################################################
# SYSTEMDS_ROOT= ../systemds/systemds-2.2.0-bin
# SYSTEMDS_JAR_FILE= ../systemds/systemds-2.2.0-bin/lib/systemds-2.2.0.jar
# SYSDS_EXEC_MODE= singlenode
# CONFIG_FILE=
# LOG4JPROP=
-Dlog4j.configuration=file:/home/baunsgaard/systemds/systemds-2.2.0-bin//conf/log4j.properties
# CLASSPATH=
../systemds/systemds-2.2.0-bin/lib/systemds-2.2.0.jar:../systemds/systemds-2.2.0-bin/lib/*:../systemds/systemds-2.2.0-bin/target/lib/*
# HADOOP_HOME= /home/baunsgaard/systemds/systemds-2.2.0-bin/lib/hadoop
#
# Running script test.dml locally with opts:
###############################################################################
Executing command: java -Xmx4g -Xms4g -Xmn400m -cp
../systemds/systemds-2.2.0-bin/lib/systemds-2.2.0.jar:../systemds/systemds-2.2.0-bin/lib/*:../systemds/systemds-2.2.0-bin/target/lib/*
-Dlog4j.configuration=file:/home/baunsgaard/systemds/systemds-2.2.0-bin//conf/log4j.properties
org.apache.sysds.api.DMLScript -f test.dml -exec singlenode
Exception in thread "main" java.lang.NoClassDefFoundError: scala/Function0
at org.apache.sysds.lops.Checkpoint.<clinit>(Checkpoint.java:43)
at
org.apache.sysds.runtime.instructions.spark.utils.SparkUtils.<clinit>(SparkUtils.java:69)
at
org.apache.sysds.api.DMLScript.cleanupHadoopExecution(DMLScript.java:522)
at org.apache.sysds.api.DMLScript.initHadoopExecution(DMLScript.java:494)
at org.apache.sysds.api.DMLScript.execute(DMLScript.java:402)
at org.apache.sysds.api.DMLScript.executeScript(DMLScript.java:274)
at org.apache.sysds.api.DMLScript.main(DMLScript.java:169)
Caused by: java.lang.ClassNotFoundException: scala.Function0
at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
at java.lang.ClassLoader.loadClass(ClassLoader.java:418)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352)
at java.lang.ClassLoader.loadClass(ClassLoader.java:351)
... 7 more