niketanpansare edited a comment on issue #857: [SYSTEMML-2523] Update SystemML 
to Support Spark 2.3.0
URL: https://github.com/apache/systemml/pull/857#issuecomment-475394794
 
 
   @romeokienzler You are getting the error because the setup contains two 
SystemML (possibly conflicting dependencies) jars. There are two possible 
solutions to your problem:
   1. *Recommended:* Remove the older incubating jar and do not include the 
corresponding 1.2.0 or 1.3.0-snapshot jars (i.e. no need for `ln -s` trick).
   2. Use the python package compiled by this PR.
   
   Since there is something weird happening here, I am including the logs. I 
apologize it advance for the long trace. Please ignore the below logs if you 
agree to the above statements.
   
   Setup 1. With only incubating jar (FAILS !!)
   ```
   $ ~/spark-2.3.0-bin-hadoop2.7/bin/pyspark --driver-memory 20g --master 
local[*] --driver-class-path systemml-0.14.0-incubating.jar
   Python 3.6.3 (default, Mar 20 2018, 13:50:41) 
   [GCC 4.8.5 20150623 (Red Hat 4.8.5-16)] on linux
   Type "help", "copyright", "credits" or "license" for more information.
   2019-03-21 13:07:11 WARN  NativeCodeLoader:62 - Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
   Setting default log level to "WARN".
   To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use 
setLogLevel(newLevel).
   Welcome to
         ____              __
        / __/__  ___ _____/ /__
       _\ \/ _ \/ _ `/ __/  '_/
      /__ / .__/\_,_/_/ /_/\_\   version 2.3.0
         /_/
   
   Using Python version 3.6.3 (default, Mar 20 2018 13:50:41)
   SparkSession available as 'spark'.
   >>> from systemml import MLContext
   >>> ml = MLContext(spark)
   2019-03-21 13:07:20 WARN  ObjectStore:568 - Failed to get database 
global_temp, returning NoSuchObjectException
   
   Welcome to Apache SystemML!
   
   >>> ml.version()
   '0.14.0-incubating'
   >>> df=spark.read.parquet('shake.parquet')
   >>> df.show()
   +-----+---------+-----+-----+-----+
   |CLASS| SENSORID|    X|    Y|    Z|
   +-----+---------+-----+-----+-----+
   |    2| qqqqqqqq| 0.12| 0.12| 0.12|
   |    2|aUniqueID| 0.03| 0.03| 0.03|
   |    2| qqqqqqqq|-3.84|-3.84|-3.84|
   |    2| 12345678| -0.1| -0.1| -0.1|
   |    2| 12345678|-0.15|-0.15|-0.15|
   |    2| 12345678| 0.47| 0.47| 0.47|
   |    2| 12345678|-0.06|-0.06|-0.06|
   |    2| 12345678|-0.09|-0.09|-0.09|
   |    2| 12345678| 0.21| 0.21| 0.21|
   |    2| 12345678|-0.08|-0.08|-0.08|
   |    2| 12345678| 0.44| 0.44| 0.44|
   |    2|    gholi| 0.76| 0.76| 0.76|
   |    2|    gholi| 1.62| 1.62| 1.62|
   |    2|    gholi| 5.81| 5.81| 5.81|
   |    2| bcbcbcbc| 0.58| 0.58| 0.58|
   |    2| bcbcbcbc|-8.24|-8.24|-8.24|
   |    2| bcbcbcbc|-0.45|-0.45|-0.45|
   |    2| bcbcbcbc| 1.03| 1.03| 1.03|
   |    2|aUniqueID|-0.05|-0.05|-0.05|
   |    2| qqqqqqqq|-0.44|-0.44|-0.44|
   +-----+---------+-----+-----+-----+
   only showing top 20 rows
   
   >>> df.createOrReplaceTempView("df")
   ANTLR Tool version 4.7 used for code generation does not match the current 
runtime version 4.5.3ANTLR Runtime version 4.7 used for parser compilation does 
not match the current runtime version 4.5.3Traceback (most recent call last):
     File "<stdin>", line 1, in <module>
     File 
"/home/npansar/spark-2.3.0-bin-hadoop2.7/python/pyspark/sql/dataframe.py", line 
176, in createOrReplaceTempView
       self._jdf.createOrReplaceTempView(name)
     File 
"/home/npansar/spark-2.3.0-bin-hadoop2.7/python/lib/py4j-0.10.6-src.zip/py4j/java_gateway.py",
 line 1160, in __call__
     File 
"/home/npansar/spark-2.3.0-bin-hadoop2.7/python/pyspark/sql/utils.py", line 63, 
in deco
       return f(*a, **kw)
     File 
"/home/npansar/spark-2.3.0-bin-hadoop2.7/python/lib/py4j-0.10.6-src.zip/py4j/protocol.py",
 line 320, in get_return_value
   py4j.protocol.Py4JJavaError: An error occurred while calling 
o52.createOrReplaceTempView.
   : java.lang.ExceptionInInitializerError
        at 
org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parse(ParseDriver.scala:84)
        at 
org.apache.spark.sql.execution.SparkSqlParser.parse(SparkSqlParser.scala:48)
        at 
org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parseTableIdentifier(ParseDriver.scala:49)
        at 
org.apache.spark.sql.Dataset.createTempViewCommand(Dataset.scala:3079)
        at 
org.apache.spark.sql.Dataset.createOrReplaceTempView(Dataset.scala:3034)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
        at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
        at py4j.Gateway.invoke(Gateway.java:282)
        at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
        at py4j.commands.CallCommand.execute(CallCommand.java:79)
        at py4j.GatewayConnection.run(GatewayConnection.java:214)
        at java.lang.Thread.run(Thread.java:748)
   Caused by: java.lang.UnsupportedOperationException: 
java.io.InvalidClassException: org.antlr.v4.runtime.atn.ATN; Could not 
deserialize ATN with UUID 59627784-3be5-417a-b9eb-8131a7286089 (expected 
aadb8d7e-aeef-4415-ad2b-8204d6cf042e or a legacy UUID).
        at 
org.antlr.v4.runtime.atn.ATNDeserializer.deserialize(ATNDeserializer.java:153)
        at 
org.apache.spark.sql.catalyst.parser.SqlBaseLexer.<clinit>(SqlBaseLexer.java:1153)
        ... 16 more
   Caused by: java.io.InvalidClassException: org.antlr.v4.runtime.atn.ATN; 
Could not deserialize ATN with UUID 59627784-3be5-417a-b9eb-8131a7286089 
(expected aadb8d7e-aeef-4415-ad2b-8204d6cf042e or a legacy UUID).
        ... 18 more
   
   >>>
   ```
   
   Setup 2: Put the older incubating jar before the current SystemML 1.2.0 jars 
(FAILS !!)
   ```
   $ ~/spark-2.3.0-bin-hadoop2.7/bin/pyspark --driver-memory 20g --master 
local[*] --driver-class-path 
systemml-0.14.0-incubating.jar:systemml-1.2.0-extra.jar:systemml-1.2.0.jar
   Python 3.6.3 (default, Mar 20 2018, 13:50:41) 
   [GCC 4.8.5 20150623 (Red Hat 4.8.5-16)] on linux
   Type "help", "copyright", "credits" or "license" for more information.
   2019-03-21 13:12:11 WARN  NativeCodeLoader:62 - Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
   Setting default log level to "WARN".
   To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use 
setLogLevel(newLevel).
   Welcome to
         ____              __
        / __/__  ___ _____/ /__
       _\ \/ _ \/ _ `/ __/  '_/
      /__ / .__/\_,_/_/ /_/\_\   version 2.3.0
         /_/
   
   Using Python version 3.6.3 (default, Mar 20 2018 13:50:41)
   SparkSession available as 'spark'.
   >>> from systemml import MLContext
   >>> ml = MLContext(spark)
   2019-03-21 13:12:21 WARN  ObjectStore:568 - Failed to get database 
global_temp, returning NoSuchObjectException
   
   Welcome to Apache SystemML!
   
   >>> ml.version()
   '0.14.0-incubating'
   >>> df=spark.read.parquet('shake.parquet')
   >>> df.show()
   +-----+---------+-----+-----+-----+
   |CLASS| SENSORID|    X|    Y|    Z|
   +-----+---------+-----+-----+-----+
   |    2| qqqqqqqq| 0.12| 0.12| 0.12|
   |    2|aUniqueID| 0.03| 0.03| 0.03|
   |    2| qqqqqqqq|-3.84|-3.84|-3.84|
   |    2| 12345678| -0.1| -0.1| -0.1|
   |    2| 12345678|-0.15|-0.15|-0.15|
   |    2| 12345678| 0.47| 0.47| 0.47|
   |    2| 12345678|-0.06|-0.06|-0.06|
   |    2| 12345678|-0.09|-0.09|-0.09|
   |    2| 12345678| 0.21| 0.21| 0.21|
   |    2| 12345678|-0.08|-0.08|-0.08|
   |    2| 12345678| 0.44| 0.44| 0.44|
   |    2|    gholi| 0.76| 0.76| 0.76|
   |    2|    gholi| 1.62| 1.62| 1.62|
   |    2|    gholi| 5.81| 5.81| 5.81|
   |    2| bcbcbcbc| 0.58| 0.58| 0.58|
   |    2| bcbcbcbc|-8.24|-8.24|-8.24|
   |    2| bcbcbcbc|-0.45|-0.45|-0.45|
   |    2| bcbcbcbc| 1.03| 1.03| 1.03|
   |    2|aUniqueID|-0.05|-0.05|-0.05|
   |    2| qqqqqqqq|-0.44|-0.44|-0.44|
   +-----+---------+-----+-----+-----+
   only showing top 20 rows
   
   >>> df.createOrReplaceTempView("df")
   ANTLR Tool version 4.7 used for code generation does not match the current 
runtime version 4.5.3ANTLR Runtime version 4.7 used for parser compilation does 
not match the current runtime version 4.5.3Traceback (most recent call last):
     File "<stdin>", line 1, in <module>
     File 
"/home/npansar/spark-2.3.0-bin-hadoop2.7/python/pyspark/sql/dataframe.py", line 
176, in createOrReplaceTempView
       self._jdf.createOrReplaceTempView(name)
     File 
"/home/npansar/spark-2.3.0-bin-hadoop2.7/python/lib/py4j-0.10.6-src.zip/py4j/java_gateway.py",
 line 1160, in __call__
     File 
"/home/npansar/spark-2.3.0-bin-hadoop2.7/python/pyspark/sql/utils.py", line 63, 
in deco
       return f(*a, **kw)
     File 
"/home/npansar/spark-2.3.0-bin-hadoop2.7/python/lib/py4j-0.10.6-src.zip/py4j/protocol.py",
 line 320, in get_return_value
   py4j.protocol.Py4JJavaError: An error occurred while calling 
o52.createOrReplaceTempView.
   : java.lang.ExceptionInInitializerError
        at 
org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parse(ParseDriver.scala:84)
        at 
org.apache.spark.sql.execution.SparkSqlParser.parse(SparkSqlParser.scala:48)
        at 
org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parseTableIdentifier(ParseDriver.scala:49)
        at 
org.apache.spark.sql.Dataset.createTempViewCommand(Dataset.scala:3079)
        at 
org.apache.spark.sql.Dataset.createOrReplaceTempView(Dataset.scala:3034)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
        at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
        at py4j.Gateway.invoke(Gateway.java:282)
        at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
        at py4j.commands.CallCommand.execute(CallCommand.java:79)
        at py4j.GatewayConnection.run(GatewayConnection.java:214)
        at java.lang.Thread.run(Thread.java:748)
   Caused by: java.lang.UnsupportedOperationException: 
java.io.InvalidClassException: org.antlr.v4.runtime.atn.ATN; Could not 
deserialize ATN with UUID 59627784-3be5-417a-b9eb-8131a7286089 (expected 
aadb8d7e-aeef-4415-ad2b-8204d6cf042e or a legacy UUID).
        at 
org.antlr.v4.runtime.atn.ATNDeserializer.deserialize(ATNDeserializer.java:153)
        at 
org.apache.spark.sql.catalyst.parser.SqlBaseLexer.<clinit>(SqlBaseLexer.java:1153)
        ... 16 more
   Caused by: java.io.InvalidClassException: org.antlr.v4.runtime.atn.ATN; 
Could not deserialize ATN with UUID 59627784-3be5-417a-b9eb-8131a7286089 
(expected aadb8d7e-aeef-4415-ad2b-8204d6cf042e or a legacy UUID).
        ... 18 more
   
   >>>
   ```
   
   Setup 3: Put the the current SystemML 1.2.0 jars before the older incubating 
jar (FAILS !!)
   ```
   $ ~/spark-2.3.0-bin-hadoop2.7/bin/pyspark --driver-memory 20g --master 
local[*] --driver-class-path 
systemml-1.2.0-extra.jar:systemml-1.2.0.jar:systemml-0.14.0-incubating.jar
   Python 3.6.3 (default, Mar 20 2018, 13:50:41) 
   [GCC 4.8.5 20150623 (Red Hat 4.8.5-16)] on linux
   Type "help", "copyright", "credits" or "license" for more information.
   2019-03-21 13:14:49 WARN  NativeCodeLoader:62 - Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
   Setting default log level to "WARN".
   To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use 
setLogLevel(newLevel).
   Welcome to
         ____              __
        / __/__  ___ _____/ /__
       _\ \/ _ \/ _ `/ __/  '_/
      /__ / .__/\_,_/_/ /_/\_\   version 2.3.0
         /_/
   
   Using Python version 3.6.3 (default, Mar 20 2018 13:50:41)
   SparkSession available as 'spark'.
   >>> from systemml import MLContext
   >>> ml = MLContext(spark)
   2019-03-21 13:15:11 WARN  ObjectStore:568 - Failed to get database 
global_temp, returning NoSuchObjectException
   
   Welcome to Apache SystemML!
   Version 1.2.0
   >>> ml.version()
   '1.2.0'
   >>> df=spark.read.parquet('shake.parquet')
   >>> df.show()
   +-----+---------+-----+-----+-----+
   |CLASS| SENSORID|    X|    Y|    Z|
   +-----+---------+-----+-----+-----+
   |    2| qqqqqqqq| 0.12| 0.12| 0.12|
   |    2|aUniqueID| 0.03| 0.03| 0.03|
   |    2| qqqqqqqq|-3.84|-3.84|-3.84|
   |    2| 12345678| -0.1| -0.1| -0.1|
   |    2| 12345678|-0.15|-0.15|-0.15|
   |    2| 12345678| 0.47| 0.47| 0.47|
   |    2| 12345678|-0.06|-0.06|-0.06|
   |    2| 12345678|-0.09|-0.09|-0.09|
   |    2| 12345678| 0.21| 0.21| 0.21|
   |    2| 12345678|-0.08|-0.08|-0.08|
   |    2| 12345678| 0.44| 0.44| 0.44|
   |    2|    gholi| 0.76| 0.76| 0.76|
   |    2|    gholi| 1.62| 1.62| 1.62|
   |    2|    gholi| 5.81| 5.81| 5.81|
   |    2| bcbcbcbc| 0.58| 0.58| 0.58|
   |    2| bcbcbcbc|-8.24|-8.24|-8.24|
   |    2| bcbcbcbc|-0.45|-0.45|-0.45|
   |    2| bcbcbcbc| 1.03| 1.03| 1.03|
   |    2|aUniqueID|-0.05|-0.05|-0.05|
   |    2| qqqqqqqq|-0.44|-0.44|-0.44|
   +-----+---------+-----+-----+-----+
   only showing top 20 rows
   
   >>> df.createOrReplaceTempView("df")
   ANTLR Tool version 4.7 used for code generation does not match the current 
runtime version 4.5.3ANTLR Runtime version 4.7 used for parser compilation does 
not match the current runtime version 4.5.3Traceback (most recent call last):
     File "<stdin>", line 1, in <module>
     File 
"/home/npansar/spark-2.3.0-bin-hadoop2.7/python/pyspark/sql/dataframe.py", line 
176, in createOrReplaceTempView
       self._jdf.createOrReplaceTempView(name)
     File 
"/home/npansar/spark-2.3.0-bin-hadoop2.7/python/lib/py4j-0.10.6-src.zip/py4j/java_gateway.py",
 line 1160, in __call__
     File 
"/home/npansar/spark-2.3.0-bin-hadoop2.7/python/pyspark/sql/utils.py", line 63, 
in deco
       return f(*a, **kw)
     File 
"/home/npansar/spark-2.3.0-bin-hadoop2.7/python/lib/py4j-0.10.6-src.zip/py4j/protocol.py",
 line 320, in get_return_value
   py4j.protocol.Py4JJavaError: An error occurred while calling 
o52.createOrReplaceTempView.
   : java.lang.ExceptionInInitializerError
        at 
org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parse(ParseDriver.scala:84)
        at 
org.apache.spark.sql.execution.SparkSqlParser.parse(SparkSqlParser.scala:48)
        at 
org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parseTableIdentifier(ParseDriver.scala:49)
        at 
org.apache.spark.sql.Dataset.createTempViewCommand(Dataset.scala:3079)
        at 
org.apache.spark.sql.Dataset.createOrReplaceTempView(Dataset.scala:3034)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
        at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
        at py4j.Gateway.invoke(Gateway.java:282)
        at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
        at py4j.commands.CallCommand.execute(CallCommand.java:79)
        at py4j.GatewayConnection.run(GatewayConnection.java:214)
        at java.lang.Thread.run(Thread.java:748)
   Caused by: java.lang.UnsupportedOperationException: 
java.io.InvalidClassException: org.antlr.v4.runtime.atn.ATN; Could not 
deserialize ATN with UUID 59627784-3be5-417a-b9eb-8131a7286089 (expected 
aadb8d7e-aeef-4415-ad2b-8204d6cf042e or a legacy UUID).
        at 
org.antlr.v4.runtime.atn.ATNDeserializer.deserialize(ATNDeserializer.java:153)
        at 
org.apache.spark.sql.catalyst.parser.SqlBaseLexer.<clinit>(SqlBaseLexer.java:1153)
        ... 16 more
   Caused by: java.io.InvalidClassException: org.antlr.v4.runtime.atn.ATN; 
Could not deserialize ATN with UUID 59627784-3be5-417a-b9eb-8131a7286089 
(expected aadb8d7e-aeef-4415-ad2b-8204d6cf042e or a legacy UUID).
        ... 18 more
   
   >>>
   ```
   
   Setup 4: Put the jar from the PR before the older incubating jar (SUCCEEDS 
!!)
   ```
   $ ~/spark-2.3.0-bin-hadoop2.7/bin/pyspark --driver-memory 20g --master 
local[*] --driver-class-path 
systemml-1.3.0-SNAPSHOT-extra-pr.jar:systemml-1.3.0-SNAPSHOT-pr.jar:systemml-0.14.0-incubating.jar
   Python 3.6.3 (default, Mar 20 2018, 13:50:41) 
   [GCC 4.8.5 20150623 (Red Hat 4.8.5-16)] on linux
   Type "help", "copyright", "credits" or "license" for more information.
   2019-03-21 13:19:59 WARN  NativeCodeLoader:62 - Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
   Setting default log level to "WARN".
   To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use 
setLogLevel(newLevel).
   Welcome to
         ____              __
        / __/__  ___ _____/ /__
       _\ \/ _ \/ _ `/ __/  '_/
      /__ / .__/\_,_/_/ /_/\_\   version 2.3.0
         /_/
   
   Using Python version 3.6.3 (default, Mar 20 2018 13:50:41)
   SparkSession available as 'spark'.
   >>> from systemml import MLContext
   >>> ml = MLContext(spark)
   2019-03-21 13:20:22 WARN  ObjectStore:568 - Failed to get database 
global_temp, returning NoSuchObjectException
   
   Welcome to Apache SystemML!
   Version 1.3.0-SNAPSHOT
   >>> ml.version()
   '1.3.0-SNAPSHOT'
   >>> df=spark.read.parquet('shake.parquet')
   >>> df.show()
   +-----+---------+-----+-----+-----+
   |CLASS| SENSORID|    X|    Y|    Z|
   +-----+---------+-----+-----+-----+
   |    2| qqqqqqqq| 0.12| 0.12| 0.12|
   |    2|aUniqueID| 0.03| 0.03| 0.03|
   |    2| qqqqqqqq|-3.84|-3.84|-3.84|
   |    2| 12345678| -0.1| -0.1| -0.1|
   |    2| 12345678|-0.15|-0.15|-0.15|
   |    2| 12345678| 0.47| 0.47| 0.47|
   |    2| 12345678|-0.06|-0.06|-0.06|
   |    2| 12345678|-0.09|-0.09|-0.09|
   |    2| 12345678| 0.21| 0.21| 0.21|
   |    2| 12345678|-0.08|-0.08|-0.08|
   |    2| 12345678| 0.44| 0.44| 0.44|
   |    2|    gholi| 0.76| 0.76| 0.76|
   |    2|    gholi| 1.62| 1.62| 1.62|
   |    2|    gholi| 5.81| 5.81| 5.81|
   |    2| bcbcbcbc| 0.58| 0.58| 0.58|
   |    2| bcbcbcbc|-8.24|-8.24|-8.24|
   |    2| bcbcbcbc|-0.45|-0.45|-0.45|
   |    2| bcbcbcbc| 1.03| 1.03| 1.03|
   |    2|aUniqueID|-0.05|-0.05|-0.05|
   |    2| qqqqqqqq|-0.44|-0.44|-0.44|
   +-----+---------+-----+-----+-----+
   only showing top 20 rows
   
   >>> df.createOrReplaceTempView("df")
   >>>
   ```
   
   Setup 5: No jar provided (SUCCEEDS !!)
   
   ```
   $ ~/spark-2.3.0-bin-hadoop2.7/bin/pyspark --driver-memory 20g --master 
local[*]
   Python 3.6.3 (default, Mar 20 2018, 13:50:41) 
   [GCC 4.8.5 20150623 (Red Hat 4.8.5-16)] on linux
   Type "help", "copyright", "credits" or "license" for more information.
   2019-03-21 13:23:26 WARN  NativeCodeLoader:62 - Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
   Setting default log level to "WARN".
   To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use 
setLogLevel(newLevel).
   Welcome to
         ____              __
        / __/__  ___ _____/ /__
       _\ \/ _ \/ _ `/ __/  '_/
      /__ / .__/\_,_/_/ /_/\_\   version 2.3.0
         /_/
   
   Using Python version 3.6.3 (default, Mar 20 2018 13:50:41)
   SparkSession available as 'spark'.
   >>> from systemml import MLContext
   >>> ml = MLContext(spark)
   2019-03-21 13:23:46 WARN  ObjectStore:568 - Failed to get database 
global_temp, returning NoSuchObjectException
   
   Welcome to Apache SystemML!
   Version 1.2.0
   >>> ml.version()
   '1.2.0'
   >>> df=spark.read.parquet('shake.parquet')
   >>> df.show()
   +-----+---------+-----+-----+-----+
   |CLASS| SENSORID|    X|    Y|    Z|
   +-----+---------+-----+-----+-----+
   |    2| qqqqqqqq| 0.12| 0.12| 0.12|
   |    2|aUniqueID| 0.03| 0.03| 0.03|
   |    2| qqqqqqqq|-3.84|-3.84|-3.84|
   |    2| 12345678| -0.1| -0.1| -0.1|
   |    2| 12345678|-0.15|-0.15|-0.15|
   |    2| 12345678| 0.47| 0.47| 0.47|
   |    2| 12345678|-0.06|-0.06|-0.06|
   |    2| 12345678|-0.09|-0.09|-0.09|
   |    2| 12345678| 0.21| 0.21| 0.21|
   |    2| 12345678|-0.08|-0.08|-0.08|
   |    2| 12345678| 0.44| 0.44| 0.44|
   |    2|    gholi| 0.76| 0.76| 0.76|
   |    2|    gholi| 1.62| 1.62| 1.62|
   |    2|    gholi| 5.81| 5.81| 5.81|
   |    2| bcbcbcbc| 0.58| 0.58| 0.58|
   |    2| bcbcbcbc|-8.24|-8.24|-8.24|
   |    2| bcbcbcbc|-0.45|-0.45|-0.45|
   |    2| bcbcbcbc| 1.03| 1.03| 1.03|
   |    2|aUniqueID|-0.05|-0.05|-0.05|
   |    2| qqqqqqqq|-0.44|-0.44|-0.44|
   +-----+---------+-----+-----+-----+
   only showing top 20 rows
   
   >>> df.createOrReplaceTempView("df")
   >>>
   ```
   
   Setup 6: Provide just `1.2.0` jars (FAILS !!)
   ```
   $ ~/spark-2.3.0-bin-hadoop2.7/bin/pyspark --driver-memory 20g --master 
local[*] --driver-class-path systemml-1.2.0
   systemml-1.2.0-extra.jar  systemml-1.2.0.jar        
   [npansar@dml3 debug_classpath]$ ~/spark-2.3.0-bin-hadoop2.7/bin/pyspark 
--driver-memory 20g --master local[*] --driver-class-path 
systemml-1.2.0.jar:systemml-1.2.0-extra.jar
   Python 3.6.3 (default, Mar 20 2018, 13:50:41) 
   [GCC 4.8.5 20150623 (Red Hat 4.8.5-16)] on linux
   Type "help", "copyright", "credits" or "license" for more information.
   2019-03-21 13:32:09 WARN  NativeCodeLoader:62 - Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
   Setting default log level to "WARN".
   To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use 
setLogLevel(newLevel).
   Welcome to
         ____              __
        / __/__  ___ _____/ /__
       _\ \/ _ \/ _ `/ __/  '_/
      /__ / .__/\_,_/_/ /_/\_\   version 2.3.0
         /_/
   
   Using Python version 3.6.3 (default, Mar 20 2018 13:50:41)
   SparkSession available as 'spark'.
   >>> from systemml import MLContext
   >>> ml = MLContext(spark)
   2019-03-21 13:32:25 WARN  ObjectStore:568 - Failed to get database 
global_temp, returning NoSuchObjectException
   
   Welcome to Apache SystemML!
   Version 1.2.0
   >>> ml.version()
   '1.2.0'
   >>> df=spark.read.parquet('shake.parquet')
   >>> df.show()
   +-----+---------+-----+-----+-----+
   |CLASS| SENSORID|    X|    Y|    Z|
   +-----+---------+-----+-----+-----+
   |    2| qqqqqqqq| 0.12| 0.12| 0.12|
   |    2|aUniqueID| 0.03| 0.03| 0.03|
   |    2| qqqqqqqq|-3.84|-3.84|-3.84|
   |    2| 12345678| -0.1| -0.1| -0.1|
   |    2| 12345678|-0.15|-0.15|-0.15|
   |    2| 12345678| 0.47| 0.47| 0.47|
   |    2| 12345678|-0.06|-0.06|-0.06|
   |    2| 12345678|-0.09|-0.09|-0.09|
   |    2| 12345678| 0.21| 0.21| 0.21|
   |    2| 12345678|-0.08|-0.08|-0.08|
   |    2| 12345678| 0.44| 0.44| 0.44|
   |    2|    gholi| 0.76| 0.76| 0.76|
   |    2|    gholi| 1.62| 1.62| 1.62|
   |    2|    gholi| 5.81| 5.81| 5.81|
   |    2| bcbcbcbc| 0.58| 0.58| 0.58|
   |    2| bcbcbcbc|-8.24|-8.24|-8.24|
   |    2| bcbcbcbc|-0.45|-0.45|-0.45|
   |    2| bcbcbcbc| 1.03| 1.03| 1.03|
   |    2|aUniqueID|-0.05|-0.05|-0.05|
   |    2| qqqqqqqq|-0.44|-0.44|-0.44|
   +-----+---------+-----+-----+-----+
   only showing top 20 rows
   
   >>> df.createOrReplaceTempView("df")
   ANTLR Tool version 4.7 used for code generation does not match the current 
runtime version 4.5.3ANTLR Runtime version 4.7 used for parser compilation does 
not match the current runtime version 4.5.3Traceback (most recent call last):
     File "<stdin>", line 1, in <module>
     File 
"/home/npansar/spark-2.3.0-bin-hadoop2.7/python/pyspark/sql/dataframe.py", line 
176, in createOrReplaceTempView
       self._jdf.createOrReplaceTempView(name)
     File 
"/home/npansar/spark-2.3.0-bin-hadoop2.7/python/lib/py4j-0.10.6-src.zip/py4j/java_gateway.py",
 line 1160, in __call__
     File 
"/home/npansar/spark-2.3.0-bin-hadoop2.7/python/pyspark/sql/utils.py", line 63, 
in deco
       return f(*a, **kw)
     File 
"/home/npansar/spark-2.3.0-bin-hadoop2.7/python/lib/py4j-0.10.6-src.zip/py4j/protocol.py",
 line 320, in get_return_value
   py4j.protocol.Py4JJavaError: An error occurred while calling 
o52.createOrReplaceTempView.
   : java.lang.ExceptionInInitializerError
        at 
org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parse(ParseDriver.scala:84)
        at 
org.apache.spark.sql.execution.SparkSqlParser.parse(SparkSqlParser.scala:48)
        at 
org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parseTableIdentifier(ParseDriver.scala:49)
        at 
org.apache.spark.sql.Dataset.createTempViewCommand(Dataset.scala:3079)
        at 
org.apache.spark.sql.Dataset.createOrReplaceTempView(Dataset.scala:3034)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
        at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
        at py4j.Gateway.invoke(Gateway.java:282)
        at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
        at py4j.commands.CallCommand.execute(CallCommand.java:79)
        at py4j.GatewayConnection.run(GatewayConnection.java:214)
        at java.lang.Thread.run(Thread.java:748)
   Caused by: java.lang.UnsupportedOperationException: 
java.io.InvalidClassException: org.antlr.v4.runtime.atn.ATN; Could not 
deserialize ATN with UUID 59627784-3be5-417a-b9eb-8131a7286089 (expected 
aadb8d7e-aeef-4415-ad2b-8204d6cf042e or a legacy UUID).
        at 
org.antlr.v4.runtime.atn.ATNDeserializer.deserialize(ATNDeserializer.java:153)
        at 
org.apache.spark.sql.catalyst.parser.SqlBaseLexer.<clinit>(SqlBaseLexer.java:1153)
        ... 16 more
   Caused by: java.io.InvalidClassException: org.antlr.v4.runtime.atn.ATN; 
Could not deserialize ATN with UUID 59627784-3be5-417a-b9eb-8131a7286089 
(expected aadb8d7e-aeef-4415-ad2b-8204d6cf042e or a legacy UUID).
        ... 18 more
   
   >>> 
   ```

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
[email protected]


With regards,
Apache Git Services

Reply via email to