Hi all, I have been testing Spark ML algorithms with bigger dataset, and ran into some problems with linear regression:
It seems the executors stop without any apparent reason: 15/04/22 20:15:05 INFO BlockManagerInfo: Added rdd_12492_80 in memory on backend-node:48037 (size: 28.5 MB, free: 2.8 GB) 15/04/22 20:15:05 INFO BlockManagerInfo: Added rdd_12493_80 in memory on backend-node:48037 (size: 37.6 MB, free: 2.7 GB) 15/04/22 20:15:08 INFO BlockManagerInfo: Added rdd_12489_81 in memory on backend-node:48037 (size: 8.4 MB, free: 2.7 GB) [E 150422 20:15:12 java_gateway:483] Error while sending or receiving. Traceback (most recent call last): File "/home/azureuser/spark-1.3.0-bin-hadoop2.4/python/lib/py4j-0.8.2.1-src.zip/py4j/java_gateway.py", line 479, in send_command raise Py4JError("Answer from Java side is empty") Py4JError: Answer from Java side is empty Then sparkcontext stops, too : [E 150422 20:15:12 java_gateway:431] An error occurred while trying to connect to the Java server the problem is that it does not happen all the time, it only fails maybe once in every five attempts. any suggestions where can I get more detailed logs from? Thanks, -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/ML-regression-spark-context-dies-without-error-tp22633.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org