Re: org.apache.spark.SparkException: Task failed while writing rows
Yeah, without actually seeing what's happening on that line, it'd be difficult to say for sure. You can check what patches HortonWorks applied, or/and ask them. And yeah, seg fault is totally possible on any size of the data. But you should've seen it in the `stdout` (assuming that the regular logs go to `stderr`) On Wed, Feb 28, 2018 at 2:53 PM, unk1102 wrote: > Hi Vadim thanks I use HortonWorks package. I dont think there are any seg > faults are dataframe I am trying to write is very small in size. Can it > still create seg fault? > > > > -- > Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ > > - > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > >
Re: org.apache.spark.SparkException: Task failed while writing rows
Hi Vadim thanks I use HortonWorks package. I dont think there are any seg faults are dataframe I am trying to write is very small in size. Can it still create seg fault? -- Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ - To unsubscribe e-mail: user-unsubscr...@spark.apache.org
Re: org.apache.spark.SparkException: Task failed while writing rows
Who's your spark provider? EMR, Azure, Databricks, etc.? Maybe contact them, since they've probably applied some patches Also have you checked `stdout` for some Segfaults? I vaguely remember getting `Task failed while writing rows at` and seeing some segfaults that caused that On Wed, Feb 28, 2018 at 2:07 PM, unk1102 wrote: > Hi thanks Vadim you are right I saw that line already 468 I dont see any > code > it is just comment yes I am sure I am using all spark-* jar which is built > for spark 2.2.0 and Scala 2.11. I am also stuck unfortunately with these > errors not sure how to solve them. > > > > -- > Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ > > - > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > >
Re: org.apache.spark.SparkException: Task failed while writing rows
Hi thanks Vadim you are right I saw that line already 468 I dont see any code it is just comment yes I am sure I am using all spark-* jar which is built for spark 2.2.0 and Scala 2.11. I am also stuck unfortunately with these errors not sure how to solve them. -- Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ - To unsubscribe e-mail: user-unsubscr...@spark.apache.org
Re: org.apache.spark.SparkException: Task failed while writing rows
I'm sorry, didn't see `Caused by: java.lang.NullPointerException at org.apache.spark.sql.SparkSession$$anonfun$3.apply(SparkSession.scala:468)` Are you sure that you use 2.2.0? I don't see any code on that line https://github.com/apache/spark/blob/v2.2.0/sql/core/src/main/scala/org/apache/spark/sql/SparkSession.scala#L468 Also pretty strange that it fails there On Wed, Feb 28, 2018 at 1:55 PM, unk1102 wrote: > Hi thanks for the reply I only see NPE and Task failed while writing rows > all > over places I dont see any other errors expect SparkException job aborted > and followed by two exception I pasted earlier. > > > > -- > Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ > > - > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > >
Re: org.apache.spark.SparkException: Task failed while writing rows
Hi thanks for the reply I only see NPE and Task failed while writing rows all over places I dont see any other errors expect SparkException job aborted and followed by two exception I pasted earlier. -- Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ - To unsubscribe e-mail: user-unsubscr...@spark.apache.org
Re: org.apache.spark.SparkException: Task failed while writing rows
There should be another exception trace (basically, the actual cause) after this one, could you post it? On Wed, Feb 28, 2018 at 1:39 PM, unk1102 wrote: > Hi I am getting the following exception when I try to write DataFrame using > the following code. Please guide. I am using Spark 2.2.0. > > df.write.format("parquet").mode(SaveMode.Append); > > org.apache.spark.SparkException: Task failed while writing rows at > org.apache.spark.sql.execution.datasources.FileFormatWriter$.org$apache$ > spark$sql$execution$datasources$FileFormatWriter$$ > executeTask(FileFormatWriter.scala:270) > at > org.apache.spark.sql.execution.datasources.FileFormatWriter$$anonfun$ > write$1$$anonfun$apply$mcV$sp$1.apply(FileFormatWriter.scala:189) > at > org.apache.spark.sql.execution.datasources.FileFormatWriter$$anonfun$ > write$1$$anonfun$apply$mcV$sp$1.apply(FileFormatWriter.scala:188) > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87) at > org.apache.spark.scheduler.Task.run(Task.scala:108) at > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338) at > java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) Caused by: > java.lang.NullPointerException at > org.apache.spark.sql.SparkSession$$anonfun$3.apply(SparkSession.scala:468) > at > org.apache.spark.sql.SparkSession$$anonfun$3.apply(SparkSession.scala:468) > at scala.collection.Iterator$$anon$11.next(Iterator.scala:409) at > scala.collection.Iterator$$anon$11.next(Iterator.scala:409) at > org.apache.spark.sql.execution.datasources.FileFormatWriter$ > SingleDirectoryWriteTask.execute(FileFormatWriter.scala:324) > at > org.apache.spark.sql.execution.datasources.FileFormatWriter$$anonfun$org$ > apache$spark$sql$execution$datasources$FileFormatWriter$$ > executeTask$3.apply(FileFormatWriter.scala:256) > at > org.apache.spark.sql.execution.datasources.FileFormatWriter$$anonfun$org$ > apache$spark$sql$execution$datasources$FileFormatWriter$$ > executeTask$3.apply(FileFormatWriter.scala:254) > at > org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCa > llbacks(Utils.scala:1371) > at > org.apache.spark.sql.execution.datasources.FileFormatWriter$.org$apache$ > spark$sql$execution$datasources$FileFormatWriter$$ > executeTask(FileFormatWriter.scala:259) > > > > -- > Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ > > - > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > >