I am running code in a local machine that is single node machine. Getting into logs, it looked like the host is killed. This is happening very frequently an I am unable to find the reason of this.
Could low memory be the reason? On Fri, 18 Dec 2020, 00:11 Patrick McCarthy, <pmccar...@dstillery.com> wrote: > 'Job aborted due to stage failure: Task 1 in stage 39.0 failed 1 times' > > You may want to change the number of failures to a higher number like 4. A > single failure on a task should be able to be tolerated, especially if > you're on a shared cluster where resources can be preempted. > > It seems that a node dies or goes off the network, so perhaps you can > also debug the logs on the failing node to see why it disappears and > prevent the failures in the first place. > > On Thu, Dec 17, 2020 at 1:27 PM Vikas Garg <sperry...@gmail.com> wrote: > >> Mydomain is named by me while pasting the logs >> >> Also, there are multiple class files in my project, if I run any 1 or 2 >> at a time, then they run fine, sometimes they too give this error. But >> running all the classes at the same time always give this error. >> >> Once this error come, I can't run any program and on restarting the >> system, program starts running fine. >> This error goes away on >> >> On Thu, 17 Dec 2020, 23:50 Patrick McCarthy, <pmccar...@dstillery.com> >> wrote: >> >>> my-domain.com/192.168.166.8:63534 probably isn't a valid address on >>> your network, is it? >>> >>> On Thu, Dec 17, 2020 at 3:03 AM Vikas Garg <sperry...@gmail.com> wrote: >>> >>>> Hi, >>>> >>>> Since last few days, I am getting error message while running my >>>> project. I have searched Google for the solution but didn't got any help. >>>> >>>> Can someone help me to figure out how I could mitigate this issue? >>>> >>>> >>>> 20/12/17 13:26:57 ERROR RetryingBlockFetcher: Exception while beginning >>>> fetch of 1 outstanding blocks >>>> >>>> *java.io.IOException*: Failed to connect to >>>> my-domain.com/192.168.166.8:63534 >>>> >>>> at >>>> org.apache.spark.network.client.TransportClientFactory.createClient( >>>> *TransportClientFactory.java:253*) >>>> >>>> at >>>> org.apache.spark.network.client.TransportClientFactory.createClient( >>>> *TransportClientFactory.java:195*) >>>> >>>> at >>>> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart( >>>> *NettyBlockTransferService.scala:122*) >>>> >>>> at >>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding( >>>> *RetryingBlockFetcher.java:141*) >>>> >>>> at org.apache.spark.network.shuffle.RetryingBlockFetcher.start( >>>> *RetryingBlockFetcher.java:121*) >>>> >>>> at >>>> org.apache.spark.network.netty.NettyBlockTransferService.fetchBlocks( >>>> *NettyBlockTransferService.scala:143*) >>>> >>>> at org.apache.spark.network.BlockTransferService.fetchBlockSync( >>>> *BlockTransferService.scala:103*) >>>> >>>> at >>>> org.apache.spark.storage.BlockManager.fetchRemoteManagedBuffer( >>>> *BlockManager.scala:1010*) >>>> >>>> at >>>> org.apache.spark.storage.BlockManager.$anonfun$getRemoteBlock$8( >>>> *BlockManager.scala:954*) >>>> >>>> at scala.Option.orElse(*Option.scala:447*) >>>> >>>> at org.apache.spark.storage.BlockManager.getRemoteBlock( >>>> *BlockManager.scala:954*) >>>> >>>> at org.apache.spark.storage.BlockManager.getRemoteBytes( >>>> *BlockManager.scala:1092*) >>>> >>>> at >>>> org.apache.spark.scheduler.TaskResultGetter$$anon$3.$anonfun$run$1( >>>> *TaskResultGetter.scala:88*) >>>> >>>> at >>>> scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) >>>> >>>> at org.apache.spark.util.Utils$.logUncaughtExceptions( >>>> *Utils.scala:1932*) >>>> >>>> at org.apache.spark.scheduler.TaskResultGetter$$anon$3.run( >>>> *TaskResultGetter.scala:63*) >>>> >>>> at java.util.concurrent.ThreadPoolExecutor.runWorker( >>>> *ThreadPoolExecutor.java:1149*) >>>> >>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >>>> *ThreadPoolExecutor.java:624*) >>>> >>>> at java.lang.Thread.run(*Thread.java:748*) >>>> >>>> Caused by: *io.netty.channel.AbstractChannel$AnnotatedSocketException*: >>>> Permission denied: no further information: >>>> my-domain.com/192.168.166.8:63534 >>>> >>>> Caused by: *java.net.SocketException*: Permission denied: no further >>>> information >>>> >>>> at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*) >>>> >>>> at sun.nio.ch.SocketChannelImpl.finishConnect( >>>> *SocketChannelImpl.java:715*) >>>> >>>> at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect( >>>> *NioSocketChannel.java:330*) >>>> >>>> at >>>> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect( >>>> *AbstractNioChannel.java:334*) >>>> >>>> at io.netty.channel.nio.NioEventLoop.processSelectedKey( >>>> *NioEventLoop.java:702*) >>>> >>>> at >>>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized( >>>> *NioEventLoop.java:650*) >>>> >>>> at io.netty.channel.nio.NioEventLoop.processSelectedKeys( >>>> *NioEventLoop.java:576*) >>>> >>>> at io.netty.channel.nio.NioEventLoop.run( >>>> *NioEventLoop.java:493*) >>>> >>>> at io.netty.util.concurrent.SingleThreadEventExecutor$4.run( >>>> *SingleThreadEventExecutor.java:989*) >>>> >>>> at io.netty.util.internal.ThreadExecutorMap$2.run( >>>> *ThreadExecutorMap.java:74*) >>>> >>>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>>> *FastThreadLocalRunnable.java:30*) >>>> >>>> at java.lang.Thread.run(*Thread.java:748*) >>>> >>>> 20/12/17 13:27:02 ERROR RetryingBlockFetcher: Exception while beginning >>>> fetch of 1 outstanding blocks (after 1 retries) >>>> >>>> *java.io.IOException*: Failed to connect to >>>> my-domain.com/192.168.166.8:63534 >>>> >>>> at >>>> org.apache.spark.network.client.TransportClientFactory.createClient( >>>> *TransportClientFactory.java:253*) >>>> >>>> at >>>> org.apache.spark.network.client.TransportClientFactory.createClient( >>>> *TransportClientFactory.java:195*) >>>> >>>> at >>>> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart( >>>> *NettyBlockTransferService.scala:122*) >>>> >>>> at >>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding( >>>> *RetryingBlockFetcher.java:141*) >>>> >>>> at >>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.lambda$initiateRetry$0( >>>> *RetryingBlockFetcher.java:169*) >>>> >>>> at java.util.concurrent.Executors$RunnableAdapter.call( >>>> *Executors.java:511*) >>>> >>>> at java.util.concurrent.FutureTask.run(*FutureTask.java:266*) >>>> >>>> at java.util.concurrent.ThreadPoolExecutor.runWorker( >>>> *ThreadPoolExecutor.java:1149*) >>>> >>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >>>> *ThreadPoolExecutor.java:624*) >>>> >>>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>>> *FastThreadLocalRunnable.java:30*) >>>> >>>> at java.lang.Thread.run(*Thread.java:748*) >>>> >>>> Caused by: *io.netty.channel.AbstractChannel$AnnotatedSocketException*: >>>> Permission denied: no further information: >>>> my-domain.com/192.168.166.8:63534 >>>> >>>> Caused by: *java.net.SocketException*: Permission denied: no further >>>> information >>>> >>>> at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*) >>>> >>>> at sun.nio.ch.SocketChannelImpl.finishConnect( >>>> *SocketChannelImpl.java:715*) >>>> >>>> at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect( >>>> *NioSocketChannel.java:330*) >>>> >>>> at >>>> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect( >>>> *AbstractNioChannel.java:334*) >>>> >>>> at io.netty.channel.nio.NioEventLoop.processSelectedKey( >>>> *NioEventLoop.java:702*) >>>> >>>> at >>>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized( >>>> *NioEventLoop.java:650*) >>>> >>>> at io.netty.channel.nio.NioEventLoop.processSelectedKeys( >>>> *NioEventLoop.java:576*) >>>> >>>> at io.netty.channel.nio.NioEventLoop.run( >>>> *NioEventLoop.java:493*) >>>> >>>> at io.netty.util.concurrent.SingleThreadEventExecutor$4.run( >>>> *SingleThreadEventExecutor.java:989*) >>>> >>>> at io.netty.util.internal.ThreadExecutorMap$2.run( >>>> *ThreadExecutorMap.java:74*) >>>> >>>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>>> *FastThreadLocalRunnable.java:30*) >>>> >>>> at java.lang.Thread.run(*Thread.java:748*) >>>> >>>> 20/12/17 13:27:07 ERROR RetryingBlockFetcher: Exception while beginning >>>> fetch of 1 outstanding blocks (after 2 retries) >>>> >>>> *java.io.IOException*: Failed to connect to >>>> my-domain.com/192.168.166.8:63534 >>>> >>>> at >>>> org.apache.spark.network.client.TransportClientFactory.createClient( >>>> *TransportClientFactory.java:253*) >>>> >>>> at >>>> org.apache.spark.network.client.TransportClientFactory.createClient( >>>> *TransportClientFactory.java:195*) >>>> >>>> at >>>> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart( >>>> *NettyBlockTransferService.scala:122*) >>>> >>>> at >>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding( >>>> *RetryingBlockFetcher.java:141*) >>>> >>>> at >>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.lambda$initiateRetry$0( >>>> *RetryingBlockFetcher.java:169*) >>>> >>>> at java.util.concurrent.Executors$RunnableAdapter.call( >>>> *Executors.java:511*) >>>> >>>> at java.util.concurrent.FutureTask.run(*FutureTask.java:266*) >>>> >>>> at java.util.concurrent.ThreadPoolExecutor.runWorker( >>>> *ThreadPoolExecutor.java:1149*) >>>> >>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >>>> *ThreadPoolExecutor.java:624*) >>>> >>>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>>> *FastThreadLocalRunnable.java:30*) >>>> >>>> at java.lang.Thread.run(*Thread.java:748*) >>>> >>>> Caused by: *io.netty.channel.AbstractChannel$AnnotatedSocketException*: >>>> Permission denied: no further information: >>>> my-domain.com/192.168.166.8:63534 >>>> >>>> Caused by: *java.net.SocketException*: Permission denied: no further >>>> information >>>> >>>> at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*) >>>> >>>> at sun.nio.ch.SocketChannelImpl.finishConnect( >>>> *SocketChannelImpl.java:715*) >>>> >>>> at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect( >>>> *NioSocketChannel.java:330*) >>>> >>>> at >>>> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect( >>>> *AbstractNioChannel.java:334*) >>>> >>>> at io.netty.channel.nio.NioEventLoop.processSelectedKey( >>>> *NioEventLoop.java:702*) >>>> >>>> at >>>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized( >>>> *NioEventLoop.java:650*) >>>> >>>> at io.netty.channel.nio.NioEventLoop.processSelectedKeys( >>>> *NioEventLoop.java:576*) >>>> >>>> at io.netty.channel.nio.NioEventLoop.run( >>>> *NioEventLoop.java:493*) >>>> >>>> at io.netty.util.concurrent.SingleThreadEventExecutor$4.run( >>>> *SingleThreadEventExecutor.java:989*) >>>> >>>> at io.netty.util.internal.ThreadExecutorMap$2.run( >>>> *ThreadExecutorMap.java:74*) >>>> >>>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>>> *FastThreadLocalRunnable.java:30*) >>>> >>>> at java.lang.Thread.run(*Thread.java:748*) >>>> >>>> 20/12/17 13:27:12 ERROR RetryingBlockFetcher: Exception while beginning >>>> fetch of 1 outstanding blocks (after 3 retries) >>>> >>>> *java.io.IOException*: Failed to connect to >>>> my-domain.com/192.168.166.8:63534 >>>> >>>> at >>>> org.apache.spark.network.client.TransportClientFactory.createClient( >>>> *TransportClientFactory.java:253*) >>>> >>>> at >>>> org.apache.spark.network.client.TransportClientFactory.createClient( >>>> *TransportClientFactory.java:195*) >>>> >>>> at >>>> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart( >>>> *NettyBlockTransferService.scala:122*) >>>> >>>> at >>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding( >>>> *RetryingBlockFetcher.java:141*) >>>> >>>> at >>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.lambda$initiateRetry$0( >>>> *RetryingBlockFetcher.java:169*) >>>> >>>> at java.util.concurrent.Executors$RunnableAdapter.call( >>>> *Executors.java:511*) >>>> >>>> at java.util.concurrent.FutureTask.run(*FutureTask.java:266*) >>>> >>>> at java.util.concurrent.ThreadPoolExecutor.runWorker( >>>> *ThreadPoolExecutor.java:1149*) >>>> >>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >>>> *ThreadPoolExecutor.java:624*) >>>> >>>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>>> *FastThreadLocalRunnable.java:30*) >>>> >>>> at java.lang.Thread.run(*Thread.java:748*) >>>> >>>> Caused by: *io.netty.channel.AbstractChannel$AnnotatedSocketException*: >>>> Permission denied: no further information: >>>> my-domain.com/192.168.166.8:63534 >>>> >>>> Caused by: *java.net.SocketException*: Permission denied: no further >>>> information >>>> >>>> at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*) >>>> >>>> at sun.nio.ch.SocketChannelImpl.finishConnect( >>>> *SocketChannelImpl.java:715*) >>>> >>>> at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect( >>>> *NioSocketChannel.java:330*) >>>> >>>> at >>>> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect( >>>> *AbstractNioChannel.java:334*) >>>> >>>> at io.netty.channel.nio.NioEventLoop.processSelectedKey( >>>> *NioEventLoop.java:702*) >>>> >>>> at >>>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized( >>>> *NioEventLoop.java:650*) >>>> >>>> at io.netty.channel.nio.NioEventLoop.processSelectedKeys( >>>> *NioEventLoop.java:576*) >>>> >>>> at io.netty.channel.nio.NioEventLoop.run( >>>> *NioEventLoop.java:493*) >>>> >>>> at io.netty.util.concurrent.SingleThreadEventExecutor$4.run( >>>> *SingleThreadEventExecutor.java:989*) >>>> >>>> at io.netty.util.internal.ThreadExecutorMap$2.run( >>>> *ThreadExecutorMap.java:74*) >>>> >>>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>>> *FastThreadLocalRunnable.java:30*) >>>> >>>> at java.lang.Thread.run(*Thread.java:748*) >>>> >>>> 20/12/17 13:27:12 ERROR TaskSetManager: Task 1 in stage 39.0 failed 1 >>>> times; aborting job >>>> >>>> 20/12/17 13:27:12 ERROR FileFormatWriter: Aborting job >>>> 0db8a16a-e1d3-4091-b725-0a8bdc17b0b1. >>>> >>>> *java.util.concurrent.ExecutionException*: >>>> *org.apache.spark.SparkException*: Job aborted due to stage failure: >>>> Task 1 in stage 39.0 failed 1 times, most recent failure: Lost task 1.0 in >>>> stage 39.0 (TID 655, my-domain.com, executor driver): TaskResultLost >>>> (result lost from block manager) >>>> >>>> Driver stacktrace: >>>> >>>> at java.util.concurrent.FutureTask.report(*FutureTask.java:122* >>>> ) >>>> >>>> at java.util.concurrent.FutureTask.get(*FutureTask.java:206*) >>>> >>>> at >>>> org.apache.spark.sql.execution.exchange.BroadcastExchangeExec.doExecuteBroadcast( >>>> *BroadcastExchangeExec.scala:195*) >>>> >>>> at >>>> org.apache.spark.sql.execution.InputAdapter.doExecuteBroadcast( >>>> *WholeStageCodegenExec.scala:515*) >>>> >>>> at >>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeBroadcast$1( >>>> *SparkPlan.scala:188*) >>>> >>>> at >>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>>> *SparkPlan.scala:213*) >>>> >>>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>>> *RDDOperationScope.scala:151*) >>>> >>>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>>> *SparkPlan.scala:210*) >>>> >>>> at org.apache.spark.sql.execution.SparkPlan.executeBroadcast( >>>> *SparkPlan.scala:184*) >>>> >>>> at >>>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.prepareBroadcast( >>>> *BroadcastHashJoinExec.scala:116*) >>>> >>>> at >>>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.codegenInner( >>>> *BroadcastHashJoinExec.scala:210*) >>>> >>>> at >>>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.doConsume( >>>> *BroadcastHashJoinExec.scala:100*) >>>> >>>> at org.apache.spark.sql.execution.CodegenSupport.consume( >>>> *WholeStageCodegenExec.scala:194*) >>>> >>>> at org.apache.spark.sql.execution.CodegenSupport.consume$( >>>> *WholeStageCodegenExec.scala:149*) >>>> >>>> at org.apache.spark.sql.execution.ProjectExec.consume( >>>> *basicPhysicalOperators.scala:41*) >>>> >>>> at org.apache.spark.sql.execution.ProjectExec.doConsume( >>>> *basicPhysicalOperators.scala:71*) >>>> >>>> at org.apache.spark.sql.execution.CodegenSupport.consume( >>>> *WholeStageCodegenExec.scala:194*) >>>> >>>> at org.apache.spark.sql.execution.CodegenSupport.consume$( >>>> *WholeStageCodegenExec.scala:149*) >>>> >>>> at org.apache.spark.sql.execution.FilterExec.consume( >>>> *basicPhysicalOperators.scala:97*) >>>> >>>> at org.apache.spark.sql.execution.FilterExec.doConsume( >>>> *basicPhysicalOperators.scala:222*) >>>> >>>> at org.apache.spark.sql.execution.CodegenSupport.consume( >>>> *WholeStageCodegenExec.scala:194*) >>>> >>>> at org.apache.spark.sql.execution.CodegenSupport.consume$( >>>> *WholeStageCodegenExec.scala:149*) >>>> >>>> at org.apache.spark.sql.execution.ColumnarToRowExec.consume( >>>> *Columnar.scala:60*) >>>> >>>> at org.apache.spark.sql.execution.ColumnarToRowExec.doProduce( >>>> *Columnar.scala:185*) >>>> >>>> at >>>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >>>> *WholeStageCodegenExec.scala:95*) >>>> >>>> at >>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>>> *SparkPlan.scala:213*) >>>> >>>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>>> *RDDOperationScope.scala:151*) >>>> >>>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>>> *SparkPlan.scala:210*) >>>> >>>> at org.apache.spark.sql.execution.CodegenSupport.produce( >>>> *WholeStageCodegenExec.scala:90*) >>>> >>>> at org.apache.spark.sql.execution.CodegenSupport.produce$( >>>> *WholeStageCodegenExec.scala:90*) >>>> >>>> at org.apache.spark.sql.execution.ColumnarToRowExec.produce( >>>> *Columnar.scala:60*) >>>> >>>> at org.apache.spark.sql.execution.FilterExec.doProduce( >>>> *basicPhysicalOperators.scala:137*) >>>> >>>> at >>>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >>>> *WholeStageCodegenExec.scala:95*) >>>> >>>> at >>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>>> *SparkPlan.scala:213*) >>>> >>>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>>> *RDDOperationScope.scala:151*) >>>> >>>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>>> *SparkPlan.scala:210*) >>>> >>>> at org.apache.spark.sql.execution.CodegenSupport.produce( >>>> *WholeStageCodegenExec.scala:90*) >>>> >>>> at org.apache.spark.sql.execution.CodegenSupport.produce$( >>>> *WholeStageCodegenExec.scala:90*) >>>> >>>> at org.apache.spark.sql.execution.FilterExec.produce( >>>> *basicPhysicalOperators.scala:97*) >>>> >>>> at org.apache.spark.sql.execution.ProjectExec.doProduce( >>>> *basicPhysicalOperators.scala:51*) >>>> >>>> at >>>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >>>> *WholeStageCodegenExec.scala:95*) >>>> >>>> at >>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>>> *SparkPlan.scala:213*) >>>> >>>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>>> *RDDOperationScope.scala:151*) >>>> >>>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>>> *SparkPlan.scala:210*) >>>> >>>> at org.apache.spark.sql.execution.CodegenSupport.produce( >>>> *WholeStageCodegenExec.scala:90*) >>>> >>>> at org.apache.spark.sql.execution.CodegenSupport.produce$( >>>> *WholeStageCodegenExec.scala:90*) >>>> >>>> at org.apache.spark.sql.execution.ProjectExec.produce( >>>> *basicPhysicalOperators.scala:41*) >>>> >>>> at >>>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.doProduce( >>>> *BroadcastHashJoinExec.scala:95*) >>>> >>>> at >>>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >>>> *WholeStageCodegenExec.scala:95*) >>>> >>>> at >>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>>> *SparkPlan.scala:213*) >>>> >>>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>>> *RDDOperationScope.scala:151*) >>>> >>>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>>> *SparkPlan.scala:210*) >>>> >>>> at org.apache.spark.sql.execution.CodegenSupport.produce( >>>> *WholeStageCodegenExec.scala:90*) >>>> >>>> at org.apache.spark.sql.execution.CodegenSupport.produce$( >>>> *WholeStageCodegenExec.scala:90*) >>>> >>>> at >>>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.produce( >>>> *BroadcastHashJoinExec.scala:39*) >>>> >>>> at org.apache.spark.sql.execution.ProjectExec.doProduce( >>>> *basicPhysicalOperators.scala:51*) >>>> >>>> at >>>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >>>> *WholeStageCodegenExec.scala:95*) >>>> >>>> at >>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>>> *SparkPlan.scala:213*) >>>> >>>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>>> *RDDOperationScope.scala:151*) >>>> >>>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>>> *SparkPlan.scala:210*) >>>> >>>> at org.apache.spark.sql.execution.CodegenSupport.produce( >>>> *WholeStageCodegenExec.scala:90*) >>>> >>>> at org.apache.spark.sql.execution.CodegenSupport.produce$( >>>> *WholeStageCodegenExec.scala:90*) >>>> >>>> at org.apache.spark.sql.execution.ProjectExec.produce( >>>> *basicPhysicalOperators.scala:41*) >>>> >>>> at >>>> org.apache.spark.sql.execution.WholeStageCodegenExec.doCodeGen( >>>> *WholeStageCodegenExec.scala:632*) >>>> >>>> at >>>> org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute( >>>> *WholeStageCodegenExec.scala:692*) >>>> >>>> at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1( >>>> *SparkPlan.scala:175*) >>>> >>>> at >>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>>> *SparkPlan.scala:213*) >>>> >>>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>>> *RDDOperationScope.scala:151*) >>>> >>>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>>> *SparkPlan.scala:210*) >>>> >>>> at org.apache.spark.sql.execution.SparkPlan.execute( >>>> *SparkPlan.scala:171*) >>>> >>>> at >>>> org.apache.spark.sql.execution.datasources.FileFormatWriter$.write( >>>> *FileFormatWriter.scala:172*) >>>> >>>> at >>>> org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand.run( >>>> *InsertIntoHadoopFsRelationCommand.scala:178*) >>>> >>>> at >>>> org.apache.spark.sql.execution.command.DataWritingCommandExec.sideEffectResult$lzycompute( >>>> *commands.scala:108*) >>>> >>>> at >>>> org.apache.spark.sql.execution.command.DataWritingCommandExec.sideEffectResult( >>>> *commands.scala:106*) >>>> >>>> at >>>> org.apache.spark.sql.execution.command.DataWritingCommandExec.doExecute( >>>> *commands.scala:131*) >>>> >>>> at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1( >>>> *SparkPlan.scala:175*) >>>> >>>> at >>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>>> *SparkPlan.scala:213*) >>>> >>>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>>> *RDDOperationScope.scala:151*) >>>> >>>> at org.apache.spark.sql.execu >>>> >>> > > -- > > > *Patrick McCarthy * > > Senior Data Scientist, Machine Learning Engineering > > Dstillery > > 470 Park Ave South, 17th Floor, NYC 10016 >