'Job aborted due to stage failure: Task 1 in stage 39.0 failed 1 times' You may want to change the number of failures to a higher number like 4. A single failure on a task should be able to be tolerated, especially if you're on a shared cluster where resources can be preempted.
It seems that a node dies or goes off the network, so perhaps you can also debug the logs on the failing node to see why it disappears and prevent the failures in the first place. On Thu, Dec 17, 2020 at 1:27 PM Vikas Garg <sperry...@gmail.com> wrote: > Mydomain is named by me while pasting the logs > > Also, there are multiple class files in my project, if I run any 1 or 2 > at a time, then they run fine, sometimes they too give this error. But > running all the classes at the same time always give this error. > > Once this error come, I can't run any program and on restarting the > system, program starts running fine. > This error goes away on > > On Thu, 17 Dec 2020, 23:50 Patrick McCarthy, <pmccar...@dstillery.com> > wrote: > >> my-domain.com/192.168.166.8:63534 probably isn't a valid address on your >> network, is it? >> >> On Thu, Dec 17, 2020 at 3:03 AM Vikas Garg <sperry...@gmail.com> wrote: >> >>> Hi, >>> >>> Since last few days, I am getting error message while running my >>> project. I have searched Google for the solution but didn't got any help. >>> >>> Can someone help me to figure out how I could mitigate this issue? >>> >>> >>> 20/12/17 13:26:57 ERROR RetryingBlockFetcher: Exception while beginning >>> fetch of 1 outstanding blocks >>> >>> *java.io.IOException*: Failed to connect to >>> my-domain.com/192.168.166.8:63534 >>> >>> at >>> org.apache.spark.network.client.TransportClientFactory.createClient( >>> *TransportClientFactory.java:253*) >>> >>> at >>> org.apache.spark.network.client.TransportClientFactory.createClient( >>> *TransportClientFactory.java:195*) >>> >>> at >>> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart( >>> *NettyBlockTransferService.scala:122*) >>> >>> at >>> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding( >>> *RetryingBlockFetcher.java:141*) >>> >>> at org.apache.spark.network.shuffle.RetryingBlockFetcher.start( >>> *RetryingBlockFetcher.java:121*) >>> >>> at >>> org.apache.spark.network.netty.NettyBlockTransferService.fetchBlocks( >>> *NettyBlockTransferService.scala:143*) >>> >>> at org.apache.spark.network.BlockTransferService.fetchBlockSync( >>> *BlockTransferService.scala:103*) >>> >>> at >>> org.apache.spark.storage.BlockManager.fetchRemoteManagedBuffer( >>> *BlockManager.scala:1010*) >>> >>> at >>> org.apache.spark.storage.BlockManager.$anonfun$getRemoteBlock$8( >>> *BlockManager.scala:954*) >>> >>> at scala.Option.orElse(*Option.scala:447*) >>> >>> at org.apache.spark.storage.BlockManager.getRemoteBlock( >>> *BlockManager.scala:954*) >>> >>> at org.apache.spark.storage.BlockManager.getRemoteBytes( >>> *BlockManager.scala:1092*) >>> >>> at >>> org.apache.spark.scheduler.TaskResultGetter$$anon$3.$anonfun$run$1( >>> *TaskResultGetter.scala:88*) >>> >>> at >>> scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) >>> >>> at org.apache.spark.util.Utils$.logUncaughtExceptions( >>> *Utils.scala:1932*) >>> >>> at org.apache.spark.scheduler.TaskResultGetter$$anon$3.run( >>> *TaskResultGetter.scala:63*) >>> >>> at java.util.concurrent.ThreadPoolExecutor.runWorker( >>> *ThreadPoolExecutor.java:1149*) >>> >>> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >>> *ThreadPoolExecutor.java:624*) >>> >>> at java.lang.Thread.run(*Thread.java:748*) >>> >>> Caused by: *io.netty.channel.AbstractChannel$AnnotatedSocketException*: >>> Permission denied: no further information: >>> my-domain.com/192.168.166.8:63534 >>> >>> Caused by: *java.net.SocketException*: Permission denied: no further >>> information >>> >>> at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*) >>> >>> at sun.nio.ch.SocketChannelImpl.finishConnect( >>> *SocketChannelImpl.java:715*) >>> >>> at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect( >>> *NioSocketChannel.java:330*) >>> >>> at >>> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect( >>> *AbstractNioChannel.java:334*) >>> >>> at io.netty.channel.nio.NioEventLoop.processSelectedKey( >>> *NioEventLoop.java:702*) >>> >>> at >>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized( >>> *NioEventLoop.java:650*) >>> >>> at io.netty.channel.nio.NioEventLoop.processSelectedKeys( >>> *NioEventLoop.java:576*) >>> >>> at io.netty.channel.nio.NioEventLoop.run(*NioEventLoop.java:493* >>> ) >>> >>> at io.netty.util.concurrent.SingleThreadEventExecutor$4.run( >>> *SingleThreadEventExecutor.java:989*) >>> >>> at io.netty.util.internal.ThreadExecutorMap$2.run( >>> *ThreadExecutorMap.java:74*) >>> >>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>> *FastThreadLocalRunnable.java:30*) >>> >>> at java.lang.Thread.run(*Thread.java:748*) >>> >>> 20/12/17 13:27:02 ERROR RetryingBlockFetcher: Exception while beginning >>> fetch of 1 outstanding blocks (after 1 retries) >>> >>> *java.io.IOException*: Failed to connect to >>> my-domain.com/192.168.166.8:63534 >>> >>> at >>> org.apache.spark.network.client.TransportClientFactory.createClient( >>> *TransportClientFactory.java:253*) >>> >>> at >>> org.apache.spark.network.client.TransportClientFactory.createClient( >>> *TransportClientFactory.java:195*) >>> >>> at >>> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart( >>> *NettyBlockTransferService.scala:122*) >>> >>> at >>> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding( >>> *RetryingBlockFetcher.java:141*) >>> >>> at >>> org.apache.spark.network.shuffle.RetryingBlockFetcher.lambda$initiateRetry$0( >>> *RetryingBlockFetcher.java:169*) >>> >>> at java.util.concurrent.Executors$RunnableAdapter.call( >>> *Executors.java:511*) >>> >>> at java.util.concurrent.FutureTask.run(*FutureTask.java:266*) >>> >>> at java.util.concurrent.ThreadPoolExecutor.runWorker( >>> *ThreadPoolExecutor.java:1149*) >>> >>> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >>> *ThreadPoolExecutor.java:624*) >>> >>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>> *FastThreadLocalRunnable.java:30*) >>> >>> at java.lang.Thread.run(*Thread.java:748*) >>> >>> Caused by: *io.netty.channel.AbstractChannel$AnnotatedSocketException*: >>> Permission denied: no further information: >>> my-domain.com/192.168.166.8:63534 >>> >>> Caused by: *java.net.SocketException*: Permission denied: no further >>> information >>> >>> at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*) >>> >>> at sun.nio.ch.SocketChannelImpl.finishConnect( >>> *SocketChannelImpl.java:715*) >>> >>> at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect( >>> *NioSocketChannel.java:330*) >>> >>> at >>> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect( >>> *AbstractNioChannel.java:334*) >>> >>> at io.netty.channel.nio.NioEventLoop.processSelectedKey( >>> *NioEventLoop.java:702*) >>> >>> at >>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized( >>> *NioEventLoop.java:650*) >>> >>> at io.netty.channel.nio.NioEventLoop.processSelectedKeys( >>> *NioEventLoop.java:576*) >>> >>> at io.netty.channel.nio.NioEventLoop.run(*NioEventLoop.java:493* >>> ) >>> >>> at io.netty.util.concurrent.SingleThreadEventExecutor$4.run( >>> *SingleThreadEventExecutor.java:989*) >>> >>> at io.netty.util.internal.ThreadExecutorMap$2.run( >>> *ThreadExecutorMap.java:74*) >>> >>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>> *FastThreadLocalRunnable.java:30*) >>> >>> at java.lang.Thread.run(*Thread.java:748*) >>> >>> 20/12/17 13:27:07 ERROR RetryingBlockFetcher: Exception while beginning >>> fetch of 1 outstanding blocks (after 2 retries) >>> >>> *java.io.IOException*: Failed to connect to >>> my-domain.com/192.168.166.8:63534 >>> >>> at >>> org.apache.spark.network.client.TransportClientFactory.createClient( >>> *TransportClientFactory.java:253*) >>> >>> at >>> org.apache.spark.network.client.TransportClientFactory.createClient( >>> *TransportClientFactory.java:195*) >>> >>> at >>> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart( >>> *NettyBlockTransferService.scala:122*) >>> >>> at >>> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding( >>> *RetryingBlockFetcher.java:141*) >>> >>> at >>> org.apache.spark.network.shuffle.RetryingBlockFetcher.lambda$initiateRetry$0( >>> *RetryingBlockFetcher.java:169*) >>> >>> at java.util.concurrent.Executors$RunnableAdapter.call( >>> *Executors.java:511*) >>> >>> at java.util.concurrent.FutureTask.run(*FutureTask.java:266*) >>> >>> at java.util.concurrent.ThreadPoolExecutor.runWorker( >>> *ThreadPoolExecutor.java:1149*) >>> >>> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >>> *ThreadPoolExecutor.java:624*) >>> >>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>> *FastThreadLocalRunnable.java:30*) >>> >>> at java.lang.Thread.run(*Thread.java:748*) >>> >>> Caused by: *io.netty.channel.AbstractChannel$AnnotatedSocketException*: >>> Permission denied: no further information: >>> my-domain.com/192.168.166.8:63534 >>> >>> Caused by: *java.net.SocketException*: Permission denied: no further >>> information >>> >>> at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*) >>> >>> at sun.nio.ch.SocketChannelImpl.finishConnect( >>> *SocketChannelImpl.java:715*) >>> >>> at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect( >>> *NioSocketChannel.java:330*) >>> >>> at >>> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect( >>> *AbstractNioChannel.java:334*) >>> >>> at io.netty.channel.nio.NioEventLoop.processSelectedKey( >>> *NioEventLoop.java:702*) >>> >>> at >>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized( >>> *NioEventLoop.java:650*) >>> >>> at io.netty.channel.nio.NioEventLoop.processSelectedKeys( >>> *NioEventLoop.java:576*) >>> >>> at io.netty.channel.nio.NioEventLoop.run(*NioEventLoop.java:493* >>> ) >>> >>> at io.netty.util.concurrent.SingleThreadEventExecutor$4.run( >>> *SingleThreadEventExecutor.java:989*) >>> >>> at io.netty.util.internal.ThreadExecutorMap$2.run( >>> *ThreadExecutorMap.java:74*) >>> >>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>> *FastThreadLocalRunnable.java:30*) >>> >>> at java.lang.Thread.run(*Thread.java:748*) >>> >>> 20/12/17 13:27:12 ERROR RetryingBlockFetcher: Exception while beginning >>> fetch of 1 outstanding blocks (after 3 retries) >>> >>> *java.io.IOException*: Failed to connect to >>> my-domain.com/192.168.166.8:63534 >>> >>> at >>> org.apache.spark.network.client.TransportClientFactory.createClient( >>> *TransportClientFactory.java:253*) >>> >>> at >>> org.apache.spark.network.client.TransportClientFactory.createClient( >>> *TransportClientFactory.java:195*) >>> >>> at >>> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart( >>> *NettyBlockTransferService.scala:122*) >>> >>> at >>> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding( >>> *RetryingBlockFetcher.java:141*) >>> >>> at >>> org.apache.spark.network.shuffle.RetryingBlockFetcher.lambda$initiateRetry$0( >>> *RetryingBlockFetcher.java:169*) >>> >>> at java.util.concurrent.Executors$RunnableAdapter.call( >>> *Executors.java:511*) >>> >>> at java.util.concurrent.FutureTask.run(*FutureTask.java:266*) >>> >>> at java.util.concurrent.ThreadPoolExecutor.runWorker( >>> *ThreadPoolExecutor.java:1149*) >>> >>> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >>> *ThreadPoolExecutor.java:624*) >>> >>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>> *FastThreadLocalRunnable.java:30*) >>> >>> at java.lang.Thread.run(*Thread.java:748*) >>> >>> Caused by: *io.netty.channel.AbstractChannel$AnnotatedSocketException*: >>> Permission denied: no further information: >>> my-domain.com/192.168.166.8:63534 >>> >>> Caused by: *java.net.SocketException*: Permission denied: no further >>> information >>> >>> at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*) >>> >>> at sun.nio.ch.SocketChannelImpl.finishConnect( >>> *SocketChannelImpl.java:715*) >>> >>> at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect( >>> *NioSocketChannel.java:330*) >>> >>> at >>> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect( >>> *AbstractNioChannel.java:334*) >>> >>> at io.netty.channel.nio.NioEventLoop.processSelectedKey( >>> *NioEventLoop.java:702*) >>> >>> at >>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized( >>> *NioEventLoop.java:650*) >>> >>> at io.netty.channel.nio.NioEventLoop.processSelectedKeys( >>> *NioEventLoop.java:576*) >>> >>> at io.netty.channel.nio.NioEventLoop.run(*NioEventLoop.java:493* >>> ) >>> >>> at io.netty.util.concurrent.SingleThreadEventExecutor$4.run( >>> *SingleThreadEventExecutor.java:989*) >>> >>> at io.netty.util.internal.ThreadExecutorMap$2.run( >>> *ThreadExecutorMap.java:74*) >>> >>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>> *FastThreadLocalRunnable.java:30*) >>> >>> at java.lang.Thread.run(*Thread.java:748*) >>> >>> 20/12/17 13:27:12 ERROR TaskSetManager: Task 1 in stage 39.0 failed 1 >>> times; aborting job >>> >>> 20/12/17 13:27:12 ERROR FileFormatWriter: Aborting job >>> 0db8a16a-e1d3-4091-b725-0a8bdc17b0b1. >>> >>> *java.util.concurrent.ExecutionException*: >>> *org.apache.spark.SparkException*: Job aborted due to stage failure: >>> Task 1 in stage 39.0 failed 1 times, most recent failure: Lost task 1.0 in >>> stage 39.0 (TID 655, my-domain.com, executor driver): TaskResultLost >>> (result lost from block manager) >>> >>> Driver stacktrace: >>> >>> at java.util.concurrent.FutureTask.report(*FutureTask.java:122*) >>> >>> at java.util.concurrent.FutureTask.get(*FutureTask.java:206*) >>> >>> at >>> org.apache.spark.sql.execution.exchange.BroadcastExchangeExec.doExecuteBroadcast( >>> *BroadcastExchangeExec.scala:195*) >>> >>> at >>> org.apache.spark.sql.execution.InputAdapter.doExecuteBroadcast( >>> *WholeStageCodegenExec.scala:515*) >>> >>> at >>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeBroadcast$1( >>> *SparkPlan.scala:188*) >>> >>> at >>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>> *SparkPlan.scala:213*) >>> >>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>> *RDDOperationScope.scala:151*) >>> >>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>> *SparkPlan.scala:210*) >>> >>> at org.apache.spark.sql.execution.SparkPlan.executeBroadcast( >>> *SparkPlan.scala:184*) >>> >>> at >>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.prepareBroadcast( >>> *BroadcastHashJoinExec.scala:116*) >>> >>> at >>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.codegenInner( >>> *BroadcastHashJoinExec.scala:210*) >>> >>> at >>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.doConsume( >>> *BroadcastHashJoinExec.scala:100*) >>> >>> at org.apache.spark.sql.execution.CodegenSupport.consume( >>> *WholeStageCodegenExec.scala:194*) >>> >>> at org.apache.spark.sql.execution.CodegenSupport.consume$( >>> *WholeStageCodegenExec.scala:149*) >>> >>> at org.apache.spark.sql.execution.ProjectExec.consume( >>> *basicPhysicalOperators.scala:41*) >>> >>> at org.apache.spark.sql.execution.ProjectExec.doConsume( >>> *basicPhysicalOperators.scala:71*) >>> >>> at org.apache.spark.sql.execution.CodegenSupport.consume( >>> *WholeStageCodegenExec.scala:194*) >>> >>> at org.apache.spark.sql.execution.CodegenSupport.consume$( >>> *WholeStageCodegenExec.scala:149*) >>> >>> at org.apache.spark.sql.execution.FilterExec.consume( >>> *basicPhysicalOperators.scala:97*) >>> >>> at org.apache.spark.sql.execution.FilterExec.doConsume( >>> *basicPhysicalOperators.scala:222*) >>> >>> at org.apache.spark.sql.execution.CodegenSupport.consume( >>> *WholeStageCodegenExec.scala:194*) >>> >>> at org.apache.spark.sql.execution.CodegenSupport.consume$( >>> *WholeStageCodegenExec.scala:149*) >>> >>> at org.apache.spark.sql.execution.ColumnarToRowExec.consume( >>> *Columnar.scala:60*) >>> >>> at org.apache.spark.sql.execution.ColumnarToRowExec.doProduce( >>> *Columnar.scala:185*) >>> >>> at >>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >>> *WholeStageCodegenExec.scala:95*) >>> >>> at >>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>> *SparkPlan.scala:213*) >>> >>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>> *RDDOperationScope.scala:151*) >>> >>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>> *SparkPlan.scala:210*) >>> >>> at org.apache.spark.sql.execution.CodegenSupport.produce( >>> *WholeStageCodegenExec.scala:90*) >>> >>> at org.apache.spark.sql.execution.CodegenSupport.produce$( >>> *WholeStageCodegenExec.scala:90*) >>> >>> at org.apache.spark.sql.execution.ColumnarToRowExec.produce( >>> *Columnar.scala:60*) >>> >>> at org.apache.spark.sql.execution.FilterExec.doProduce( >>> *basicPhysicalOperators.scala:137*) >>> >>> at >>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >>> *WholeStageCodegenExec.scala:95*) >>> >>> at >>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>> *SparkPlan.scala:213*) >>> >>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>> *RDDOperationScope.scala:151*) >>> >>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>> *SparkPlan.scala:210*) >>> >>> at org.apache.spark.sql.execution.CodegenSupport.produce( >>> *WholeStageCodegenExec.scala:90*) >>> >>> at org.apache.spark.sql.execution.CodegenSupport.produce$( >>> *WholeStageCodegenExec.scala:90*) >>> >>> at org.apache.spark.sql.execution.FilterExec.produce( >>> *basicPhysicalOperators.scala:97*) >>> >>> at org.apache.spark.sql.execution.ProjectExec.doProduce( >>> *basicPhysicalOperators.scala:51*) >>> >>> at >>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >>> *WholeStageCodegenExec.scala:95*) >>> >>> at >>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>> *SparkPlan.scala:213*) >>> >>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>> *RDDOperationScope.scala:151*) >>> >>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>> *SparkPlan.scala:210*) >>> >>> at org.apache.spark.sql.execution.CodegenSupport.produce( >>> *WholeStageCodegenExec.scala:90*) >>> >>> at org.apache.spark.sql.execution.CodegenSupport.produce$( >>> *WholeStageCodegenExec.scala:90*) >>> >>> at org.apache.spark.sql.execution.ProjectExec.produce( >>> *basicPhysicalOperators.scala:41*) >>> >>> at >>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.doProduce( >>> *BroadcastHashJoinExec.scala:95*) >>> >>> at >>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >>> *WholeStageCodegenExec.scala:95*) >>> >>> at >>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>> *SparkPlan.scala:213*) >>> >>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>> *RDDOperationScope.scala:151*) >>> >>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>> *SparkPlan.scala:210*) >>> >>> at org.apache.spark.sql.execution.CodegenSupport.produce( >>> *WholeStageCodegenExec.scala:90*) >>> >>> at org.apache.spark.sql.execution.CodegenSupport.produce$( >>> *WholeStageCodegenExec.scala:90*) >>> >>> at >>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.produce( >>> *BroadcastHashJoinExec.scala:39*) >>> >>> at org.apache.spark.sql.execution.ProjectExec.doProduce( >>> *basicPhysicalOperators.scala:51*) >>> >>> at >>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >>> *WholeStageCodegenExec.scala:95*) >>> >>> at >>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>> *SparkPlan.scala:213*) >>> >>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>> *RDDOperationScope.scala:151*) >>> >>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>> *SparkPlan.scala:210*) >>> >>> at org.apache.spark.sql.execution.CodegenSupport.produce( >>> *WholeStageCodegenExec.scala:90*) >>> >>> at org.apache.spark.sql.execution.CodegenSupport.produce$( >>> *WholeStageCodegenExec.scala:90*) >>> >>> at org.apache.spark.sql.execution.ProjectExec.produce( >>> *basicPhysicalOperators.scala:41*) >>> >>> at >>> org.apache.spark.sql.execution.WholeStageCodegenExec.doCodeGen( >>> *WholeStageCodegenExec.scala:632*) >>> >>> at >>> org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute( >>> *WholeStageCodegenExec.scala:692*) >>> >>> at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1( >>> *SparkPlan.scala:175*) >>> >>> at >>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>> *SparkPlan.scala:213*) >>> >>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>> *RDDOperationScope.scala:151*) >>> >>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>> *SparkPlan.scala:210*) >>> >>> at org.apache.spark.sql.execution.SparkPlan.execute( >>> *SparkPlan.scala:171*) >>> >>> at >>> org.apache.spark.sql.execution.datasources.FileFormatWriter$.write( >>> *FileFormatWriter.scala:172*) >>> >>> at >>> org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand.run( >>> *InsertIntoHadoopFsRelationCommand.scala:178*) >>> >>> at >>> org.apache.spark.sql.execution.command.DataWritingCommandExec.sideEffectResult$lzycompute( >>> *commands.scala:108*) >>> >>> at >>> org.apache.spark.sql.execution.command.DataWritingCommandExec.sideEffectResult( >>> *commands.scala:106*) >>> >>> at >>> org.apache.spark.sql.execution.command.DataWritingCommandExec.doExecute( >>> *commands.scala:131*) >>> >>> at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1( >>> *SparkPlan.scala:175*) >>> >>> at >>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>> *SparkPlan.scala:213*) >>> >>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>> *RDDOperationScope.scala:151*) >>> >>> at org.apache.spark.sql.execu >>> >> -- *Patrick McCarthy * Senior Data Scientist, Machine Learning Engineering Dstillery 470 Park Ave South, 17th Floor, NYC 10016