Mydomain is named by me while pasting the logs Also, there are multiple class files in my project, if I run any 1 or 2 at a time, then they run fine, sometimes they too give this error. But running all the classes at the same time always give this error.
Once this error come, I can't run any program and on restarting the system, program starts running fine. This error goes away on On Thu, 17 Dec 2020, 23:50 Patrick McCarthy, <pmccar...@dstillery.com> wrote: > my-domain.com/192.168.166.8:63534 probably isn't a valid address on your > network, is it? > > On Thu, Dec 17, 2020 at 3:03 AM Vikas Garg <sperry...@gmail.com> wrote: > >> Hi, >> >> Since last few days, I am getting error message while running my project. >> I have searched Google for the solution but didn't got any help. >> >> Can someone help me to figure out how I could mitigate this issue? >> >> >> 20/12/17 13:26:57 ERROR RetryingBlockFetcher: Exception while beginning >> fetch of 1 outstanding blocks >> >> *java.io.IOException*: Failed to connect to >> my-domain.com/192.168.166.8:63534 >> >> at >> org.apache.spark.network.client.TransportClientFactory.createClient( >> *TransportClientFactory.java:253*) >> >> at >> org.apache.spark.network.client.TransportClientFactory.createClient( >> *TransportClientFactory.java:195*) >> >> at >> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart( >> *NettyBlockTransferService.scala:122*) >> >> at >> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding( >> *RetryingBlockFetcher.java:141*) >> >> at org.apache.spark.network.shuffle.RetryingBlockFetcher.start( >> *RetryingBlockFetcher.java:121*) >> >> at >> org.apache.spark.network.netty.NettyBlockTransferService.fetchBlocks( >> *NettyBlockTransferService.scala:143*) >> >> at org.apache.spark.network.BlockTransferService.fetchBlockSync( >> *BlockTransferService.scala:103*) >> >> at org.apache.spark.storage.BlockManager.fetchRemoteManagedBuffer( >> *BlockManager.scala:1010*) >> >> at >> org.apache.spark.storage.BlockManager.$anonfun$getRemoteBlock$8( >> *BlockManager.scala:954*) >> >> at scala.Option.orElse(*Option.scala:447*) >> >> at org.apache.spark.storage.BlockManager.getRemoteBlock( >> *BlockManager.scala:954*) >> >> at org.apache.spark.storage.BlockManager.getRemoteBytes( >> *BlockManager.scala:1092*) >> >> at >> org.apache.spark.scheduler.TaskResultGetter$$anon$3.$anonfun$run$1( >> *TaskResultGetter.scala:88*) >> >> at >> scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) >> >> at org.apache.spark.util.Utils$.logUncaughtExceptions( >> *Utils.scala:1932*) >> >> at org.apache.spark.scheduler.TaskResultGetter$$anon$3.run( >> *TaskResultGetter.scala:63*) >> >> at java.util.concurrent.ThreadPoolExecutor.runWorker( >> *ThreadPoolExecutor.java:1149*) >> >> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >> *ThreadPoolExecutor.java:624*) >> >> at java.lang.Thread.run(*Thread.java:748*) >> >> Caused by: *io.netty.channel.AbstractChannel$AnnotatedSocketException*: >> Permission denied: no further information: >> my-domain.com/192.168.166.8:63534 >> >> Caused by: *java.net.SocketException*: Permission denied: no further >> information >> >> at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*) >> >> at sun.nio.ch.SocketChannelImpl.finishConnect( >> *SocketChannelImpl.java:715*) >> >> at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect( >> *NioSocketChannel.java:330*) >> >> at >> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect( >> *AbstractNioChannel.java:334*) >> >> at io.netty.channel.nio.NioEventLoop.processSelectedKey( >> *NioEventLoop.java:702*) >> >> at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized( >> *NioEventLoop.java:650*) >> >> at io.netty.channel.nio.NioEventLoop.processSelectedKeys( >> *NioEventLoop.java:576*) >> >> at io.netty.channel.nio.NioEventLoop.run(*NioEventLoop.java:493*) >> >> at io.netty.util.concurrent.SingleThreadEventExecutor$4.run( >> *SingleThreadEventExecutor.java:989*) >> >> at io.netty.util.internal.ThreadExecutorMap$2.run( >> *ThreadExecutorMap.java:74*) >> >> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >> *FastThreadLocalRunnable.java:30*) >> >> at java.lang.Thread.run(*Thread.java:748*) >> >> 20/12/17 13:27:02 ERROR RetryingBlockFetcher: Exception while beginning >> fetch of 1 outstanding blocks (after 1 retries) >> >> *java.io.IOException*: Failed to connect to >> my-domain.com/192.168.166.8:63534 >> >> at >> org.apache.spark.network.client.TransportClientFactory.createClient( >> *TransportClientFactory.java:253*) >> >> at >> org.apache.spark.network.client.TransportClientFactory.createClient( >> *TransportClientFactory.java:195*) >> >> at >> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart( >> *NettyBlockTransferService.scala:122*) >> >> at >> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding( >> *RetryingBlockFetcher.java:141*) >> >> at >> org.apache.spark.network.shuffle.RetryingBlockFetcher.lambda$initiateRetry$0( >> *RetryingBlockFetcher.java:169*) >> >> at java.util.concurrent.Executors$RunnableAdapter.call( >> *Executors.java:511*) >> >> at java.util.concurrent.FutureTask.run(*FutureTask.java:266*) >> >> at java.util.concurrent.ThreadPoolExecutor.runWorker( >> *ThreadPoolExecutor.java:1149*) >> >> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >> *ThreadPoolExecutor.java:624*) >> >> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >> *FastThreadLocalRunnable.java:30*) >> >> at java.lang.Thread.run(*Thread.java:748*) >> >> Caused by: *io.netty.channel.AbstractChannel$AnnotatedSocketException*: >> Permission denied: no further information: >> my-domain.com/192.168.166.8:63534 >> >> Caused by: *java.net.SocketException*: Permission denied: no further >> information >> >> at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*) >> >> at sun.nio.ch.SocketChannelImpl.finishConnect( >> *SocketChannelImpl.java:715*) >> >> at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect( >> *NioSocketChannel.java:330*) >> >> at >> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect( >> *AbstractNioChannel.java:334*) >> >> at io.netty.channel.nio.NioEventLoop.processSelectedKey( >> *NioEventLoop.java:702*) >> >> at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized( >> *NioEventLoop.java:650*) >> >> at io.netty.channel.nio.NioEventLoop.processSelectedKeys( >> *NioEventLoop.java:576*) >> >> at io.netty.channel.nio.NioEventLoop.run(*NioEventLoop.java:493*) >> >> at io.netty.util.concurrent.SingleThreadEventExecutor$4.run( >> *SingleThreadEventExecutor.java:989*) >> >> at io.netty.util.internal.ThreadExecutorMap$2.run( >> *ThreadExecutorMap.java:74*) >> >> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >> *FastThreadLocalRunnable.java:30*) >> >> at java.lang.Thread.run(*Thread.java:748*) >> >> 20/12/17 13:27:07 ERROR RetryingBlockFetcher: Exception while beginning >> fetch of 1 outstanding blocks (after 2 retries) >> >> *java.io.IOException*: Failed to connect to >> my-domain.com/192.168.166.8:63534 >> >> at >> org.apache.spark.network.client.TransportClientFactory.createClient( >> *TransportClientFactory.java:253*) >> >> at >> org.apache.spark.network.client.TransportClientFactory.createClient( >> *TransportClientFactory.java:195*) >> >> at >> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart( >> *NettyBlockTransferService.scala:122*) >> >> at >> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding( >> *RetryingBlockFetcher.java:141*) >> >> at >> org.apache.spark.network.shuffle.RetryingBlockFetcher.lambda$initiateRetry$0( >> *RetryingBlockFetcher.java:169*) >> >> at java.util.concurrent.Executors$RunnableAdapter.call( >> *Executors.java:511*) >> >> at java.util.concurrent.FutureTask.run(*FutureTask.java:266*) >> >> at java.util.concurrent.ThreadPoolExecutor.runWorker( >> *ThreadPoolExecutor.java:1149*) >> >> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >> *ThreadPoolExecutor.java:624*) >> >> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >> *FastThreadLocalRunnable.java:30*) >> >> at java.lang.Thread.run(*Thread.java:748*) >> >> Caused by: *io.netty.channel.AbstractChannel$AnnotatedSocketException*: >> Permission denied: no further information: >> my-domain.com/192.168.166.8:63534 >> >> Caused by: *java.net.SocketException*: Permission denied: no further >> information >> >> at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*) >> >> at sun.nio.ch.SocketChannelImpl.finishConnect( >> *SocketChannelImpl.java:715*) >> >> at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect( >> *NioSocketChannel.java:330*) >> >> at >> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect( >> *AbstractNioChannel.java:334*) >> >> at io.netty.channel.nio.NioEventLoop.processSelectedKey( >> *NioEventLoop.java:702*) >> >> at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized( >> *NioEventLoop.java:650*) >> >> at io.netty.channel.nio.NioEventLoop.processSelectedKeys( >> *NioEventLoop.java:576*) >> >> at io.netty.channel.nio.NioEventLoop.run(*NioEventLoop.java:493*) >> >> at io.netty.util.concurrent.SingleThreadEventExecutor$4.run( >> *SingleThreadEventExecutor.java:989*) >> >> at io.netty.util.internal.ThreadExecutorMap$2.run( >> *ThreadExecutorMap.java:74*) >> >> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >> *FastThreadLocalRunnable.java:30*) >> >> at java.lang.Thread.run(*Thread.java:748*) >> >> 20/12/17 13:27:12 ERROR RetryingBlockFetcher: Exception while beginning >> fetch of 1 outstanding blocks (after 3 retries) >> >> *java.io.IOException*: Failed to connect to >> my-domain.com/192.168.166.8:63534 >> >> at >> org.apache.spark.network.client.TransportClientFactory.createClient( >> *TransportClientFactory.java:253*) >> >> at >> org.apache.spark.network.client.TransportClientFactory.createClient( >> *TransportClientFactory.java:195*) >> >> at >> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart( >> *NettyBlockTransferService.scala:122*) >> >> at >> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding( >> *RetryingBlockFetcher.java:141*) >> >> at >> org.apache.spark.network.shuffle.RetryingBlockFetcher.lambda$initiateRetry$0( >> *RetryingBlockFetcher.java:169*) >> >> at java.util.concurrent.Executors$RunnableAdapter.call( >> *Executors.java:511*) >> >> at java.util.concurrent.FutureTask.run(*FutureTask.java:266*) >> >> at java.util.concurrent.ThreadPoolExecutor.runWorker( >> *ThreadPoolExecutor.java:1149*) >> >> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >> *ThreadPoolExecutor.java:624*) >> >> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >> *FastThreadLocalRunnable.java:30*) >> >> at java.lang.Thread.run(*Thread.java:748*) >> >> Caused by: *io.netty.channel.AbstractChannel$AnnotatedSocketException*: >> Permission denied: no further information: >> my-domain.com/192.168.166.8:63534 >> >> Caused by: *java.net.SocketException*: Permission denied: no further >> information >> >> at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*) >> >> at sun.nio.ch.SocketChannelImpl.finishConnect( >> *SocketChannelImpl.java:715*) >> >> at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect( >> *NioSocketChannel.java:330*) >> >> at >> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect( >> *AbstractNioChannel.java:334*) >> >> at io.netty.channel.nio.NioEventLoop.processSelectedKey( >> *NioEventLoop.java:702*) >> >> at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized( >> *NioEventLoop.java:650*) >> >> at io.netty.channel.nio.NioEventLoop.processSelectedKeys( >> *NioEventLoop.java:576*) >> >> at io.netty.channel.nio.NioEventLoop.run(*NioEventLoop.java:493*) >> >> at io.netty.util.concurrent.SingleThreadEventExecutor$4.run( >> *SingleThreadEventExecutor.java:989*) >> >> at io.netty.util.internal.ThreadExecutorMap$2.run( >> *ThreadExecutorMap.java:74*) >> >> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >> *FastThreadLocalRunnable.java:30*) >> >> at java.lang.Thread.run(*Thread.java:748*) >> >> 20/12/17 13:27:12 ERROR TaskSetManager: Task 1 in stage 39.0 failed 1 >> times; aborting job >> >> 20/12/17 13:27:12 ERROR FileFormatWriter: Aborting job >> 0db8a16a-e1d3-4091-b725-0a8bdc17b0b1. >> >> *java.util.concurrent.ExecutionException*: >> *org.apache.spark.SparkException*: Job aborted due to stage failure: >> Task 1 in stage 39.0 failed 1 times, most recent failure: Lost task 1.0 in >> stage 39.0 (TID 655, my-domain.com, executor driver): TaskResultLost >> (result lost from block manager) >> >> Driver stacktrace: >> >> at java.util.concurrent.FutureTask.report(*FutureTask.java:122*) >> >> at java.util.concurrent.FutureTask.get(*FutureTask.java:206*) >> >> at >> org.apache.spark.sql.execution.exchange.BroadcastExchangeExec.doExecuteBroadcast( >> *BroadcastExchangeExec.scala:195*) >> >> at org.apache.spark.sql.execution.InputAdapter.doExecuteBroadcast( >> *WholeStageCodegenExec.scala:515*) >> >> at >> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeBroadcast$1( >> *SparkPlan.scala:188*) >> >> at >> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >> *SparkPlan.scala:213*) >> >> at org.apache.spark.rdd.RDDOperationScope$.withScope( >> *RDDOperationScope.scala:151*) >> >> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >> *SparkPlan.scala:210*) >> >> at org.apache.spark.sql.execution.SparkPlan.executeBroadcast( >> *SparkPlan.scala:184*) >> >> at >> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.prepareBroadcast( >> *BroadcastHashJoinExec.scala:116*) >> >> at >> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.codegenInner( >> *BroadcastHashJoinExec.scala:210*) >> >> at >> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.doConsume( >> *BroadcastHashJoinExec.scala:100*) >> >> at org.apache.spark.sql.execution.CodegenSupport.consume( >> *WholeStageCodegenExec.scala:194*) >> >> at org.apache.spark.sql.execution.CodegenSupport.consume$( >> *WholeStageCodegenExec.scala:149*) >> >> at org.apache.spark.sql.execution.ProjectExec.consume( >> *basicPhysicalOperators.scala:41*) >> >> at org.apache.spark.sql.execution.ProjectExec.doConsume( >> *basicPhysicalOperators.scala:71*) >> >> at org.apache.spark.sql.execution.CodegenSupport.consume( >> *WholeStageCodegenExec.scala:194*) >> >> at org.apache.spark.sql.execution.CodegenSupport.consume$( >> *WholeStageCodegenExec.scala:149*) >> >> at org.apache.spark.sql.execution.FilterExec.consume( >> *basicPhysicalOperators.scala:97*) >> >> at org.apache.spark.sql.execution.FilterExec.doConsume( >> *basicPhysicalOperators.scala:222*) >> >> at org.apache.spark.sql.execution.CodegenSupport.consume( >> *WholeStageCodegenExec.scala:194*) >> >> at org.apache.spark.sql.execution.CodegenSupport.consume$( >> *WholeStageCodegenExec.scala:149*) >> >> at org.apache.spark.sql.execution.ColumnarToRowExec.consume( >> *Columnar.scala:60*) >> >> at org.apache.spark.sql.execution.ColumnarToRowExec.doProduce( >> *Columnar.scala:185*) >> >> at >> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >> *WholeStageCodegenExec.scala:95*) >> >> at >> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >> *SparkPlan.scala:213*) >> >> at org.apache.spark.rdd.RDDOperationScope$.withScope( >> *RDDOperationScope.scala:151*) >> >> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >> *SparkPlan.scala:210*) >> >> at org.apache.spark.sql.execution.CodegenSupport.produce( >> *WholeStageCodegenExec.scala:90*) >> >> at org.apache.spark.sql.execution.CodegenSupport.produce$( >> *WholeStageCodegenExec.scala:90*) >> >> at org.apache.spark.sql.execution.ColumnarToRowExec.produce( >> *Columnar.scala:60*) >> >> at org.apache.spark.sql.execution.FilterExec.doProduce( >> *basicPhysicalOperators.scala:137*) >> >> at >> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >> *WholeStageCodegenExec.scala:95*) >> >> at >> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >> *SparkPlan.scala:213*) >> >> at org.apache.spark.rdd.RDDOperationScope$.withScope( >> *RDDOperationScope.scala:151*) >> >> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >> *SparkPlan.scala:210*) >> >> at org.apache.spark.sql.execution.CodegenSupport.produce( >> *WholeStageCodegenExec.scala:90*) >> >> at org.apache.spark.sql.execution.CodegenSupport.produce$( >> *WholeStageCodegenExec.scala:90*) >> >> at org.apache.spark.sql.execution.FilterExec.produce( >> *basicPhysicalOperators.scala:97*) >> >> at org.apache.spark.sql.execution.ProjectExec.doProduce( >> *basicPhysicalOperators.scala:51*) >> >> at >> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >> *WholeStageCodegenExec.scala:95*) >> >> at >> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >> *SparkPlan.scala:213*) >> >> at org.apache.spark.rdd.RDDOperationScope$.withScope( >> *RDDOperationScope.scala:151*) >> >> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >> *SparkPlan.scala:210*) >> >> at org.apache.spark.sql.execution.CodegenSupport.produce( >> *WholeStageCodegenExec.scala:90*) >> >> at org.apache.spark.sql.execution.CodegenSupport.produce$( >> *WholeStageCodegenExec.scala:90*) >> >> at org.apache.spark.sql.execution.ProjectExec.produce( >> *basicPhysicalOperators.scala:41*) >> >> at >> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.doProduce( >> *BroadcastHashJoinExec.scala:95*) >> >> at >> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >> *WholeStageCodegenExec.scala:95*) >> >> at >> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >> *SparkPlan.scala:213*) >> >> at org.apache.spark.rdd.RDDOperationScope$.withScope( >> *RDDOperationScope.scala:151*) >> >> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >> *SparkPlan.scala:210*) >> >> at org.apache.spark.sql.execution.CodegenSupport.produce( >> *WholeStageCodegenExec.scala:90*) >> >> at org.apache.spark.sql.execution.CodegenSupport.produce$( >> *WholeStageCodegenExec.scala:90*) >> >> at >> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.produce( >> *BroadcastHashJoinExec.scala:39*) >> >> at org.apache.spark.sql.execution.ProjectExec.doProduce( >> *basicPhysicalOperators.scala:51*) >> >> at >> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >> *WholeStageCodegenExec.scala:95*) >> >> at >> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >> *SparkPlan.scala:213*) >> >> at org.apache.spark.rdd.RDDOperationScope$.withScope( >> *RDDOperationScope.scala:151*) >> >> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >> *SparkPlan.scala:210*) >> >> at org.apache.spark.sql.execution.CodegenSupport.produce( >> *WholeStageCodegenExec.scala:90*) >> >> at org.apache.spark.sql.execution.CodegenSupport.produce$( >> *WholeStageCodegenExec.scala:90*) >> >> at org.apache.spark.sql.execution.ProjectExec.produce( >> *basicPhysicalOperators.scala:41*) >> >> at org.apache.spark.sql.execution.WholeStageCodegenExec.doCodeGen( >> *WholeStageCodegenExec.scala:632*) >> >> at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute( >> *WholeStageCodegenExec.scala:692*) >> >> at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1( >> *SparkPlan.scala:175*) >> >> at >> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >> *SparkPlan.scala:213*) >> >> at org.apache.spark.rdd.RDDOperationScope$.withScope( >> *RDDOperationScope.scala:151*) >> >> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >> *SparkPlan.scala:210*) >> >> at org.apache.spark.sql.execution.SparkPlan.execute( >> *SparkPlan.scala:171*) >> >> at >> org.apache.spark.sql.execution.datasources.FileFormatWriter$.write( >> *FileFormatWriter.scala:172*) >> >> at >> org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand.run( >> *InsertIntoHadoopFsRelationCommand.scala:178*) >> >> at >> org.apache.spark.sql.execution.command.DataWritingCommandExec.sideEffectResult$lzycompute( >> *commands.scala:108*) >> >> at >> org.apache.spark.sql.execution.command.DataWritingCommandExec.sideEffectResult( >> *commands.scala:106*) >> >> at >> org.apache.spark.sql.execution.command.DataWritingCommandExec.doExecute( >> *commands.scala:131*) >> >> at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1( >> *SparkPlan.scala:175*) >> >> at >> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >> *SparkPlan.scala:213*) >> >> at org.apache.spark.rdd.RDDOperationScope$.withScope( >> *RDDOperationScope.scala:151*) >> >> at org.apache.spark.sql.execu >> >