Possibly. In that case maybe you should step back from spark and see if
there are OS-level tools to understand what's going on, like looking for
evidence of the OOM killer -
https://docs.memset.com/other/linux-s-oom-process-killer

On Thu, Dec 17, 2020 at 1:45 PM Vikas Garg <sperry...@gmail.com> wrote:

> I am running code in a local machine that is single node machine.
>
> Getting into logs,  it looked like the host is killed.  This is happening
> very frequently an I am unable to find the reason of this.
>
> Could low memory be the reason?
>
> On Fri, 18 Dec 2020, 00:11 Patrick McCarthy, <pmccar...@dstillery.com>
> wrote:
>
>> 'Job aborted due to stage failure: Task 1 in stage 39.0 failed 1 times'
>>
>> You may want to change the number of failures to a higher number like 4.
>> A single failure on a task should be able to be tolerated, especially if
>> you're on a shared cluster where resources can be preempted.
>>
>>  It seems that a node dies or goes off the network, so perhaps you can
>> also debug the logs on the failing node to see why it disappears and
>> prevent the failures in the first place.
>>
>> On Thu, Dec 17, 2020 at 1:27 PM Vikas Garg <sperry...@gmail.com> wrote:
>>
>>> Mydomain is named by me while pasting the logs
>>>
>>> Also,  there are multiple class files in my project, if I run any 1 or 2
>>> at a time,  then they run fine,  sometimes they too give this error. But
>>> running all the classes at the same time always give this error.
>>>
>>> Once this error come, I can't run any program and on restarting the
>>> system, program starts running fine.
>>> This error goes away on
>>>
>>> On Thu, 17 Dec 2020, 23:50 Patrick McCarthy, <pmccar...@dstillery.com>
>>> wrote:
>>>
>>>> my-domain.com/192.168.166.8:63534 probably isn't a valid address on
>>>> your network, is it?
>>>>
>>>> On Thu, Dec 17, 2020 at 3:03 AM Vikas Garg <sperry...@gmail.com> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> Since last few days, I am getting error message while running my
>>>>> project. I have searched Google for the solution but didn't got any help.
>>>>>
>>>>> Can someone help me to figure out how I could mitigate this issue?
>>>>>
>>>>>
>>>>> 20/12/17 13:26:57 ERROR RetryingBlockFetcher: Exception while
>>>>> beginning fetch of 1 outstanding blocks
>>>>>
>>>>> *java.io.IOException*: Failed to connect to
>>>>> my-domain.com/192.168.166.8:63534
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.client.TransportClientFactory.createClient(
>>>>> *TransportClientFactory.java:253*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.client.TransportClientFactory.createClient(
>>>>> *TransportClientFactory.java:195*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart(
>>>>> *NettyBlockTransferService.scala:122*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding(
>>>>> *RetryingBlockFetcher.java:141*)
>>>>>
>>>>>         at org.apache.spark.network.shuffle.RetryingBlockFetcher.start(
>>>>> *RetryingBlockFetcher.java:121*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.netty.NettyBlockTransferService.fetchBlocks(
>>>>> *NettyBlockTransferService.scala:143*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.BlockTransferService.fetchBlockSync(
>>>>> *BlockTransferService.scala:103*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.storage.BlockManager.fetchRemoteManagedBuffer(
>>>>> *BlockManager.scala:1010*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.storage.BlockManager.$anonfun$getRemoteBlock$8(
>>>>> *BlockManager.scala:954*)
>>>>>
>>>>>         at scala.Option.orElse(*Option.scala:447*)
>>>>>
>>>>>         at org.apache.spark.storage.BlockManager.getRemoteBlock(
>>>>> *BlockManager.scala:954*)
>>>>>
>>>>>         at org.apache.spark.storage.BlockManager.getRemoteBytes(
>>>>> *BlockManager.scala:1092*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.scheduler.TaskResultGetter$$anon$3.$anonfun$run$1(
>>>>> *TaskResultGetter.scala:88*)
>>>>>
>>>>>         at
>>>>> scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
>>>>>
>>>>>         at org.apache.spark.util.Utils$.logUncaughtExceptions(
>>>>> *Utils.scala:1932*)
>>>>>
>>>>>         at org.apache.spark.scheduler.TaskResultGetter$$anon$3.run(
>>>>> *TaskResultGetter.scala:63*)
>>>>>
>>>>>         at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>> *ThreadPoolExecutor.java:1149*)
>>>>>
>>>>>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(
>>>>> *ThreadPoolExecutor.java:624*)
>>>>>
>>>>>         at java.lang.Thread.run(*Thread.java:748*)
>>>>>
>>>>> Caused by: *io.netty.channel.AbstractChannel$AnnotatedSocketException*:
>>>>> Permission denied: no further information:
>>>>> my-domain.com/192.168.166.8:63534
>>>>>
>>>>> Caused by: *java.net.SocketException*: Permission denied: no further
>>>>> information
>>>>>
>>>>>         at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*)
>>>>>
>>>>>         at sun.nio.ch.SocketChannelImpl.finishConnect(
>>>>> *SocketChannelImpl.java:715*)
>>>>>
>>>>>         at
>>>>> io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(
>>>>> *NioSocketChannel.java:330*)
>>>>>
>>>>>         at
>>>>> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(
>>>>> *AbstractNioChannel.java:334*)
>>>>>
>>>>>         at io.netty.channel.nio.NioEventLoop.processSelectedKey(
>>>>> *NioEventLoop.java:702*)
>>>>>
>>>>>         at
>>>>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(
>>>>> *NioEventLoop.java:650*)
>>>>>
>>>>>         at io.netty.channel.nio.NioEventLoop.processSelectedKeys(
>>>>> *NioEventLoop.java:576*)
>>>>>
>>>>>         at io.netty.channel.nio.NioEventLoop.run(
>>>>> *NioEventLoop.java:493*)
>>>>>
>>>>>         at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(
>>>>> *SingleThreadEventExecutor.java:989*)
>>>>>
>>>>>         at io.netty.util.internal.ThreadExecutorMap$2.run(
>>>>> *ThreadExecutorMap.java:74*)
>>>>>
>>>>>         at io.netty.util.concurrent.FastThreadLocalRunnable.run(
>>>>> *FastThreadLocalRunnable.java:30*)
>>>>>
>>>>>         at java.lang.Thread.run(*Thread.java:748*)
>>>>>
>>>>> 20/12/17 13:27:02 ERROR RetryingBlockFetcher: Exception while
>>>>> beginning fetch of 1 outstanding blocks (after 1 retries)
>>>>>
>>>>> *java.io.IOException*: Failed to connect to
>>>>> my-domain.com/192.168.166.8:63534
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.client.TransportClientFactory.createClient(
>>>>> *TransportClientFactory.java:253*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.client.TransportClientFactory.createClient(
>>>>> *TransportClientFactory.java:195*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart(
>>>>> *NettyBlockTransferService.scala:122*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding(
>>>>> *RetryingBlockFetcher.java:141*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.lambda$initiateRetry$0(
>>>>> *RetryingBlockFetcher.java:169*)
>>>>>
>>>>>         at java.util.concurrent.Executors$RunnableAdapter.call(
>>>>> *Executors.java:511*)
>>>>>
>>>>>         at java.util.concurrent.FutureTask.run(*FutureTask.java:266*)
>>>>>
>>>>>         at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>> *ThreadPoolExecutor.java:1149*)
>>>>>
>>>>>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(
>>>>> *ThreadPoolExecutor.java:624*)
>>>>>
>>>>>         at io.netty.util.concurrent.FastThreadLocalRunnable.run(
>>>>> *FastThreadLocalRunnable.java:30*)
>>>>>
>>>>>         at java.lang.Thread.run(*Thread.java:748*)
>>>>>
>>>>> Caused by: *io.netty.channel.AbstractChannel$AnnotatedSocketException*:
>>>>> Permission denied: no further information:
>>>>> my-domain.com/192.168.166.8:63534
>>>>>
>>>>> Caused by: *java.net.SocketException*: Permission denied: no further
>>>>> information
>>>>>
>>>>>         at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*)
>>>>>
>>>>>         at sun.nio.ch.SocketChannelImpl.finishConnect(
>>>>> *SocketChannelImpl.java:715*)
>>>>>
>>>>>         at
>>>>> io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(
>>>>> *NioSocketChannel.java:330*)
>>>>>
>>>>>         at
>>>>> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(
>>>>> *AbstractNioChannel.java:334*)
>>>>>
>>>>>         at io.netty.channel.nio.NioEventLoop.processSelectedKey(
>>>>> *NioEventLoop.java:702*)
>>>>>
>>>>>         at
>>>>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(
>>>>> *NioEventLoop.java:650*)
>>>>>
>>>>>         at io.netty.channel.nio.NioEventLoop.processSelectedKeys(
>>>>> *NioEventLoop.java:576*)
>>>>>
>>>>>         at io.netty.channel.nio.NioEventLoop.run(
>>>>> *NioEventLoop.java:493*)
>>>>>
>>>>>         at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(
>>>>> *SingleThreadEventExecutor.java:989*)
>>>>>
>>>>>         at io.netty.util.internal.ThreadExecutorMap$2.run(
>>>>> *ThreadExecutorMap.java:74*)
>>>>>
>>>>>         at io.netty.util.concurrent.FastThreadLocalRunnable.run(
>>>>> *FastThreadLocalRunnable.java:30*)
>>>>>
>>>>>         at java.lang.Thread.run(*Thread.java:748*)
>>>>>
>>>>> 20/12/17 13:27:07 ERROR RetryingBlockFetcher: Exception while
>>>>> beginning fetch of 1 outstanding blocks (after 2 retries)
>>>>>
>>>>> *java.io.IOException*: Failed to connect to
>>>>> my-domain.com/192.168.166.8:63534
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.client.TransportClientFactory.createClient(
>>>>> *TransportClientFactory.java:253*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.client.TransportClientFactory.createClient(
>>>>> *TransportClientFactory.java:195*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart(
>>>>> *NettyBlockTransferService.scala:122*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding(
>>>>> *RetryingBlockFetcher.java:141*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.lambda$initiateRetry$0(
>>>>> *RetryingBlockFetcher.java:169*)
>>>>>
>>>>>         at java.util.concurrent.Executors$RunnableAdapter.call(
>>>>> *Executors.java:511*)
>>>>>
>>>>>         at java.util.concurrent.FutureTask.run(*FutureTask.java:266*)
>>>>>
>>>>>         at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>> *ThreadPoolExecutor.java:1149*)
>>>>>
>>>>>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(
>>>>> *ThreadPoolExecutor.java:624*)
>>>>>
>>>>>         at io.netty.util.concurrent.FastThreadLocalRunnable.run(
>>>>> *FastThreadLocalRunnable.java:30*)
>>>>>
>>>>>         at java.lang.Thread.run(*Thread.java:748*)
>>>>>
>>>>> Caused by: *io.netty.channel.AbstractChannel$AnnotatedSocketException*:
>>>>> Permission denied: no further information:
>>>>> my-domain.com/192.168.166.8:63534
>>>>>
>>>>> Caused by: *java.net.SocketException*: Permission denied: no further
>>>>> information
>>>>>
>>>>>         at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*)
>>>>>
>>>>>         at sun.nio.ch.SocketChannelImpl.finishConnect(
>>>>> *SocketChannelImpl.java:715*)
>>>>>
>>>>>         at
>>>>> io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(
>>>>> *NioSocketChannel.java:330*)
>>>>>
>>>>>         at
>>>>> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(
>>>>> *AbstractNioChannel.java:334*)
>>>>>
>>>>>         at io.netty.channel.nio.NioEventLoop.processSelectedKey(
>>>>> *NioEventLoop.java:702*)
>>>>>
>>>>>         at
>>>>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(
>>>>> *NioEventLoop.java:650*)
>>>>>
>>>>>         at io.netty.channel.nio.NioEventLoop.processSelectedKeys(
>>>>> *NioEventLoop.java:576*)
>>>>>
>>>>>         at io.netty.channel.nio.NioEventLoop.run(
>>>>> *NioEventLoop.java:493*)
>>>>>
>>>>>         at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(
>>>>> *SingleThreadEventExecutor.java:989*)
>>>>>
>>>>>         at io.netty.util.internal.ThreadExecutorMap$2.run(
>>>>> *ThreadExecutorMap.java:74*)
>>>>>
>>>>>         at io.netty.util.concurrent.FastThreadLocalRunnable.run(
>>>>> *FastThreadLocalRunnable.java:30*)
>>>>>
>>>>>         at java.lang.Thread.run(*Thread.java:748*)
>>>>>
>>>>> 20/12/17 13:27:12 ERROR RetryingBlockFetcher: Exception while
>>>>> beginning fetch of 1 outstanding blocks (after 3 retries)
>>>>>
>>>>> *java.io.IOException*: Failed to connect to
>>>>> my-domain.com/192.168.166.8:63534
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.client.TransportClientFactory.createClient(
>>>>> *TransportClientFactory.java:253*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.client.TransportClientFactory.createClient(
>>>>> *TransportClientFactory.java:195*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart(
>>>>> *NettyBlockTransferService.scala:122*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding(
>>>>> *RetryingBlockFetcher.java:141*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.lambda$initiateRetry$0(
>>>>> *RetryingBlockFetcher.java:169*)
>>>>>
>>>>>         at java.util.concurrent.Executors$RunnableAdapter.call(
>>>>> *Executors.java:511*)
>>>>>
>>>>>         at java.util.concurrent.FutureTask.run(*FutureTask.java:266*)
>>>>>
>>>>>         at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>> *ThreadPoolExecutor.java:1149*)
>>>>>
>>>>>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(
>>>>> *ThreadPoolExecutor.java:624*)
>>>>>
>>>>>         at io.netty.util.concurrent.FastThreadLocalRunnable.run(
>>>>> *FastThreadLocalRunnable.java:30*)
>>>>>
>>>>>         at java.lang.Thread.run(*Thread.java:748*)
>>>>>
>>>>> Caused by: *io.netty.channel.AbstractChannel$AnnotatedSocketException*:
>>>>> Permission denied: no further information:
>>>>> my-domain.com/192.168.166.8:63534
>>>>>
>>>>> Caused by: *java.net.SocketException*: Permission denied: no further
>>>>> information
>>>>>
>>>>>         at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*)
>>>>>
>>>>>         at sun.nio.ch.SocketChannelImpl.finishConnect(
>>>>> *SocketChannelImpl.java:715*)
>>>>>
>>>>>         at
>>>>> io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(
>>>>> *NioSocketChannel.java:330*)
>>>>>
>>>>>         at
>>>>> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(
>>>>> *AbstractNioChannel.java:334*)
>>>>>
>>>>>         at io.netty.channel.nio.NioEventLoop.processSelectedKey(
>>>>> *NioEventLoop.java:702*)
>>>>>
>>>>>         at
>>>>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(
>>>>> *NioEventLoop.java:650*)
>>>>>
>>>>>         at io.netty.channel.nio.NioEventLoop.processSelectedKeys(
>>>>> *NioEventLoop.java:576*)
>>>>>
>>>>>         at io.netty.channel.nio.NioEventLoop.run(
>>>>> *NioEventLoop.java:493*)
>>>>>
>>>>>         at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(
>>>>> *SingleThreadEventExecutor.java:989*)
>>>>>
>>>>>         at io.netty.util.internal.ThreadExecutorMap$2.run(
>>>>> *ThreadExecutorMap.java:74*)
>>>>>
>>>>>         at io.netty.util.concurrent.FastThreadLocalRunnable.run(
>>>>> *FastThreadLocalRunnable.java:30*)
>>>>>
>>>>>         at java.lang.Thread.run(*Thread.java:748*)
>>>>>
>>>>> 20/12/17 13:27:12 ERROR TaskSetManager: Task 1 in stage 39.0 failed 1
>>>>> times; aborting job
>>>>>
>>>>> 20/12/17 13:27:12 ERROR FileFormatWriter: Aborting job
>>>>> 0db8a16a-e1d3-4091-b725-0a8bdc17b0b1.
>>>>>
>>>>> *java.util.concurrent.ExecutionException*:
>>>>> *org.apache.spark.SparkException*: Job aborted due to stage failure:
>>>>> Task 1 in stage 39.0 failed 1 times, most recent failure: Lost task 1.0 in
>>>>> stage 39.0 (TID 655, my-domain.com, executor driver): TaskResultLost
>>>>> (result lost from block manager)
>>>>>
>>>>> Driver stacktrace:
>>>>>
>>>>>         at java.util.concurrent.FutureTask.report(
>>>>> *FutureTask.java:122*)
>>>>>
>>>>>         at java.util.concurrent.FutureTask.get(*FutureTask.java:206*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.exchange.BroadcastExchangeExec.doExecuteBroadcast(
>>>>> *BroadcastExchangeExec.scala:195*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.InputAdapter.doExecuteBroadcast(
>>>>> *WholeStageCodegenExec.scala:515*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeBroadcast$1(
>>>>> *SparkPlan.scala:188*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(
>>>>> *SparkPlan.scala:213*)
>>>>>
>>>>>         at org.apache.spark.rdd.RDDOperationScope$.withScope(
>>>>> *RDDOperationScope.scala:151*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.SparkPlan.executeQuery(
>>>>> *SparkPlan.scala:210*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.SparkPlan.executeBroadcast(
>>>>> *SparkPlan.scala:184*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.prepareBroadcast(
>>>>> *BroadcastHashJoinExec.scala:116*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.codegenInner(
>>>>> *BroadcastHashJoinExec.scala:210*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.doConsume(
>>>>> *BroadcastHashJoinExec.scala:100*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.CodegenSupport.consume(
>>>>> *WholeStageCodegenExec.scala:194*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.CodegenSupport.consume$(
>>>>> *WholeStageCodegenExec.scala:149*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.ProjectExec.consume(
>>>>> *basicPhysicalOperators.scala:41*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.ProjectExec.doConsume(
>>>>> *basicPhysicalOperators.scala:71*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.CodegenSupport.consume(
>>>>> *WholeStageCodegenExec.scala:194*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.CodegenSupport.consume$(
>>>>> *WholeStageCodegenExec.scala:149*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.FilterExec.consume(
>>>>> *basicPhysicalOperators.scala:97*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.FilterExec.doConsume(
>>>>> *basicPhysicalOperators.scala:222*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.CodegenSupport.consume(
>>>>> *WholeStageCodegenExec.scala:194*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.CodegenSupport.consume$(
>>>>> *WholeStageCodegenExec.scala:149*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.ColumnarToRowExec.consume(
>>>>> *Columnar.scala:60*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.ColumnarToRowExec.doProduce(
>>>>> *Columnar.scala:185*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1(
>>>>> *WholeStageCodegenExec.scala:95*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(
>>>>> *SparkPlan.scala:213*)
>>>>>
>>>>>         at org.apache.spark.rdd.RDDOperationScope$.withScope(
>>>>> *RDDOperationScope.scala:151*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.SparkPlan.executeQuery(
>>>>> *SparkPlan.scala:210*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.CodegenSupport.produce(
>>>>> *WholeStageCodegenExec.scala:90*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.CodegenSupport.produce$(
>>>>> *WholeStageCodegenExec.scala:90*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.ColumnarToRowExec.produce(
>>>>> *Columnar.scala:60*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.FilterExec.doProduce(
>>>>> *basicPhysicalOperators.scala:137*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1(
>>>>> *WholeStageCodegenExec.scala:95*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(
>>>>> *SparkPlan.scala:213*)
>>>>>
>>>>>         at org.apache.spark.rdd.RDDOperationScope$.withScope(
>>>>> *RDDOperationScope.scala:151*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.SparkPlan.executeQuery(
>>>>> *SparkPlan.scala:210*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.CodegenSupport.produce(
>>>>> *WholeStageCodegenExec.scala:90*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.CodegenSupport.produce$(
>>>>> *WholeStageCodegenExec.scala:90*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.FilterExec.produce(
>>>>> *basicPhysicalOperators.scala:97*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.ProjectExec.doProduce(
>>>>> *basicPhysicalOperators.scala:51*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1(
>>>>> *WholeStageCodegenExec.scala:95*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(
>>>>> *SparkPlan.scala:213*)
>>>>>
>>>>>         at org.apache.spark.rdd.RDDOperationScope$.withScope(
>>>>> *RDDOperationScope.scala:151*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.SparkPlan.executeQuery(
>>>>> *SparkPlan.scala:210*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.CodegenSupport.produce(
>>>>> *WholeStageCodegenExec.scala:90*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.CodegenSupport.produce$(
>>>>> *WholeStageCodegenExec.scala:90*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.ProjectExec.produce(
>>>>> *basicPhysicalOperators.scala:41*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.doProduce(
>>>>> *BroadcastHashJoinExec.scala:95*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1(
>>>>> *WholeStageCodegenExec.scala:95*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(
>>>>> *SparkPlan.scala:213*)
>>>>>
>>>>>         at org.apache.spark.rdd.RDDOperationScope$.withScope(
>>>>> *RDDOperationScope.scala:151*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.SparkPlan.executeQuery(
>>>>> *SparkPlan.scala:210*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.CodegenSupport.produce(
>>>>> *WholeStageCodegenExec.scala:90*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.CodegenSupport.produce$(
>>>>> *WholeStageCodegenExec.scala:90*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.produce(
>>>>> *BroadcastHashJoinExec.scala:39*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.ProjectExec.doProduce(
>>>>> *basicPhysicalOperators.scala:51*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1(
>>>>> *WholeStageCodegenExec.scala:95*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(
>>>>> *SparkPlan.scala:213*)
>>>>>
>>>>>         at org.apache.spark.rdd.RDDOperationScope$.withScope(
>>>>> *RDDOperationScope.scala:151*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.SparkPlan.executeQuery(
>>>>> *SparkPlan.scala:210*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.CodegenSupport.produce(
>>>>> *WholeStageCodegenExec.scala:90*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.CodegenSupport.produce$(
>>>>> *WholeStageCodegenExec.scala:90*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.ProjectExec.produce(
>>>>> *basicPhysicalOperators.scala:41*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.WholeStageCodegenExec.doCodeGen(
>>>>> *WholeStageCodegenExec.scala:632*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(
>>>>> *WholeStageCodegenExec.scala:692*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(
>>>>> *SparkPlan.scala:175*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(
>>>>> *SparkPlan.scala:213*)
>>>>>
>>>>>         at org.apache.spark.rdd.RDDOperationScope$.withScope(
>>>>> *RDDOperationScope.scala:151*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.SparkPlan.executeQuery(
>>>>> *SparkPlan.scala:210*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.SparkPlan.execute(
>>>>> *SparkPlan.scala:171*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.datasources.FileFormatWriter$.write(
>>>>> *FileFormatWriter.scala:172*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand.run(
>>>>> *InsertIntoHadoopFsRelationCommand.scala:178*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.command.DataWritingCommandExec.sideEffectResult$lzycompute(
>>>>> *commands.scala:108*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.command.DataWritingCommandExec.sideEffectResult(
>>>>> *commands.scala:106*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.command.DataWritingCommandExec.doExecute(
>>>>> *commands.scala:131*)
>>>>>
>>>>>         at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(
>>>>> *SparkPlan.scala:175*)
>>>>>
>>>>>         at
>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(
>>>>> *SparkPlan.scala:213*)
>>>>>
>>>>>         at org.apache.spark.rdd.RDDOperationScope$.withScope(
>>>>> *RDDOperationScope.scala:151*)
>>>>>
>>>>>         at org.apache.spark.sql.execu
>>>>>
>>>>
>>
>> --
>>
>>
>> *Patrick McCarthy  *
>>
>> Senior Data Scientist, Machine Learning Engineering
>>
>> Dstillery
>>
>> 470 Park Ave South, 17th Floor, NYC 10016
>>
>

-- 


*Patrick McCarthy  *

Senior Data Scientist, Machine Learning Engineering

Dstillery

470 Park Ave South, 17th Floor, NYC 10016

Reply via email to