Please find below the exact exception

Exception in thread "streaming-job-executor-3" java.lang.OutOfMemoryError:
Java heap space
        at java.util.Arrays.copyOf(Arrays.java:3332)
        at
java.lang.AbstractStringBuilder.ensureCapacityInternal(AbstractStringBuilder.java:124)
        at
java.lang.AbstractStringBuilder.append(AbstractStringBuilder.java:448)
        at java.lang.StringBuilder.append(StringBuilder.java:136)
        at scala.StringContext.standardInterpolator(StringContext.scala:126)
        at scala.StringContext.s(StringContext.scala:95)
        at sparkStreaming.TRReview.getTRReviews(TRReview.scala:307)
        at
sparkStreaming.KafkaListener$$anonfun$1$$anonfun$apply$1$$anonfun$3.apply(KafkaListener.scala:154)
        at
sparkStreaming.KafkaListener$$anonfun$1$$anonfun$apply$1$$anonfun$3.apply(KafkaListener.scala:138)
        at scala.util.Success$$anonfun$map$1.apply(Try.scala:237)
        at scala.util.Try$.apply(Try.scala:192)
        at scala.util.Success.map(Try.scala:237)

On Sun, Nov 8, 2020 at 1:35 PM Amit Sharma <resolve...@gmail.com> wrote:

> Hi , I am using 16 nodes spark cluster with below config
> 1. Executor memory  8 GB
> 2. 5 cores per executor
> 3. Driver memory 12 GB.
>
>
> We have streaming job. We do not see problem but sometimes we get
> exception executor-1 heap memory issue. I am not understanding if data size
> is same and this job receive a request and process it but suddenly it’s
> start giving out of memory error . It will throw exception for 1 executor
> then throw for other executor also and it stop processing the request.
>
> Thanks
> Amit
>

Reply via email to