On Fri, Nov 20, 2020, 8:25 AM Amit Sharma <resolve...@gmail.com> wrote:
> please help. > > > Thanks > Amit > > On Mon, Nov 9, 2020 at 4:18 PM Amit Sharma <resolve...@gmail.com> wrote: > >> Please find below the exact exception >> >> Exception in thread "streaming-job-executor-3" >> java.lang.OutOfMemoryError: Java heap space >> at java.util.Arrays.copyOf(Arrays.java:3332) >> at >> java.lang.AbstractStringBuilder.ensureCapacityInternal(AbstractStringBuilder.java:124) >> at >> java.lang.AbstractStringBuilder.append(AbstractStringBuilder.java:448) >> at java.lang.StringBuilder.append(StringBuilder.java:136) >> at >> scala.StringContext.standardInterpolator(StringContext.scala:126) >> at scala.StringContext.s(StringContext.scala:95) >> at sparkStreaming.TRReview.getTRReviews(TRReview.scala:307) >> at >> sparkStreaming.KafkaListener$$anonfun$1$$anonfun$apply$1$$anonfun$3.apply(KafkaListener.scala:154) >> at >> sparkStreaming.KafkaListener$$anonfun$1$$anonfun$apply$1$$anonfun$3.apply(KafkaListener.scala:138) >> at scala.util.Success$$anonfun$map$1.apply(Try.scala:237) >> at scala.util.Try$.apply(Try.scala:192) >> at scala.util.Success.map(Try.scala:237) >> >> On Sun, Nov 8, 2020 at 1:35 PM Amit Sharma <resolve...@gmail.com> wrote: >> >>> Hi , I am using 16 nodes spark cluster with below config >>> 1. Executor memory 8 GB >>> 2. 5 cores per executor >>> 3. Driver memory 12 GB. >>> >>> >>> We have streaming job. We do not see problem but sometimes we get >>> exception executor-1 heap memory issue. I am not understanding if data size >>> is same and this job receive a request and process it but suddenly it’s >>> start giving out of memory error . It will throw exception for 1 executor >>> then throw for other executor also and it stop processing the request. >>> >>> Thanks >>> Amit >>> >>