[ 
https://issues.apache.org/jira/browse/SPARK-19644?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shixiong Zhu updated SPARK-19644:
---------------------------------
    Description: 
I am using streaming on the production for some aggregation and fetching data 
from cassandra and saving data back to cassandra. 

I see a gradual increase in old generation heap capacity from 1161216 Bytes to 
1397760 Bytes over a period of six hours.

After 50 hours of processing instances of class 
scala.collection.immutable.$colon$colon incresed to 12,811,793 which is a huge 
number. 

I think this is a clear case of memory leak

Updated: The major issue is when creating an encoder object, it leaks several 
Scala internal objects due to a Scala memory leak issue.

  was:
I am using streaming on the production for some aggregation and fetching data 
from cassandra and saving data back to cassandra. 

I see a gradual increase in old generation heap capacity from 1161216 Bytes to 
1397760 Bytes over a period of six hours.

After 50 hours of processing instances of class 
scala.collection.immutable.$colon$colon incresed to 12,811,793 which is a huge 
number. 

I think this is a clear case of memory leak


> Memory leak in Spark Streaming (Encoder/Scala Reflection)
> ---------------------------------------------------------
>
>                 Key: SPARK-19644
>                 URL: https://issues.apache.org/jira/browse/SPARK-19644
>             Project: Spark
>          Issue Type: Bug
>          Components: DStreams, SQL, Structured Streaming
>    Affects Versions: 2.0.2
>         Environment: 3 AWS EC2 c3.xLarge
> Number of cores - 3
> Number of executors 3 
> Memory to each executor 2GB
>            Reporter: Deenbandhu Agarwal
>            Priority: Major
>              Labels: memory_leak, performance
>         Attachments: Dominator_tree.png, Path2GCRoot.png, heapdump.png
>
>
> I am using streaming on the production for some aggregation and fetching data 
> from cassandra and saving data back to cassandra. 
> I see a gradual increase in old generation heap capacity from 1161216 Bytes 
> to 1397760 Bytes over a period of six hours.
> After 50 hours of processing instances of class 
> scala.collection.immutable.$colon$colon incresed to 12,811,793 which is a 
> huge number. 
> I think this is a clear case of memory leak
> Updated: The major issue is when creating an encoder object, it leaks several 
> Scala internal objects due to a Scala memory leak issue.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to