No, I am not collect  the result to driver,I sample send the result to kafka.


BTW, the image address are:
https://cloud.githubusercontent.com/assets/5170878/7389463/ac03bf34-eea0-11e4-9e6b-1d2fba170c1c.png
and 
https://cloud.githubusercontent.com/assets/5170878/7389480/c629d236-eea0-11e4-983a-dc5aa97c2554.png



At 2015-04-29 18:48:33,zhangxiongfei <zhangxiongfei0...@163.com> wrote:



The mount of memory that the driver consumes depends on your program logic,did 
you try to collect the result of Spark job?




At 2015-04-29 18:42:04, "wyphao.2007" <wyphao.2...@163.com> wrote:

Hi, Dear developer, I am using Spark Streaming to read data from kafka, the 
program already run about 120 hours, but today the program failed because of 
driver's OOM as follow:


Container [pid=49133,containerID=container_1429773909253_0050_02_000001] is 
running beyond physical memory limits. Current usage: 2.5 GB of 2.5 GB physical 
memory used; 3.2 GB of 50 GB virtual memory used. Killing container.


I set --driver-memory to 2g, In my mind, driver is responsibility for job 
scheduler and job monitor(Please correct me If I'm wrong), Why it using so much 
memory?


So I using jmap to monitor other program(already run about 48 hours): 
sudo /home/q/java7/jdk1.7.0_45/bin/jmap -histo:live 31256, the result as follow:
the java.util.HashMap$Entry and java.lang.Long  object using about 600Mb memory!


and I also using jmap to monitor other program(already run about 1 hours),  the 
result as follow:
the java.util.HashMap$Entry and java.lang.Long object doesn't using so many 
memory, But I found, as time goes by, the java.util.HashMap$Entry and 
java.lang.Long object will occupied more and more memory,
It is driver's memory leak question? or other reason?
Thanks
Best Regards















Reply via email to