Hi all,

I am seeing the following error in one of the supervisor nodes which is similar 
to the following question on stackoverflow - 
http://stackoverflow.com/questions/23008467/java-lang-outofmemoryerror-unable-to-create-new-native-thread-while-running-s

java.lang.OutOfMemoryError: unable to create new native thread at 
java.lang.Thread.start0(Native Method) at 
java.lang.Thread.start(Thread.java:691) at 
java.util.concurrent.ThreadPoolExecutor.addWorker(ThreadPoolExecutor.java:943) 
at 
java.util.concurrent.ThreadPoolExecutor.ensurePrestart(ThreadPoolExecutor.java:1555)
 at 
java.util.concurrent.ScheduledThreadPoolExecutor.delayedExecute(ScheduledThreadPoolExecutor.java:333)
 at 
java.util.concurrent.ScheduledThreadPoolExecutor.scheduleAtFixedRate(ScheduledThreadPoolExecutor.java:570)
 at 
java.util.concurrent.Executors$DelegatedScheduledExecutorService.scheduleAtFixedRate(Executors.java:695)
 at backtype.storm.task.ShellBolt.prepare(ShellBolt.java:128) at 
backtype.storm.daemon.executor$fn__4722$fn__4734.invoke(executor.clj:692) at 
backtype.storm.util$async_loop$fn__458.invoke(util.clj:461) at 
clojure.lang.AFn.run(AFn.java:24) at java.lang.Thread.run(Thread.java:722)

I checked the memory on my supervisor and it still has 40% of memory free. I am 
not sure if it is the same problem as the one discussed in the link above. 
Also, I am only seeing this on 1 supervisor and not on my other supervisor 
nodes. It was the same config we were having on our old cluster that was 
running 0.8.2. I also have multiple topologies running on the cluster but I 
doubt any application is causing memory leak. If it was shouldn't all memory be 
used up and not list anything as free?

Cluster summary -
1 Nimbus - 5 Supervisors (12 ports each)

drpc.childopts

-Xmx768m






nimbus.childopts

-Xmx1024m


supervisor.childopts

-Xmx256m


worker.childopts

-Xmx1024m


Thanks,
Ganesh




Reply via email to