Did you restart Zeppelin after you export OPTS in zeppelin-env.sh?



> On Dec 10, 2015, at 7:42 AM, Florian Leitner 
> <florian.leit...@seleritycorp.com> wrote:
> 
> Hi there, I need some help, please.
> 
> I'm using Zeppelin 0.5.5 (locally), and am trying to increase my Executors' 
> Memory sizes. They only get the default 1G according to the web panel. In the 
> conf/zeppelin-env.sh file, I've configured as follows:
> 
> export ZEPPELIN_JAVA_OPTS="-Dspark.executor.memory=10g 
> -Dspark.driver.memory=5g -Dspark.cores.max=8"
> 
> However, if, in a notebook, I then run sc.getConf.toDebugString I only see 
> the driver memory change to whatever I set here. Neither the spark cores or 
> the max. number of cores changes with this setting. Also, if I look at 
> running Executors on the web panel, they are always provisioned with 1g. So 
> downstream, as soon as I do DataFrame work even on just modestly sized 
> datasets, I immediately run into java.lang.OutOfMemoryError: Java heap space 
> errors...
> 
>  How/where do I then correctly define how much memory Spark Executors get 
> when running via a local Zeppelin instance?
> 
> Regards,
> Florian

Reply via email to