Did you restart Zeppelin after you export OPTS in zeppelin-env.sh?
> On Dec 10, 2015, at 7:42 AM, Florian Leitner > <florian.leit...@seleritycorp.com> wrote: > > Hi there, I need some help, please. > > I'm using Zeppelin 0.5.5 (locally), and am trying to increase my Executors' > Memory sizes. They only get the default 1G according to the web panel. In the > conf/zeppelin-env.sh file, I've configured as follows: > > export ZEPPELIN_JAVA_OPTS="-Dspark.executor.memory=10g > -Dspark.driver.memory=5g -Dspark.cores.max=8" > > However, if, in a notebook, I then run sc.getConf.toDebugString I only see > the driver memory change to whatever I set here. Neither the spark cores or > the max. number of cores changes with this setting. Also, if I look at > running Executors on the web panel, they are always provisioned with 1g. So > downstream, as soon as I do DataFrame work even on just modestly sized > datasets, I immediately run into java.lang.OutOfMemoryError: Java heap space > errors... > > How/where do I then correctly define how much memory Spark Executors get > when running via a local Zeppelin instance? > > Regards, > Florian