Re: Zeppelin out of memory issue - (GC overhead limit exceeded)

2017-03-26 Thread Jianfeng (Jeff) Zhang

I verify it in master branch, it works for me. Set it in interpreter setting 
page as following.


[cid:8CB49F76-39F5-4A53-816B-9E47F7993050]


Best Regard,
Jeff Zhang


From: RUSHIKESH RAUT 
<rushikeshraut...@gmail.com<mailto:rushikeshraut...@gmail.com>>
Reply-To: "users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>" 
<users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>>
Date: Sunday, March 26, 2017 at 8:02 PM
To: "users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>" 
<users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>>
Subject: Re: Zeppelin out of memory issue - (GC overhead limit exceeded)

Thanks Jianfeng,

But i am still not able to solve the issue. I have set it to 4g but still no 
luck.Can you please explain it to me how can I set SPARK_DRIVER_MEMORY  
property.
Also as I have read that GC overhead limit exceeded error occurs when the heap 
memory is insufficient. So How can I increase the heap memory. Please correct 
me if I am wrong as I am still trying to learn these things.
Reagrds,
Rushikesh Raut

On Sun, Mar 26, 2017 at 4:25 PM, Jianfeng (Jeff) Zhang 
<jzh...@hortonworks.com<mailto:jzh...@hortonworks.com>> wrote:

This is a bug of zeppelin. spark.driver.memory won't take effect. As for now it 
isn't passed to spark through -conf parameter. See 
https://issues.apache.org/jira/browse/ZEPPELIN-1263
The workaround is to specify SPARK_DRIVER_MEMORY in interpreter setting page.



Best Regard,
Jeff Zhang


From: RUSHIKESH RAUT 
<rushikeshraut...@gmail.com<mailto:rushikeshraut...@gmail.com>>
Reply-To: "users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>" 
<users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>>
Date: Sunday, March 26, 2017 at 5:03 PM
To: "users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>" 
<users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>>
Subject: Re: Zeppelin out of memory issue - (GC overhead limit exceeded)

ZEPPELIN_INTP_JAVA_OPTS



Re: Zeppelin out of memory issue - (GC overhead limit exceeded)

2017-03-26 Thread RUSHIKESH RAUT
Thanks Jianfeng,

But i am still not able to solve the issue. I have set it to 4g but still
no luck.Can you please explain it to me how can I set SPARK_DRIVER_MEMORY
 property.
Also as I have read that GC overhead limit exceeded error occurs when the
heap memory is insufficient. So How can I increase the heap memory. Please
correct me if I am wrong as I am still trying to learn these things.
Reagrds,
Rushikesh Raut

On Sun, Mar 26, 2017 at 4:25 PM, Jianfeng (Jeff) Zhang <
jzh...@hortonworks.com> wrote:

>
> This is a bug of zeppelin. spark.driver.memory won’t take effect. As for
> now it isn’t passed to spark through —conf parameter. See
> https://issues.apache.org/jira/browse/ZEPPELIN-1263
> The workaround is to specify SPARK_DRIVER_MEMORY in interpreter setting
> page.
>
>
>
> Best Regard,
> Jeff Zhang
>
>
> From: RUSHIKESH RAUT <rushikeshraut...@gmail.com>
> Reply-To: "users@zeppelin.apache.org" <users@zeppelin.apache.org>
> Date: Sunday, March 26, 2017 at 5:03 PM
> To: "users@zeppelin.apache.org" <users@zeppelin.apache.org>
> Subject: Re: Zeppelin out of memory issue - (GC overhead limit exceeded)
>
> ZEPPELIN_INTP_JAVA_OPTS
>


Re: Zeppelin out of memory issue - (GC overhead limit exceeded)

2017-03-26 Thread Jianfeng (Jeff) Zhang

This is a bug of zeppelin. spark.driver.memory won't take effect. As for now it 
isn't passed to spark through -conf parameter. See 
https://issues.apache.org/jira/browse/ZEPPELIN-1263
The workaround is to specify SPARK_DRIVER_MEMORY in interpreter setting page.



Best Regard,
Jeff Zhang


From: RUSHIKESH RAUT 
<rushikeshraut...@gmail.com<mailto:rushikeshraut...@gmail.com>>
Reply-To: "users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>" 
<users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>>
Date: Sunday, March 26, 2017 at 5:03 PM
To: "users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>" 
<users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>>
Subject: Re: Zeppelin out of memory issue - (GC overhead limit exceeded)

ZEPPELIN_INTP_JAVA_OPTS


Re: Zeppelin out of memory issue - (GC overhead limit exceeded)

2017-03-26 Thread Eric Charles
You don't have to set spark.driver.memory with -X... but simply with 
memory size.


Look at http://spark.apache.org/docs/latest/configuration.html

spark.driver.memory 	1g 	Amount of memory to use for the driver process, 
i.e. where SparkContext is initialized. (e.g. 1g, 2g).
Note: In client mode, this config must not be set through the SparkConf 
directly in your application, because the driver JVM has already started 
at that point. Instead, please set this through the --driver-memory 
command line option or in your default properties file.





On 26/03/17 09:57, RUSHIKESH RAUT wrote:

What value should I set there?
Currently I have set it as

spark.driver.memory  -Xms4096m -Xmx4096m -XX:MaxPermSize=2048m

But still same error

On Mar 26, 2017 1:19 PM, "Eric Charles" <e...@apache.org
<mailto:e...@apache.org>> wrote:

You also have to check the memory you give to the spark driver
(spark.driver.memory property)

On 26/03/17 07:40, RUSHIKESH RAUT wrote:

Yes I know it inevitable if the data is large. I want to know
how do I
increase the interpreter memory to handle large data?

Thanks,
Rushikesh Raut

On Mar 26, 2017 8:56 AM, "Jianfeng (Jeff) Zhang"
<jzh...@hortonworks.com <mailto:jzh...@hortonworks.com>
<mailto:jzh...@hortonworks.com <mailto:jzh...@hortonworks.com>>>
wrote:


How large is your data ? This problem is inevitable if your
data is
too large, you can try to use spark data frame if that works
for you.





Best Regard,
Jeff Zhang


From: RUSHIKESH RAUT <rushikeshraut...@gmail.com
<mailto:rushikeshraut...@gmail.com>
<mailto:rushikeshraut...@gmail.com
<mailto:rushikeshraut...@gmail.com>>>
Reply-To: "users@zeppelin.apache.org
<mailto:users@zeppelin.apache.org>
<mailto:users@zeppelin.apache.org
<mailto:users@zeppelin.apache.org>>" <users@zeppelin.apache.org
<mailto:users@zeppelin.apache.org>
<mailto:users@zeppelin.apache.org
<mailto:users@zeppelin.apache.org>>>
Date: Saturday, March 25, 2017 at 5:06 PM
To: "users@zeppelin.apache.org
<mailto:users@zeppelin.apache.org>
<mailto:users@zeppelin.apache.org
<mailto:users@zeppelin.apache.org>>"
<users@zeppelin.apache.org
<mailto:users@zeppelin.apache.org>
    <mailto:users@zeppelin.apache.org
<mailto:users@zeppelin.apache.org>>>
Subject: Zeppelin out of memory issue - (GC overhead limit
exceeded)

Hi everyone,

I am trying to load some data from hive table into my
notebook and
then convert this dataframe into r dataframe using spark.r
interpreter. This works perfectly for small amount of data.
But if the data is increased then it gives me error

java.lang.OutOfMemoryError: GC overhead limit exceeded

I have tried increasing the ZEPPELIN_MEM and
ZEPPELIN_INTP_MEM in
the zeppelin-env.cmd file but i am still facing this issue.
I have
used the following configuration

set ZEPPELIN_MEM="-Xms4096m -Xmx4096m -XX:MaxPermSize=2048m"
set ZEPPELIN_INTP_MEM="-Xmx4096m -Xms4096m
-XX:MaxPermSize=2048m"

I am sure that this much size should be sufficient for my
data but
still i am getting this same error. Any guidance will be much
appreciated.

Thanks,
Rushikesh Raut



Re: Zeppelin out of memory issue - (GC overhead limit exceeded)

2017-03-25 Thread RUSHIKESH RAUT
Yes I know it inevitable if the data is large. I want to know how do I
increase the interpreter memory to handle large data?

Thanks,
Rushikesh Raut

On Mar 26, 2017 8:56 AM, "Jianfeng (Jeff) Zhang" <jzh...@hortonworks.com>
wrote:

>
> How large is your data ? This problem is inevitable if your data is too
> large, you can try to use spark data frame if that works for you.
>
>
>
>
>
> Best Regard,
> Jeff Zhang
>
>
> From: RUSHIKESH RAUT <rushikeshraut...@gmail.com>
> Reply-To: "users@zeppelin.apache.org" <users@zeppelin.apache.org>
> Date: Saturday, March 25, 2017 at 5:06 PM
> To: "users@zeppelin.apache.org" <users@zeppelin.apache.org>
> Subject: Zeppelin out of memory issue - (GC overhead limit exceeded)
>
> Hi everyone,
>
> I am trying to load some data from hive table into my notebook and then
> convert this dataframe into r dataframe using spark.r interpreter. This
> works perfectly for small amount of data.
> But if the data is increased then it gives me error
>
> java.lang.OutOfMemoryError: GC overhead limit exceeded
>
> I have tried increasing the ZEPPELIN_MEM and ZEPPELIN_INTP_MEM in
> the zeppelin-env.cmd file but i am still facing this issue. I have used the
> following configuration
>
> set ZEPPELIN_MEM="-Xms4096m -Xmx4096m -XX:MaxPermSize=2048m"
> set ZEPPELIN_INTP_MEM="-Xmx4096m -Xms4096m -XX:MaxPermSize=2048m"
>
> I am sure that this much size should be sufficient for my data but still i
> am getting this same error. Any guidance will be much appreciated.
>
> Thanks,
> Rushikesh Raut
>


Zeppelin out of memory issue - (GC overhead limit exceeded)

2017-03-25 Thread RUSHIKESH RAUT
Hi everyone,

I am trying to load some data from hive table into my notebook and then
convert this dataframe into r dataframe using spark.r interpreter. This
works perfectly for small amount of data.
But if the data is increased then it gives me error

java.lang.OutOfMemoryError: GC overhead limit exceeded

I have tried increasing the ZEPPELIN_MEM and ZEPPELIN_INTP_MEM in
the zeppelin-env.cmd file but i am still facing this issue. I have used the
following configuration

set ZEPPELIN_MEM="-Xms4096m -Xmx4096m -XX:MaxPermSize=2048m"
set ZEPPELIN_INTP_MEM="-Xmx4096m -Xms4096m -XX:MaxPermSize=2048m"

I am sure that this much size should be sufficient for my data but still i
am getting this same error. Any guidance will be much appreciated.

Thanks,
Rushikesh Raut