sc <- sparkR.session(master =
"spark://ip-172-31-6-116:7077",sparkConfig=list(spark.executor.memory="10g",
spark.app.name="Testing",spark.driver.memory="14g",spark.executor.extraJavaOption="-Xms2g
-Xmx5g -XX:-UseGCOverheadLimit",spark.driver.extraJavaOption="-Xms2g -Xmx5g
-XX:-UseGCOverheadLimit",spark.cores.max="2",spark.sql.autoBroadcastJoinThreshold="-1"))

On Mon, Oct 24, 2016 at 7:33 PM, Mich Talebzadeh <mich.talebza...@gmail.com>
wrote:

> OK so what is your full launch code now? I mean equivalent to spark-submit
>
>
>
> Dr Mich Talebzadeh
>
>
>
> LinkedIn * 
> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>
>
>
> http://talebzadehmich.wordpress.com
>
>
> *Disclaimer:* Use it at your own risk. Any and all responsibility for any
> loss, damage or destruction of data or any other property which may arise
> from relying on this email's technical content is explicitly disclaimed.
> The author will in no case be liable for any monetary damages arising from
> such loss, damage or destruction.
>
>
>
> On 24 October 2016 at 14:57, Sankar Mittapally <sankar.mittapally@
> creditvidya.com> wrote:
>
>> Hi Mich,
>>
>>  I am able to write the files to storage after adding extra parameter.
>>
>> FYI..
>>
>> This one I used.
>>
>> spark.sql.autoBroadcastJoinThreshold="-1"
>>
>>
>>
>> On Mon, Oct 24, 2016 at 7:22 PM, Mich Talebzadeh <
>> mich.talebza...@gmail.com> wrote:
>>
>>> Rather strange as you have plenty free memory there.
>>>
>>> Try reducing driver memory to 2GB and executer memory to 2GB and run it
>>> again
>>>
>>> ${SPARK_HOME}/bin/spark-submit \
>>>                --driver-memory 2G \
>>>                 --num-executors 2 \
>>>                 --executor-cores 1 \
>>>                 --executor-memory 2G \
>>>                 --master spark://IPAddress:7077 \
>>>
>>> HTH
>>>
>>>
>>>
>>> Dr Mich Talebzadeh
>>>
>>>
>>>
>>> LinkedIn * 
>>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>>>
>>>
>>>
>>> http://talebzadehmich.wordpress.com
>>>
>>>
>>> *Disclaimer:* Use it at your own risk. Any and all responsibility for
>>> any loss, damage or destruction of data or any other property which may
>>> arise from relying on this email's technical content is explicitly
>>> disclaimed. The author will in no case be liable for any monetary damages
>>> arising from such loss, damage or destruction.
>>>
>>>
>>>
>>> On 24 October 2016 at 13:15, Sankar Mittapally <
>>> sankar.mittapa...@creditvidya.com> wrote:
>>>
>>>> Hi Mich,
>>>>
>>>>  Yes, I am using standalone mode cluster, We have two executors with
>>>> 10G memory each.  We have two workers.
>>>>
>>>> FYI..
>>>>
>>>>
>>>>
>>>> On Mon, Oct 24, 2016 at 5:22 PM, Mich Talebzadeh <
>>>> mich.talebza...@gmail.com> wrote:
>>>>
>>>>> Sounds like you are running in standalone mode.
>>>>>
>>>>> Have you checked the UI on port 4040 (default) to see where memory is
>>>>> going. Why do you need executor memory of 10GB?
>>>>>
>>>>> How many executors are running and plus how many slaves started?
>>>>>
>>>>> In standalone mode executors run on workers (UI 8080)
>>>>>
>>>>>
>>>>> HTH
>>>>>
>>>>> Dr Mich Talebzadeh
>>>>>
>>>>>
>>>>>
>>>>> LinkedIn * 
>>>>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>>>>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>>>>>
>>>>>
>>>>>
>>>>> http://talebzadehmich.wordpress.com
>>>>>
>>>>>
>>>>> *Disclaimer:* Use it at your own risk. Any and all responsibility for
>>>>> any loss, damage or destruction of data or any other property which may
>>>>> arise from relying on this email's technical content is explicitly
>>>>> disclaimed. The author will in no case be liable for any monetary damages
>>>>> arising from such loss, damage or destruction.
>>>>>
>>>>>
>>>>>
>>>>> On 24 October 2016 at 12:19, sankarmittapally <
>>>>> sankar.mittapa...@creditvidya.com> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>>  I have a three node cluster with 30G of Memory. I am trying to
>>>>>> analyzing
>>>>>> the data of 200MB and running out of memory every time. This is the
>>>>>> command
>>>>>> I am using
>>>>>>
>>>>>> Driver Memory = 10G
>>>>>> Executor memory=10G
>>>>>>
>>>>>> sc <- sparkR.session(master =
>>>>>> "spark://ip-172-31-6-116:7077",sparkConfig=list(spark.execut
>>>>>> or.memory="10g",spark.app.name="Testing",spark.driver.memory
>>>>>> ="14g",spark.executor.extraJavaOption="-Xms2g
>>>>>> -Xmx5g -XX:MaxPermSize=1024M",spark.driver.extraJavaOption="-Xms2g
>>>>>> -Xmx5g
>>>>>> -XX:MaxPermSize=1024M",spark.cores.max="2"))
>>>>>>
>>>>>>
>>>>>> [D 16:43:51.437 NotebookApp] 200 GET
>>>>>> /api/contents?type=directory&_=1477289197671 (123.176.38.226) 7.96ms
>>>>>> Exception in thread "broadcast-exchange-0"
>>>>>> java.lang.OutOfMemoryError: Java
>>>>>> heap space
>>>>>>         at
>>>>>> org.apache.spark.sql.execution.joins.LongToUnsafeRowMap.appe
>>>>>> nd(HashedRelation.scala:539)
>>>>>>         at
>>>>>> org.apache.spark.sql.execution.joins.LongHashedRelation$.app
>>>>>> ly(HashedRelation.scala:803)
>>>>>>         at
>>>>>> org.apache.spark.sql.execution.joins.HashedRelation$.apply(H
>>>>>> ashedRelation.scala:105)
>>>>>>         at
>>>>>> org.apache.spark.sql.execution.joins.HashedRelationBroadcast
>>>>>> Mode.transform(HashedRelation.scala:816)
>>>>>>         at
>>>>>> org.apache.spark.sql.execution.joins.HashedRelationBroadcast
>>>>>> Mode.transform(HashedRelation.scala:812)
>>>>>>         at
>>>>>> org.apache.spark.sql.execution.exchange.BroadcastExchangeExe
>>>>>> c$$anonfun$relationFuture$1$$anonfun$apply$1.apply(Broadcast
>>>>>> ExchangeExec.
>>>>>> scala:90)
>>>>>>         at
>>>>>> org.apache.spark.sql.execution.exchange.BroadcastExchangeExe
>>>>>> c$$anonfun$relationFuture$1$$anonfun$apply$1.apply(Broadcast
>>>>>> ExchangeExec.
>>>>>> scala:72)
>>>>>>         at
>>>>>> org.apache.spark.sql.execution.SQLExecution$.withExecutionId
>>>>>> (SQLExecution.scala:94)
>>>>>>         at
>>>>>> org.apache.spark.sql.execution.exchange.BroadcastExchangeExe
>>>>>> c$$anonfun$relationFuture$1.apply(BroadcastExchangeExec.scala:72)
>>>>>>         at
>>>>>> org.apache.spark.sql.execution.exchange.BroadcastExchangeExe
>>>>>> c$$anonfun$relationFuture$1.apply(BroadcastExchangeExec.scala:72)
>>>>>>         at
>>>>>> scala.concurrent.impl.Future$PromiseCompletingRunnable.lifte
>>>>>> dTree1$1(Future.scala:24)
>>>>>>         at
>>>>>> scala.concurrent.impl.Future$PromiseCompletingRunnable.run(F
>>>>>> uture.scala:24)
>>>>>>         at
>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool
>>>>>> Executor.java:1142)
>>>>>>         at
>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo
>>>>>> lExecutor.java:617)
>>>>>>         at java.lang.Thread.run(Thread.java:745)
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> View this message in context: http://apache-spark-user-list.
>>>>>> 1001560.n3.nabble.com/JAVA-heap-space-issue-tp27950.html
>>>>>> Sent from the Apache Spark User List mailing list archive at
>>>>>> Nabble.com.
>>>>>>
>>>>>> ---------------------------------------------------------------------
>>>>>> To unsubscribe e-mail: user-unsubscr...@spark.apache.org
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to