Vira Vitanska created DLAB-953:
----------------------------------
Summary: [Notebook]: Investigate how it is allocated memory for
spark
Key: DLAB-953
URL: https://issues.apache.org/jira/browse/DLAB-953
Project: Apache DLab
Issue Type: Task
Components: DLab Main
Reporter: Vira Vitanska
Assignee: Mykola
Fix For: v.2.2
Attachments: image-2019-07-26-17-49-24-994.png,
image-2019-07-26-17-54-18-544.png, image-2019-07-26-17-55-04-023.png
*Previously it was:*
1. the value of spark memory is equal 75% from total (if instance shape is up
to 8 GB)
2. the value of spark.executor.memory is equal formula value: total value
minus 3.5 GB (if instance shape is over than 8 GB).
*Currently it is:*
+Case 1:+ Rstudio [4 GB]
3.4*0.75*1024=3072MB (by theory)
But by practical:
!image-2019-07-26-17-55-04-023.png!
+Case2:+
Rstudio [122 GB]
(122-3.5)*1024= 121344
But by practical: (by theory)
!image-2019-07-26-17-54-18-544.png!
Please, investigate how memory should be allocated
--
This message was sent by Atlassian JIRA
(v7.6.14#76016)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]