[
https://issues.apache.org/jira/browse/SPARK-25194?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16589283#comment-16589283
]
Yinan Li commented on SPARK-25194:
----------------------------------
The upcoming Spark 2.4 gets rid of the init-container and switch to running
{{spark-submit}} in client mode in the driver to download remote dependencies.
Given that 2.4 is coming soon, I would suggest waiting for and using it
instead.
> Kubernetes - Define cpu and memory limit to init container
> ----------------------------------------------------------
>
> Key: SPARK-25194
> URL: https://issues.apache.org/jira/browse/SPARK-25194
> Project: Spark
> Issue Type: New Feature
> Components: Kubernetes
> Affects Versions: 2.3.1
> Reporter: Daniel Majano
> Priority: Major
> Labels: features
>
> Hi,
>
> Recently I have started to work with spark under kubernetes. We have all our
> kubernetes clusters with resources quotes, so if you want to do a deploy yo
> need to define container cpu and memory limit.
>
> With driver and executors this is ok due to with spark submit props you can
> define this limits. But today for one of my projects, I need to load an
> external dependency. I have tried to define the dependency with --jars and
> the link with https so then, the init container will pop up and you don't
> have the possibility to define limits and the submitter failed due to he
> can't start the pod with driver + init container.
>
>
> Thanks.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]