[jira] [Commented] (SPARK-36872) Decommissioning executors get killed before transferring their data because of the hardcoded timeout of 60 secs

2022-01-27 Thread Abhishek Rao (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-36872?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17483537#comment-17483537
 ] 

Abhishek Rao commented on SPARK-36872:
--

Thanks. We'll have a look at this.

> Decommissioning executors get killed before transferring their data because 
> of the hardcoded timeout of 60 secs
> ---
>
> Key: SPARK-36872
> URL: https://issues.apache.org/jira/browse/SPARK-36872
> Project: Spark
>  Issue Type: Improvement
>  Components: Kubernetes
>Affects Versions: 3.1.1, 3.1.2, 3.2.0
>Reporter: Shekhar Gupta
>Priority: Trivial
>
> During the graceful decommissioning phase, executors need to transfer all of 
> their shuffle and cache data to the peer executors. However, they get killed 
> before transferring all the data because of the hardcoded timeout value of 60 
> secs in the decommissioning script. As a result of executors dying 
> prematurely, the spark tasks on other executors fail which causes application 
> failures, and it is hard to debug those failures. To fix the issue, we ended 
> up writing a custom script with a different timeout and rebuilt the spark 
> image but we would prefer an easier solution that does not require rebuilding 
> the image. 



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-36872) Decommissioning executors get killed before transferring their data because of the hardcoded timeout of 60 secs

2022-01-26 Thread Abhishek Rao (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-36872?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17482861#comment-17482861
 ] 

Abhishek Rao commented on SPARK-36872:
--

[~shkhrgpt], could you please share more details on which script you are 
referring to? We're facing similar issues and we're looking for options to fix 
this.

> Decommissioning executors get killed before transferring their data because 
> of the hardcoded timeout of 60 secs
> ---
>
> Key: SPARK-36872
> URL: https://issues.apache.org/jira/browse/SPARK-36872
> Project: Spark
>  Issue Type: Improvement
>  Components: Kubernetes
>Affects Versions: 3.1.1, 3.1.2, 3.2.0
>Reporter: Shekhar Gupta
>Priority: Trivial
>
> During the graceful decommissioning phase, executors need to transfer all of 
> their shuffle and cache data to the peer executors. However, they get killed 
> before transferring all the data because of the hardcoded timeout value of 60 
> secs in the decommissioning script. As a result of executors dying 
> prematurely, the spark tasks on other executors fail which causes application 
> failures, and it is hard to debug those failures. To fix the issue, we ended 
> up writing a custom script with a different timeout and rebuilt the spark 
> image but we would prefer an easier solution that does not require rebuilding 
> the image. 



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-35952) Inclusive terminology usage in Spark

2021-06-30 Thread Abhishek Rao (Jira)
Abhishek Rao created SPARK-35952:


 Summary: Inclusive terminology usage in Spark
 Key: SPARK-35952
 URL: https://issues.apache.org/jira/browse/SPARK-35952
 Project: Spark
  Issue Type: Improvement
  Components: Spark Core
Affects Versions: 3.0.3
Reporter: Abhishek Rao


Terms such as Blacklist/Whitelist and master/slave is used at different places 
in Spark Code. Do we have any plans to modify this to more inclusive 
terminology, for eg: Denylist/Allowlist and Leader/Follower?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-31825) Spark History Server UI does not come up when hosted on a custom path

2020-05-26 Thread Abhishek Rao (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-31825?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abhishek Rao updated SPARK-31825:
-
Attachment: Faulty Spark History UI.PNG

> Spark History Server UI does not come up when hosted on a custom path
> -
>
> Key: SPARK-31825
> URL: https://issues.apache.org/jira/browse/SPARK-31825
> Project: Spark
>  Issue Type: Bug
>  Components: Spark Core
>Affects Versions: 2.4.5
> Environment: Bring up Spark-History Server on any linux machine using 
> start-history-server.sh script.
>Reporter: Abhishek Rao
>Priority: Major
> Attachments: Faulty Spark History UI.PNG
>
>
> I tried to bringup spark-history server using the start-history-server.sh 
> script. The UI works perfectly fine when there is no path specified.
> i.e. http://:18080
> But If I bringup history server using custom path, I do not see the UI 
> working properly.
> Following is my configuration
> spark.history.fs.logDirectory=
> spark.ui.proxyBase=/test
> When I hit the url  http://:18080/test, I do not 
> see the History Server UI working properly. Attaching the screenshot of the 
> faulty UI.
> Wanted to know if I'm missing any configuration
>  
> !image-2020-05-26-15-26-21-616.png!
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-31825) Spark History Server UI does not come up when hosted on a custom path

2020-05-26 Thread Abhishek Rao (Jira)
Abhishek Rao created SPARK-31825:


 Summary: Spark History Server UI does not come up when hosted on a 
custom path
 Key: SPARK-31825
 URL: https://issues.apache.org/jira/browse/SPARK-31825
 Project: Spark
  Issue Type: Bug
  Components: Spark Core
Affects Versions: 2.4.5
 Environment: Bring up Spark-History Server on any linux machine using 
start-history-server.sh script.
Reporter: Abhishek Rao


I tried to bringup spark-history server using the start-history-server.sh 
script. The UI works perfectly fine when there is no path specified.

i.e. http://:18080

But If I bringup history server using custom path, I do not see the UI working 
properly.

Following is my configuration

spark.history.fs.logDirectory=
spark.ui.proxyBase=/test

When I hit the url  http://:18080/test, I do not see 
the History Server UI working properly. Attaching the screenshot of the faulty 
UI.

Wanted to know if I'm missing any configuration

 

!image-2020-05-26-15-26-21-616.png!

 

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-30619) org.slf4j.Logger and org.apache.commons.collections classes not built as part of hadoop-provided profile

2020-02-18 Thread Abhishek Rao (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-30619?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17039728#comment-17039728
 ] 

Abhishek Rao commented on SPARK-30619:
--

[~hyukjin.kwon] Were you able to reproduce this using the above procedure?

> org.slf4j.Logger and org.apache.commons.collections classes not built as part 
> of hadoop-provided profile
> 
>
> Key: SPARK-30619
> URL: https://issues.apache.org/jira/browse/SPARK-30619
> Project: Spark
>  Issue Type: Bug
>  Components: Build
>Affects Versions: 2.4.2, 2.4.4
> Environment: Spark on kubernetes
>Reporter: Abhishek Rao
>Priority: Major
>
> We're using spark-2.4.4-bin-without-hadoop.tgz and executing Java Word count 
> (org.apache.spark.examples.JavaWordCount) example on local files.
> But we're seeing that it is expecting org.slf4j.Logger and 
> org.apache.commons.collections classes to be available for executing this.
> We expected the binary to work as it is for local files. Is there anything 
> which we're missing?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-30619) org.slf4j.Logger and org.apache.commons.collections classes not built as part of hadoop-provided profile

2020-02-09 Thread Abhishek Rao (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-30619?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033374#comment-17033374
 ] 

Abhishek Rao commented on SPARK-30619:
--

Hi [~hyukjin.kwon]

I just built container using spark-2.4.4-bin-without-hadoop.tgz. Here is the 
spark-submit command that I used.

./spark-submit --deploy-mode cluster --class 
org.apache.spark.examples.JavaWordCount --master k8s://https:// 
--name spark-test --conf spark.kubernetes.container.image= --conf 
spark.kubernetes.authenticate.driver.serviceAccountName=spark-sa --conf 
spark.kubernetes.namespace=spark 
local:///opt/spark/examples/jars/spark-examples_2.11-2.4.4.jar 
file:///opt/spark/RELEASE

 

> org.slf4j.Logger and org.apache.commons.collections classes not built as part 
> of hadoop-provided profile
> 
>
> Key: SPARK-30619
> URL: https://issues.apache.org/jira/browse/SPARK-30619
> Project: Spark
>  Issue Type: Bug
>  Components: Build
>Affects Versions: 2.4.2, 2.4.4
> Environment: Spark on kubernetes
>Reporter: Abhishek Rao
>Priority: Major
>
> We're using spark-2.4.4-bin-without-hadoop.tgz and executing Java Word count 
> (org.apache.spark.examples.JavaWordCount) example on local files.
> But we're seeing that it is expecting org.slf4j.Logger and 
> org.apache.commons.collections classes to be available for executing this.
> We expected the binary to work as it is for local files. Is there anything 
> which we're missing?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-30619) org.slf4j.Logger and org.apache.commons.collections classes not built as part of hadoop-provided profile

2020-02-06 Thread Abhishek Rao (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-30619?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17032106#comment-17032106
 ] 

Abhishek Rao commented on SPARK-30619:
--

Hi,

Any updates on this?

> org.slf4j.Logger and org.apache.commons.collections classes not built as part 
> of hadoop-provided profile
> 
>
> Key: SPARK-30619
> URL: https://issues.apache.org/jira/browse/SPARK-30619
> Project: Spark
>  Issue Type: Bug
>  Components: Build
>Affects Versions: 2.4.2, 2.4.4
> Environment: Spark on kubernetes
>Reporter: Abhishek Rao
>Priority: Major
>
> We're using spark-2.4.4-bin-without-hadoop.tgz and executing Java Word count 
> (org.apache.spark.examples.JavaWordCount) example on local files.
> But we're seeing that it is expecting org.slf4j.Logger and 
> org.apache.commons.collections classes to be available for executing this.
> We expected the binary to work as it is for local files. Is there anything 
> which we're missing?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-30619) org.slf4j.Logger and org.apache.commons.collections classes not built as part of hadoop-provided profile

2020-01-30 Thread Abhishek Rao (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-30619?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17026612#comment-17026612
 ] 

Abhishek Rao commented on SPARK-30619:
--

Here are the 2 exceptions

SLF4J:

Error: A JNI error has occurred, please check your installation and try again

Exception in thread "main" java.lang.NoClassDefFoundError: org/slf4j/Logger

    at java.lang.Class.getDeclaredMethods0(Native Method)

    at java.lang.Class.privateGetDeclaredMethods(Class.java:2701)

    at java.lang.Class.privateGetMethodRecursive(Class.java:3048)

    at java.lang.Class.getMethod0(Class.java:3018)

    at java.lang.Class.getMethod(Class.java:1784)

    at 
sun.launcher.LauncherHelper.validateMainClass(LauncherHelper.java:544)

    at sun.launcher.LauncherHelper.checkAndLoadMain(LauncherHelper.java:526)

Caused by: java.lang.ClassNotFoundException: org.slf4j.Logger

    at java.net.URLClassLoader.findClass(URLClassLoader.java:382)

    at java.lang.ClassLoader.loadClass(ClassLoader.java:424)

    at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:349)

    at java.lang.ClassLoader.loadClass(ClassLoader.java:357)

    ... 7 more

 

Commons.collections

Exception in thread "main" java.lang.NoClassDefFoundError: 
org/apache/commons/collections/map/ReferenceMap
 at 
org.apache.spark.broadcast.BroadcastManager.(BroadcastManager.scala:58)
 at org.apache.spark.SparkEnv$.create(SparkEnv.scala:302)
 at org.apache.spark.SparkEnv$.createDriverEnv(SparkEnv.scala:185)
 at org.apache.spark.SparkContext.createSparkEnv(SparkContext.scala:257)
 at org.apache.spark.SparkContext.(SparkContext.scala:424)
 at org.apache.spark.api.java.JavaSparkContext.(JavaSparkContext.scala:58)
 at org.apache.spark.examples.JavaWordCount.main(JavaWordCount.java:28)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
 at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:498)
 at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
 at 
org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:850)
 at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167)
 at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195)
 at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
 at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:925)
 at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:934)
 at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.ClassNotFoundException: 
org.apache.commons.collections.map.ReferenceMap
 at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
 at java.lang.ClassLoader.loadClass(ClassLoader.java:418)
 at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352)
 at java.lang.ClassLoader.loadClass(ClassLoader.java:351)
 ... 19 more

> org.slf4j.Logger and org.apache.commons.collections classes not built as part 
> of hadoop-provided profile
> 
>
> Key: SPARK-30619
> URL: https://issues.apache.org/jira/browse/SPARK-30619
> Project: Spark
>  Issue Type: Bug
>  Components: Build
>Affects Versions: 2.4.2, 2.4.4
> Environment: Spark on kubernetes
>Reporter: Abhishek Rao
>Priority: Major
>
> We're using spark-2.4.4-bin-without-hadoop.tgz and executing Java Word count 
> (org.apache.spark.examples.JavaWordCount) example on local files.
> But we're seeing that it is expecting org.slf4j.Logger and 
> org.apache.commons.collections classes to be available for executing this.
> We expected the binary to work as it is for local files. Is there anything 
> which we're missing?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-30619) org.slf4j.Logger and org.apache.commons.collections classes not built as part of hadoop-provided profile

2020-01-28 Thread Abhishek Rao (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-30619?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abhishek Rao updated SPARK-30619:
-
Issue Type: Bug  (was: Question)

> org.slf4j.Logger and org.apache.commons.collections classes not built as part 
> of hadoop-provided profile
> 
>
> Key: SPARK-30619
> URL: https://issues.apache.org/jira/browse/SPARK-30619
> Project: Spark
>  Issue Type: Bug
>  Components: Build
>Affects Versions: 2.4.2, 2.4.4
> Environment: Spark on kubernetes
>Reporter: Abhishek Rao
>Priority: Major
>
> We're using spark-2.4.4-bin-without-hadoop.tgz and executing Java Word count 
> (org.apache.spark.examples.JavaWordCount) example on local files.
> But we're seeing that it is expecting org.slf4j.Logger and 
> org.apache.commons.collections classes to be available for executing this.
> We expected the binary to work as it is for local files. Is there anything 
> which we're missing?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-30619) org.slf4j.Logger and org.apache.commons.collections classes not built as part of hadoop-provided profile

2020-01-23 Thread Abhishek Rao (Jira)
Abhishek Rao created SPARK-30619:


 Summary: org.slf4j.Logger and org.apache.commons.collections 
classes not built as part of hadoop-provided profile
 Key: SPARK-30619
 URL: https://issues.apache.org/jira/browse/SPARK-30619
 Project: Spark
  Issue Type: Question
  Components: Build
Affects Versions: 2.4.4, 2.4.2
 Environment: Spark on kubernetes
Reporter: Abhishek Rao


We're using spark-2.4.4-bin-without-hadoop.tgz and executing Java Word count 
(org.apache.spark.examples.JavaWordCount) example on local files.

But we're seeing that it is expecting org.slf4j.Logger and 
org.apache.commons.collections classes to be available for executing this.

We expected the binary to work as it is for local files. Is there anything 
which we're missing?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-25500) Specify configmap and secrets in Spark driver and executor pods in Kubernetes

2018-09-25 Thread Abhishek Rao (JIRA)


 [ 
https://issues.apache.org/jira/browse/SPARK-25500?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abhishek Rao resolved SPARK-25500.
--
Resolution: Won't Fix

As mentioned in ticket, we'll use the templates way to mount configmap. Hence 
closing this ticket

> Specify configmap and secrets in Spark driver and executor pods in Kubernetes
> -
>
> Key: SPARK-25500
> URL: https://issues.apache.org/jira/browse/SPARK-25500
> Project: Spark
>  Issue Type: Improvement
>  Components: Kubernetes
>Affects Versions: 2.3.1
>Reporter: Abhishek Rao
>Priority: Minor
>
> This uses SPARK-23529. Support for specifying configmap and secrets as 
> spark-configuration is requested.
> Using PR #22146, the above functionality can be achieved by passing template 
> file. However, for spark properties (like log4j.properties, fairscheduler.xml 
> and metrics.properties), we are proposing this approach as this is native to 
> other configuration options specifications in spark.
> The configmaps and secrets have to be pre-created before using this as spark 
> configuration.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Comment Edited] (SPARK-25500) Specify configmap and secrets in Spark driver and executor pods in Kubernetes

2018-09-24 Thread Abhishek Rao (JIRA)


[ 
https://issues.apache.org/jira/browse/SPARK-25500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16625661#comment-16625661
 ] 

Abhishek Rao edited comment on SPARK-25500 at 9/24/18 11:21 AM:


Thanks for the information on existing configuration for configuring secrets. 
We'll continue to use the same.

Even though templates work for all scenarios, it would be better for native 
spark feature functionality to have it on command-line especially for spark 
configuration files like fairscheduler.xml, log4j.properties and 
metrics.properties.

This can of course be re-used for other usecases going forward if required.

The above PR has changes for specifying configmap during spark-submit


was (Author: abhisrao):
Even thought templates work for all scenarios, it would be better for native 
spark feature functionality to have it on command-line especially for spark 
configuration files like fairscheduler.xml, log4j.properties and 
metrics.properties.

This can of course be re-used for other usecases going forward if required.

> Specify configmap and secrets in Spark driver and executor pods in Kubernetes
> -
>
> Key: SPARK-25500
> URL: https://issues.apache.org/jira/browse/SPARK-25500
> Project: Spark
>  Issue Type: Improvement
>  Components: Kubernetes
>Affects Versions: 2.3.1
>Reporter: Abhishek Rao
>Priority: Minor
>
> This uses SPARK-23529. Support for specifying configmap and secrets as 
> spark-configuration is requested.
> Using PR #22146, the above functionality can be achieved by passing template 
> file. However, for spark properties (like log4j.properties, fairscheduler.xml 
> and metrics.properties), we are proposing this approach as this is native to 
> other configuration options specifications in spark.
> The configmaps and secrets have to be pre-created before using this as spark 
> configuration.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-25500) Specify configmap and secrets in Spark driver and executor pods in Kubernetes

2018-09-24 Thread Abhishek Rao (JIRA)


[ 
https://issues.apache.org/jira/browse/SPARK-25500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16625661#comment-16625661
 ] 

Abhishek Rao commented on SPARK-25500:
--

Even thought templates work for all scenarios, it would be better for native 
spark feature functionality to have it on command-line especially for spark 
configuration files like fairscheduler.xml, log4j.properties and 
metrics.properties.

This can of course be re-used for other usecases going forward if required.

> Specify configmap and secrets in Spark driver and executor pods in Kubernetes
> -
>
> Key: SPARK-25500
> URL: https://issues.apache.org/jira/browse/SPARK-25500
> Project: Spark
>  Issue Type: Improvement
>  Components: Kubernetes
>Affects Versions: 2.3.1
>Reporter: Abhishek Rao
>Priority: Minor
>
> This uses SPARK-23529. Support for specifying configmap and secrets as 
> spark-configuration is requested.
> Using PR #22146, the above functionality can be achieved by passing template 
> file. However, for spark properties (like log4j.properties, fairscheduler.xml 
> and metrics.properties), we are proposing this approach as this is native to 
> other configuration options specifications in spark.
> The configmaps and secrets have to be pre-created before using this as spark 
> configuration.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-25500) Specify configmap and secrets in Spark driver and executor pods in Kubernetes

2018-09-21 Thread Abhishek Rao (JIRA)


[ 
https://issues.apache.org/jira/browse/SPARK-25500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623335#comment-16623335
 ] 

Abhishek Rao commented on SPARK-25500:
--

Solution proposed is as follows

Add the following parameters while doing spark-submit

*For Configmap*

spark.kubernetes.driver.volumes.configMap..mount.path=

spark.kubernetes.driver.volumes.configMap..options.name=

 

*For Secrets*

spark.kubernetes.driver.volumes.secret..mount.path=

spark.kubernetes.driver.volumes.secret..options.name=

 

Using these properties, mount the configmap and secrets to spark-driver and 
spark-executor pods.

Changes are required in the "KubernetesVolumeSpec" and 
"MountVolumesFeatureStep" . It will be on similar lines as -SPARK-23529.-

I'll plan to submit a PR for this shortly.

> Specify configmap and secrets in Spark driver and executor pods in Kubernetes
> -
>
> Key: SPARK-25500
> URL: https://issues.apache.org/jira/browse/SPARK-25500
> Project: Spark
>  Issue Type: Improvement
>  Components: Kubernetes
>Affects Versions: 2.3.1
>Reporter: Abhishek Rao
>Priority: Minor
>
> This uses SPARK-23529. Support for specifying configmap and secrets as 
> spark-configuration is requested.
> Using PR #22146, the above functionality can be achieved by passing template 
> file. However, for spark properties (like log4j.properties, fairscheduler.xml 
> and metrics.properties), we are proposing this approach as this is native to 
> other configuration options specifications in spark.
> The configmaps and secrets have to be pre-created before using this as spark 
> configuration.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-25500) Specify configmap and secrets in Spark driver and executor pods in Kubernetes

2018-09-21 Thread Abhishek Rao (JIRA)
Abhishek Rao created SPARK-25500:


 Summary: Specify configmap and secrets in Spark driver and 
executor pods in Kubernetes
 Key: SPARK-25500
 URL: https://issues.apache.org/jira/browse/SPARK-25500
 Project: Spark
  Issue Type: Improvement
  Components: Kubernetes
Affects Versions: 2.3.1
Reporter: Abhishek Rao


This uses SPARK-23529. Support for specifying configmap and secrets as 
spark-configuration is requested.

Using PR #22146, the above functionality can be achieved by passing template 
file. However, for spark properties (like log4j.properties, fairscheduler.xml 
and metrics.properties), we are proposing this approach as this is native to 
other configuration options specifications in spark.

The configmaps and secrets have to be pre-created before using this as spark 
configuration.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-24179) History Server for Kubernetes

2018-05-11 Thread Abhishek Rao (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-24179?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16472035#comment-16472035
 ] 

Abhishek Rao commented on SPARK-24179:
--

We have brought up Spark History Server on Kubernetes using following approach. 
1) Prepare History-server docker container.

    a. Inherit Dockerfile from Kubespark/spark-base.
    b. Set Environment variable SPARK_NO_DAEMONIZE so that history-server can 
start as a Daemon.
    c. As part of Docker entry point, we’re invoking start-history-server.sh

2) Bringup Kubernetes pod using the docker image built in step 1
3) Bringup Kubernetes service for the histoty-server pod with type ClusterIP or 
NodePort.
4) If we bringup service with NodePort, History server can be accessed using 
NodeIP:NodePort
5) If we bringup service with ClusterIP, then we need to create Kubernetes 
ingress which will forward the request to the service brought up in step 3. 
Once we have the ingress up, history server can be accessed using Edge Node IP 
with the “path” specified in ingress.
6) Only limitation with ingress is that redirection is happening only when we 
use “/” in path. If we use any other string, redirection is not happening

This is validated using spark binaries from apache-spark-on-k8s forked from 
spark 2.2 as well as apache spark 2.3. Attached are the screen shots for both 
versions of spark.

> History Server for Kubernetes
> -
>
> Key: SPARK-24179
> URL: https://issues.apache.org/jira/browse/SPARK-24179
> Project: Spark
>  Issue Type: New Feature
>  Components: Kubernetes
>Affects Versions: 2.3.0
>Reporter: Eric Charles
>Priority: Major
> Attachments: Spark2_2_History_Server.PNG, Spark2_3_History_Server.PNG
>
>
> The History server is missing when running on Kubernetes, with the side 
> effect we can not debug post-mortem or analyze after-the-fact.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-24179) History Server for Kubernetes

2018-05-11 Thread Abhishek Rao (JIRA)

 [ 
https://issues.apache.org/jira/browse/SPARK-24179?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abhishek Rao updated SPARK-24179:
-
Attachment: Spark2_3_History_Server.PNG

> History Server for Kubernetes
> -
>
> Key: SPARK-24179
> URL: https://issues.apache.org/jira/browse/SPARK-24179
> Project: Spark
>  Issue Type: New Feature
>  Components: Kubernetes
>Affects Versions: 2.3.0
>Reporter: Eric Charles
>Priority: Major
> Attachments: Spark2_2_History_Server.PNG, Spark2_3_History_Server.PNG
>
>
> The History server is missing when running on Kubernetes, with the side 
> effect we can not debug post-mortem or analyze after-the-fact.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-24179) History Server for Kubernetes

2018-05-11 Thread Abhishek Rao (JIRA)

 [ 
https://issues.apache.org/jira/browse/SPARK-24179?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abhishek Rao updated SPARK-24179:
-
Attachment: Spark2_2_History_Server.PNG

> History Server for Kubernetes
> -
>
> Key: SPARK-24179
> URL: https://issues.apache.org/jira/browse/SPARK-24179
> Project: Spark
>  Issue Type: New Feature
>  Components: Kubernetes
>Affects Versions: 2.3.0
>Reporter: Eric Charles
>Priority: Major
> Attachments: Spark2_2_History_Server.PNG, Spark2_3_History_Server.PNG
>
>
> The History server is missing when running on Kubernetes, with the side 
> effect we can not debug post-mortem or analyze after-the-fact.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org