Github user vanzin commented on a diff in the pull request:
    --- Diff: 
    @@ -0,0 +1,85 @@
    + * Licensed to the Apache Software Foundation (ASF) under one or more
    + * contributor license agreements.  See the NOTICE file distributed with
    + * this work for additional information regarding copyright ownership.
    + * The ASF licenses this file to You under the Apache License, Version 2.0
    + * (the "License"); you may not use this file except in compliance with
    + * the License.  You may obtain a copy of the License at
    + *
    + *
    + *
    + * Unless required by applicable law or agreed to in writing, software
    + * distributed under the License is distributed on an "AS IS" BASIS,
    + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
    + * See the License for the specific language governing permissions and
    + * limitations under the License.
    + */
    +package org.apache.spark.deploy.k8s.submit.steps
    +import java.util.Properties
    +import io.fabric8.kubernetes.api.model._
    +import org.apache.spark.SparkConf
    +import org.apache.spark.deploy.k8s.Config._
    +import org.apache.spark.deploy.k8s.Constants._
    +import org.apache.spark.deploy.k8s.submit.KubernetesDriverSpec
    + * Create a config map with the driver configuration and attach it to the 
pod. This needs to
    + * come at the end of the driver configuration so that all modifications 
to the Spark config
    + * are reflected in the generated config map.
    + */
    +private[spark] class DriverConfigPropertiesStep(resourceNamePrefix: String)
    +    extends DriverConfigurationStep {
    +  override def configureDriver(spec: KubernetesDriverSpec): 
KubernetesDriverSpec = {
    +    val configMapName = s"$resourceNamePrefix-driver-conf-map"
    +    val configMap = buildConfigMap(configMapName, spec.driverSparkConf)
    --- End diff --
    > Just to double check, will this logic need to be re-used anywhere? I 
don't think the executor pods have to mount the same properties config map.
    Currently, I don't think so, since executors aren't started with 
spark-submit. That would be nice to do at some point, to simplify how executors 
are started, but that's a little beyond what this PR is doing.


To unsubscribe, e-mail:
For additional commands, e-mail:

Reply via email to