Github user rahulsinghaliitd commented on a diff in the pull request:

    https://github.com/apache/spark/pull/433#discussion_r11920472
  
    --- Diff: 
yarn/common/src/main/scala/org/apache/spark/deploy/yarn/ClientBase.scala ---
    @@ -354,63 +354,85 @@ trait ClientBase extends Logging {
       }
     }
     
    -object ClientBase {
    +object ClientBase extends Logging {
       val SPARK_JAR: String = "spark.jar"
       val APP_JAR: String = "app.jar"
       val LOG4J_PROP: String = "log4j.properties"
       val LOG4J_CONF_ENV_KEY: String = "SPARK_LOG4J_CONF"
       val LOCAL_SCHEME = "local"
     
    -  // Based on code from org.apache.hadoop.mapreduce.v2.util.MRApps
    -  def populateHadoopClasspath(conf: Configuration, env: HashMap[String, 
String]) {
    -    val classpathEntries = Option(conf.getStrings(
    -      YarnConfiguration.YARN_APPLICATION_CLASSPATH)).getOrElse(
    -        getDefaultYarnApplicationClasspath())
    -    for (c <- classpathEntries) {
    -      YarnSparkHadoopUtil.addToEnvironment(env, 
Environment.CLASSPATH.name, c.trim,
    +  def populateHadoopClasspath(conf: Configuration, env: HashMap[String, 
String]) = {
    +    val classPathElementsToAdd = getYarnAppClasspath(conf) ++ 
getMRAppClasspath(conf)
    +    for (c <- classPathElementsToAdd.flatten) {
    +      YarnSparkHadoopUtil.addToEnvironment(
    +        env,
    +        Environment.CLASSPATH.name,
    +        c.trim,
             File.pathSeparator)
         }
    +    classPathElementsToAdd
    +  }
     
    -    val mrClasspathEntries = Option(conf.getStrings(
    -      "mapreduce.application.classpath")).getOrElse(
    -        getDefaultMRApplicationClasspath())
    -    if (mrClasspathEntries != null) {
    -      for (c <- mrClasspathEntries) {
    -        YarnSparkHadoopUtil.addToEnvironment(env, 
Environment.CLASSPATH.name, c.trim,
    -          File.pathSeparator)
    -      }
    -    }
    +  private def getYarnAppClasspath(conf: Configuration): 
Option[Seq[String]] =
    +    Option(conf.getStrings(YarnConfiguration.YARN_APPLICATION_CLASSPATH)) 
match {
    +      case Some(s) => Some(s.toSeq)
    +      case None => getDefaultYarnApplicationClasspath
       }
     
    -  def getDefaultYarnApplicationClasspath(): Array[String] = {
    -    try {
    -      val field = 
classOf[MRJobConfig].getField("DEFAULT_YARN_APPLICATION_CLASSPATH")
    -      field.get(null).asInstanceOf[Array[String]]
    -    } catch {
    -      case err: NoSuchFieldError => null
    -      case err: NoSuchFieldException => null
    +  private def getMRAppClasspath(conf: Configuration): Option[Seq[String]] =
    +    Option(conf.getStrings("mapreduce.application.classpath")) match {
    +      case Some(s) => Some(s.toSeq)
    +      case None => getDefaultMRApplicationClasspath
    +    }
    +
    +  def getDefaultYarnApplicationClasspath: Option[Seq[String]] = {
    +    val triedDefault = Try[Seq[String]] {
    +      val field = 
classOf[YarnConfiguration].getField("DEFAULT_YARN_APPLICATION_CLASSPATH")
    --- End diff --
    
    Got it!
    
    I now see that the DEFAULT version is not present in hadoop 0.23. Then I 
wondered that if we truly want to be YARN API agnostic then shouldn't we get  
YARN_APPLICATION_CLASSPATH also via reflection. But I guess it is safe to 
assume that YARN_APPLICATION_CLASSPATH is here to stay.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

Reply via email to