Partha Pratim Ghosh created SPARK-17722:
-------------------------------------------

             Summary: YarnScheduler: Initial job has not accepted any resources
                 Key: SPARK-17722
                 URL: https://issues.apache.org/jira/browse/SPARK-17722
             Project: Spark
          Issue Type: Bug
            Reporter: Partha Pratim Ghosh


Connected spark in yarn mode from eclipse java. On trying to run task it is 
giving the following - 
YarnScheduler: Initial job has not accepted any resources; check your cluster 
UI to ensure that workers are registered and have sufficient resources. The 
request is going to Hadoop cluster scheduler and from there we can see the job 
in spark UI. But there it is saying that no task has been assigned for this.

Same code is running from spark-submit where we need to remove the following 
lines - 

System.setProperty("java.security.krb5.conf", "C:\\xxx\\krb5.conf");
                
                org.apache.hadoop.conf.Configuration conf = new     
                org.apache.hadoop.conf.Configuration();
                conf.set("hadoop.security.authentication", "kerberos");
                UserGroupInformation.setConfiguration(conf);

Following is the configuration - 

import org.apache.hadoop.security.UserGroupInformation;
import org.apache.spark.SparkConf;
import org.apache.spark.api.java.JavaRDD;
import org.apache.spark.api.java.JavaSparkContext;
import org.apache.spark.sql.DataFrame;
import org.apache.spark.sql.SQLContext;


public class TestConnectivity {

        /**
         * @param args
         */
        public static void main(String[] args) {
                System.setProperty("java.security.krb5.conf", 
"C:\\xxx\\krb5.conf");
                
                org.apache.hadoop.conf.Configuration conf = new     
                org.apache.hadoop.conf.Configuration();
                conf.set("hadoop.security.authentication", "kerberos");
                UserGroupInformation.setConfiguration(conf);

                 SparkConf config = new SparkConf().setAppName("Test Spark ");
                 config = config.setMaster("yarn-client");
                 config .set("spark.dynamicAllocation.enabled", "false");
                 config.set("spark.executor.memory", "2g");
                 config.set("spark.executor.instances", "1");
                 config.set("spark.executor.cores", "2");
                 //config.set("spark.driver.memory", "2g");
                 //config.set("spark.driver.cores", "1");
                 /*config.set("spark.executor.am.memory", "2g");
                 config.set("spark.executor.am.cores", "2");*/
                 config.set("spark.cores.max", "4");
                 config.set("yarn.nodemanager.resource.cpu-vcores","4");
                 config.set("spark.yarn.queue","root.root");
                 /*config.set("spark.deploy.defaultCores", "2");
                 config.set("spark.task.cpus", "2");*/
                 config.set("spark.yarn.jar", 
"file:/C:/xxx/spark-assembly_2.10-1.6.0-cdh5.7.1.jar");

                    JavaSparkContext sc = new JavaSparkContext(config);

                    SQLContext sqlcontext = new SQLContext(sc);

                            DataFrame df = sqlcontext.jsonFile(logFile);
                            JavaRDD<String> logData = 
sc.textFile("sparkexamples/Employee.json").cache();
                           DataFrame df = sqlcontext.jsonRDD(logData);
                           
                            df.show();
                            df.printSchema();
                    
                    //UserGroupInformation.setConfiguration(conf);
  
        }

}




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to