That's not the error, that's just telling you the application failed.
You have to look at the YARN logs for application_1479877553404_0041
to see why it failed.

On Mon, Dec 5, 2016 at 10:44 AM, Gerard Casey <gerardhughca...@gmail.com> wrote:
> Thanks Marcelo,
>
> My understanding from a few pointers is that this may be due to insufficient 
> read permissions to the key tab or a corrupt key tab. I have checked the read 
> permissions and they are ok. I can see that it is initially configuring 
> correctly:
>
>        INFO security.UserGroupInformation: Login successful for user 
> user@login_node using keytab file /path/to/keytab
>
> I’ve added the full trace below.
>
> Gerry
>
> Full trace:
>
> Multiple versions of Spark are installed but SPARK_MAJOR_VERSION is not set
> Spark1 will be picked by default
> 16/12/05 18:23:27 WARN util.NativeCodeLoader: Unable to load native-hadoop 
> library for your platform... using builtin-java classes where applicable
> 16/12/05 18:23:27 INFO security.UserGroupInformation: Login successful for 
> user me@login_nodeusing keytab file /path/to/keytab
> 16/12/05 18:23:27 INFO yarn.Client: Attempting to login to the Kerberos using 
> principal: me@login_node and keytab: /path/to/keytab
> 16/12/05 18:23:28 INFO impl.TimelineClientImpl: Timeline service address: 
> http://login_node1.xcat.cluster:8188/ws/v1/timeline/
> 16/12/05 18:23:28 INFO client.RMProxy: Connecting to ResourceManager at 
> login_node1.xcat.cluster/
> 16/12/05 18:23:28 INFO client.AHSProxy: Connecting to Application History 
> server at login_node1.xcat.cluster/
> 16/12/05 18:23:28 WARN shortcircuit.DomainSocketFactory: The short-circuit 
> local reads feature cannot be used because libhadoop cannot be loaded.
> 16/12/05 18:23:28 INFO yarn.Client: Requesting a new application from cluster 
> with 32 NodeManagers
> 16/12/05 18:23:28 INFO yarn.Client: Verifying our application has not 
> requested more than the maximum memory capability of the cluster (15360 MB 
> per container)
> 16/12/05 18:23:28 INFO yarn.Client: Will allocate AM container, with 1408 MB 
> memory including 384 MB overhead
> 16/12/05 18:23:28 INFO yarn.Client: Setting up container launch context for 
> our AM
> 16/12/05 18:23:28 INFO yarn.Client: Setting up the launch environment for our 
> AM container
> 16/12/05 18:23:28 INFO yarn.Client: Using the spark assembly jar on HDFS 
> because you are using HDP, 
> defaultSparkAssembly:hdfs://login_node1.xcat.cluster:8020/hdp/apps/2.5.0.0-1245/spark/spark-hdp-assembly.jar
> 16/12/05 18:23:28 INFO yarn.Client: Credentials file set to:
> 16/12/05 18:23:28 INFO yarn.YarnSparkHadoopUtil: getting token for namenode: 
> hdfs://login_node1.xcat.cluster:8020/user/me/.sparkStaging/application_
> 16/12/05 18:23:28 INFO hdfs.DFSClient: Created HDFS_DELEGATION_TOKEN token 
> 1856 for me on
> 16/12/05 18:23:28 INFO yarn.Client: Renewal Interval set to 86400009
> 16/12/05 18:23:28 INFO yarn.Client: Preparing resources for our AM container
> 16/12/05 18:23:28 INFO yarn.YarnSparkHadoopUtil: getting token for namenode: 
> hdfs://login_node1.xcat.cluster:8020/user/me/.sparkStaging/application_
> 16/12/05 18:23:28 INFO hdfs.DFSClient: Created HDFS_DELEGATION_TOKEN token 
> 1857 for me on
> 16/12/05 18:23:29 INFO yarn.YarnSparkHadoopUtil: HBase class not found 
> java.lang.ClassNotFoundException: org.apache.hadoop.hbase.HBaseConfiguration
> 16/12/05 18:23:29 INFO yarn.Client: To enable the AM to login from keytab, 
> credentials are being copied over to the AM via the YARN Secure Distributed 
> Cache.
> 16/12/05 18:23:29 INFO yarn.Client: Uploading resource file:/path/to/keytab 
> -> 
> hdfs://login_node1.xcat.cluster:8020/user/me/.sparkStaging/application_1479877553404_0041/keytab
> 16/12/05 18:23:29 INFO yarn.Client: Using the spark assembly jar on HDFS 
> because you are using HDP, 
> defaultSparkAssembly:hdfs://login_node1.xcat.cluster:8020/hdp/apps/2.5.0.0-1245/spark/spark-hdp-assembly.jar
> 16/12/05 18:23:29 INFO yarn.Client: Source and destination file systems are 
> the same. Not copying 
> hdfs://login_node1.xcat.cluster:8020/hdp/apps/2.5.0.0-1245/spark/spark-hdp-assembly.jar
> 16/12/05 18:23:29 INFO yarn.Client: Uploading resource 
> file:/home/me/Aoife/spark-abm/target/scala-2.10/graphx_sp_2.10-1.0.jar -> 
> hdfs://login_node1.xcat.cluster:8020/user/me/.sparkStaging/application_1479877553404_0041/graphx_sp_2.10-1.0.jar
> 16/12/05 18:23:29 INFO yarn.Client: Uploading resource 
> file:/tmp/spark-2e566133-d50a-4904-920e-ab5cec07c644/__spark_conf__6538744395325375994.zip
>  -> 
> hdfs://login_node1.xcat.cluster:8020/user/me/.sparkStaging/application_1479877553404_0041/__spark_conf__6538744395325375994.zip
> 16/12/05 18:23:29 INFO spark.SecurityManager: Changing view acls to: me
> 16/12/05 18:23:29 INFO spark.SecurityManager: Changing modify acls to: me
> 16/12/05 18:23:29 INFO spark.SecurityManager: SecurityManager: authentication 
> disabled; ui acls disabled; users with view permissions: Set(me); users with 
> modify permissions: Set(me)
> 16/12/05 18:23:29 INFO yarn.Client: Submitting application 41 to 
> ResourceManager
> 16/12/05 18:23:30 INFO impl.YarnClientImpl: Submitted application 
> application_1479877553404_0041
> 16/12/05 18:23:31 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:23:31 INFO yarn.Client:
>          client token: Token { kind: YARN_CLIENT_TOKEN, service:  }
>          diagnostics: AM container is launched, waiting for AM container to 
> Register with RM
>          ApplicationMaster host: N/A
>          ApplicationMaster RPC port: -1
>          queue: default
>          start time: 1480962209903
>          final status: UNDEFINED
>          tracking URL: 
> http://login_node1.xcat.cluster:8088/proxy/application_1479877553404_0041/
>          user: me
> 16/12/05 18:23:32 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:23:33 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:23:34 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:23:35 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:23:36 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:23:37 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:23:38 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:23:39 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:23:40 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:23:40 INFO yarn.Client:
>          client token: Token { kind: YARN_CLIENT_TOKEN, service:  }
>          diagnostics: N/A
>          ApplicationMaster host:
>          ApplicationMaster RPC port: 0
>          queue: default
>          start time: 1480962209903
>          final status: UNDEFINED
>          tracking URL: 
> http://login_node1.xcat.cluster:8088/proxy/application_1479877553404_0041/
>          user: me
> 16/12/05 18:23:41 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:23:42 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:23:43 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:23:44 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:23:45 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:23:46 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:23:47 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:23:48 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:23:49 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:23:50 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:23:51 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:23:52 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:23:53 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:23:54 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:23:55 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:23:55 INFO yarn.Client:
>          client token: Token { kind: YARN_CLIENT_TOKEN, service:  }
>          diagnostics: AM container is launched, waiting for AM container to 
> Register with RM
>          ApplicationMaster host: N/A
>          ApplicationMaster RPC port: -1
>          queue: default
>          start time: 1480962209903
>          final status: UNDEFINED
>          tracking URL: 
> http://login_node1.xcat.cluster:8088/proxy/application_1479877553404_0041/
>          user: me
> 16/12/05 18:23:56 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:23:57 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:23:58 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:23:59 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:24:00 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:24:01 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:24:02 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:24:03 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:24:04 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: ACCEPTED)
> 16/12/05 18:24:05 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:24:05 INFO yarn.Client:
>          client token: Token { kind: YARN_CLIENT_TOKEN, service:  }
>          diagnostics: N/A
>          ApplicationMaster host:
>          ApplicationMaster RPC port: 0
>          queue: default
>          start time: 1480962209903
>          final status: UNDEFINED
>          tracking URL: 
> http://login_node1.xcat.cluster:8088/proxy/application_1479877553404_0041/
>          user: me
> 16/12/05 18:24:06 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:24:07 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:24:08 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:24:09 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:24:10 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:24:11 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:24:12 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:24:13 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:24:14 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:24:15 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:24:16 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:24:17 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: RUNNING)
> 16/12/05 18:24:18 INFO yarn.Client: Application report for 
> application_1479877553404_0041 (state: FINISHED)
> 16/12/05 18:24:18 INFO yarn.Client:
>          client token: Token { kind: YARN_CLIENT_TOKEN, service:  }
>          diagnostics: User class threw exception: 
> org.apache.hadoop.security.AccessControlException: Authentication required
>          ApplicationMaster host:
>          ApplicationMaster RPC port: 0
>          queue: default
>          start time: 1480962209903
>          final status: FAILED
>          tracking URL: 
> http://login_node1.xcat.cluster:8088/proxy/application_1479877553404_0041/
>          user: me
> Exception in thread "main" org.apache.spark.SparkException: Application 
> application_1479877553404_0041 finished with failed status
>         at org.apache.spark.deploy.yarn.Client.run(Client.scala:1122)
>         at org.apache.spark.deploy.yarn.Client$.main(Client.scala:1169)
>         at org.apache.spark.deploy.yarn.Client.main(Client.scala)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>         at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:497)
>         at 
> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
>         at 
> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
>         at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
>         at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
>         at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> 16/12/05 18:24:18 INFO util.ShutdownHookManager: Shutdown hook called
> 16/12/05 18:24:18 INFO util.ShutdownHookManager: Deleting directory 
> /tmp/spark-2e566133-d50a-4904-920e-ab5cec07c644
>
>
>
> On Mon, Dec 5, 2016 at 10:30 AM, Gerard Casey <gerardhughca...@gmail.com> 
> wrote:
>
>>
>>> On 5 Dec 2016, at 19:26, Marcelo Vanzin <van...@cloudera.com> wrote:
>>>
>>> There's generally an exception in these cases, and you haven't posted
>>> it, so it's hard to tell you what's wrong. The most probable cause,
>>> without the extra information the exception provides, is that you're
>>> using the wrong Hadoop configuration when submitting the job to YARN.
>>>
>>> On Mon, Dec 5, 2016 at 4:35 AM, Gerard Casey <gerardhughca...@gmail.com> 
>>> wrote:
>>>> Hello all,
>>>>
>>>> I am using Spark with Kerberos authentication.
>>>>
>>>> I can run my code using `spark-shell` fine and I can also use 
>>>> `spark-submit`
>>>> in local mode (e.g. —master local[16]). Both function as expected.
>>>>
>>>> local mode -
>>>>
>>>> spark-submit --class "graphx_sp" --master local[16] --driver-memory 20G
>>>> target/scala-2.10/graphx_sp_2.10-1.0.jar
>>>>
>>>> I am now progressing to run in cluster mode using YARN.
>>>>
>>>> cluster mode with YARN -
>>>>
>>>> spark-submit --class "graphx_sp" --master yarn --deploy-mode cluster
>>>> --executor-memory 13G --total-executor-cores 32
>>>> target/scala-2.10/graphx_sp_2.10-1.0.jar
>>>>
>>>> However, this returns:
>>>>
>>>> diagnostics: User class threw exception:
>>>> org.apache.hadoop.security.AccessControlException: Authentication required
>>>>
>>>> Before I run using spark-shell or on local mode in spark-submit I do the
>>>> following kerberos setup:
>>>>
>>>> kinit -k -t ~/keytab -r 7d `whoami`
>>>>
>>>> Clearly, this setup is not extending to the YARN setup. How do I fix the
>>>> Kerberos issue with YARN in cluster mode? Is this something which must be 
>>>> in
>>>> my /src/main/scala/graphx_sp.scala file?
>>>>
>>>> Many thanks
>>>>
>>>> Geroid
>>>
>>>
>>>
>>> --
>>> Marcelo
>>
>
>
>
> --
> Marcelo



-- 
Marcelo

---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Reply via email to