Github user marsishandsome commented on the pull request:

    https://github.com/apache/spark/pull/9168#issuecomment-149394322
  
    The Scenario is as follows:
    1. Kerberos is enabled.
    
    2. NameNode HA is enabled.
    
    3. In order to test Token expired, I change the configuration of the 
NameNode
    dfs.namenode.delegation.token.max-lifetime = 40min
    dfs.namenode.delegation.key.update-interval = 20min
    dfs.namenode.delegation.token.renew-interval = 20min
    
    4. The Spark Test Application will write a HDFS file every minute.
    
    5. Yarn Cluster Mode is used.
    
    6. --principal --keytab argument is used.
    
    
    After running 40 min, I got  the Error:
    15/10/16 16:09:19 ERROR ApplicationMaster: User class threw exception: 
org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in 
stage 12.0 failed 4 times, most recent failure: Lost task 0.3 in stage 12.0 
(TID 30, node153-81-74-jylt.qiyi.hadoop): 
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.token.SecretManager$InvalidToken):
 token (HDFS_DELEGATION_TOKEN token 324309 for test) is expired
            at org.apache.hadoop.ipc.Client.call(Client.java:1411)
            at org.apache.hadoop.ipc.Client.call(Client.java:1364)
            at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
            at com.sun.proxy.$Proxy14.create(Unknown Source)
            at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.create(ClientNamenodeProtocolTranslatorPB.java:287)
            at sun.reflect.GeneratedMethodAccessor45.invoke(Unknown Source)
            at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
            at java.lang.reflect.Method.invoke(Method.java:606)
            at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
            at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
            at com.sun.proxy.$Proxy15.create(Unknown Source)
            at 
org.apache.hadoop.hdfs.DFSOutputStream.newStreamForCreate(DFSOutputStream.java:1645)
            at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1627)
            at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1552)
            at 
org.apache.hadoop.hdfs.DistributedFileSystem$6.doCall(DistributedFileSystem.java:396)
            at 
org.apache.hadoop.hdfs.DistributedFileSystem$6.doCall(DistributedFileSystem.java:392)
            at 
org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
            at 
org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:392)
            at 
org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:336)
            at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:908)
            at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:801)
            at 
org.apache.hadoop.mapred.TextOutputFormat.getRecordWriter(TextOutputFormat.java:123)
            at 
org.apache.spark.SparkHadoopWriter.open(SparkHadoopWriter.scala:91)
            at 
org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopDataset$1$$anonfun$13.apply(PairRDDFunctions.scala:1104)
            at 
org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopDataset$1$$anonfun$13.apply(PairRDDFunctions.scala:1095)
            at 
org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
            at org.apache.spark.scheduler.Task.run(Task.scala:88)
            at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
            at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
            at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
            at java.lang.Thread.run(Thread.java:745)



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to