[jira] [Commented] (SPARK-28634) Failed to start SparkSession with Keytab file
[ https://issues.apache.org/jira/browse/SPARK-28634?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16908450#comment-16908450 ] Marcelo Vanzin commented on SPARK-28634: I think it's still worth it to fix it, so that users with old configuration are not surprised by this. > Failed to start SparkSession with Keytab file > -- > > Key: SPARK-28634 > URL: https://issues.apache.org/jira/browse/SPARK-28634 > Project: Spark > Issue Type: Bug > Components: YARN >Affects Versions: 3.0.0 >Reporter: Yuming Wang >Priority: Major > > {noformat} > [user-etl@hermesdevour002-700165 spark-3.0.0-SNAPSHOT-bin-2.7.4]$ > bin/spark-sql --master yarn --conf > spark.yarn.keytab=/apache/spark-2.3.0-bin-2.7.3/conf/user-etl.keytab --conf > spark.yarn.principal=user-...@prod.example.com > log4j:WARN No such property [maxFileSize] in > org.apache.log4j.rolling.RollingFileAppender. > log4j:WARN No such property [maxBackupIndex] in > org.apache.log4j.rolling.RollingFileAppender. > Exception in thread "main" org.apache.spark.SparkException: Application > application_1564558112805_1794 failed 2 times due to AM Container for > appattempt_1564558112805_1794_02 exited with exitCode: 1 > For more detailed output, check the application tracking page: > https://0.0.0.0:8190/applicationhistory/app/application_1564558112805_1794 > Then click on links to logs of each attempt. > Diagnostics: Exception from container-launch. > Container id: container_e1987_1564558112805_1794_02_01 > Exit code: 1 > Shell output: main : command provided 1 > main : run as user is user-etl > main : requested yarn user is user-etl > Getting exit code file... > Creating script paths... > Writing pid file... > Writing to tmp file > /hadoop/2/yarn/local/nmPrivate/application_1564558112805_1794/container_e1987_1564558112805_1794_02_01/container_e1987_1564558112805_1794_02_01.pid.tmp > Writing to cgroup task files... > Creating local dirs... > Launching container... > Getting exit code file... > Creating script paths... > Container exited with a non-zero exit code 1. Error file: prelaunch.err. > Last 4096 bytes of prelaunch.err : > Last 4096 bytes of stderr : > log4j:WARN No such property [maxFileSize] in > org.apache.log4j.rolling.RollingFileAppender. > log4j:WARN No such property [maxBackupIndex] in > org.apache.log4j.rolling.RollingFileAppender. > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in > [jar:file:/hadoop/2/yarn/local/usercache/user-etl/filecache/58/__spark_libs__4358879230136591830.zip/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > [jar:file:/apache/releases/hbase-1.1.2.2.6.4.1/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > [jar:file:/apache/releases/hadoop-2.7.3.2.6.4.3/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > explanation. > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] > Exception in thread "main" org.apache.spark.SparkException: Keytab file: > /apache/spark-2.3.0-bin-2.7.3/conf/user-etl.keytab does not exist > at > org.apache.spark.deploy.SparkHadoopUtil.loginUserFromKeytab(SparkHadoopUtil.scala:131) > at > org.apache.spark.deploy.yarn.ApplicationMaster$.main(ApplicationMaster.scala:846) > at > org.apache.spark.deploy.yarn.ExecutorLauncher$.main(ApplicationMaster.scala:889) > at > org.apache.spark.deploy.yarn.ExecutorLauncher.main(ApplicationMaster.scala) > Failing this attempt. Failing the application. > at > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.waitForApplication(YarnClientSchedulerBackend.scala:95) > at > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:62) > at > org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:185) > at org.apache.spark.SparkContext.(SparkContext.scala:509) > at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2466) > at > org.apache.spark.sql.SparkSession$Builder.$anonfun$getOrCreate$5(SparkSession.scala:948) > at scala.Option.getOrElse(Option.scala:138) > at > org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:939) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLEnv$.init(SparkSQLEnv.scala:48) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.(SparkSQLCLIDriver.scala:315) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:166) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLID
[jira] [Commented] (SPARK-28634) Failed to start SparkSession with Keytab file
[ https://issues.apache.org/jira/browse/SPARK-28634?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16902702#comment-16902702 ] Yuming Wang commented on SPARK-28634: - Thank you [~vanzin] It works. > Failed to start SparkSession with Keytab file > -- > > Key: SPARK-28634 > URL: https://issues.apache.org/jira/browse/SPARK-28634 > Project: Spark > Issue Type: Bug > Components: YARN >Affects Versions: 3.0.0 >Reporter: Yuming Wang >Priority: Major > > {noformat} > [user-etl@hermesdevour002-700165 spark-3.0.0-SNAPSHOT-bin-2.7.4]$ > bin/spark-sql --master yarn --conf > spark.yarn.keytab=/apache/spark-2.3.0-bin-2.7.3/conf/user-etl.keytab --conf > spark.yarn.principal=user-...@prod.example.com > log4j:WARN No such property [maxFileSize] in > org.apache.log4j.rolling.RollingFileAppender. > log4j:WARN No such property [maxBackupIndex] in > org.apache.log4j.rolling.RollingFileAppender. > Exception in thread "main" org.apache.spark.SparkException: Application > application_1564558112805_1794 failed 2 times due to AM Container for > appattempt_1564558112805_1794_02 exited with exitCode: 1 > For more detailed output, check the application tracking page: > https://0.0.0.0:8190/applicationhistory/app/application_1564558112805_1794 > Then click on links to logs of each attempt. > Diagnostics: Exception from container-launch. > Container id: container_e1987_1564558112805_1794_02_01 > Exit code: 1 > Shell output: main : command provided 1 > main : run as user is user-etl > main : requested yarn user is user-etl > Getting exit code file... > Creating script paths... > Writing pid file... > Writing to tmp file > /hadoop/2/yarn/local/nmPrivate/application_1564558112805_1794/container_e1987_1564558112805_1794_02_01/container_e1987_1564558112805_1794_02_01.pid.tmp > Writing to cgroup task files... > Creating local dirs... > Launching container... > Getting exit code file... > Creating script paths... > Container exited with a non-zero exit code 1. Error file: prelaunch.err. > Last 4096 bytes of prelaunch.err : > Last 4096 bytes of stderr : > log4j:WARN No such property [maxFileSize] in > org.apache.log4j.rolling.RollingFileAppender. > log4j:WARN No such property [maxBackupIndex] in > org.apache.log4j.rolling.RollingFileAppender. > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in > [jar:file:/hadoop/2/yarn/local/usercache/user-etl/filecache/58/__spark_libs__4358879230136591830.zip/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > [jar:file:/apache/releases/hbase-1.1.2.2.6.4.1/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > [jar:file:/apache/releases/hadoop-2.7.3.2.6.4.3/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > explanation. > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] > Exception in thread "main" org.apache.spark.SparkException: Keytab file: > /apache/spark-2.3.0-bin-2.7.3/conf/user-etl.keytab does not exist > at > org.apache.spark.deploy.SparkHadoopUtil.loginUserFromKeytab(SparkHadoopUtil.scala:131) > at > org.apache.spark.deploy.yarn.ApplicationMaster$.main(ApplicationMaster.scala:846) > at > org.apache.spark.deploy.yarn.ExecutorLauncher$.main(ApplicationMaster.scala:889) > at > org.apache.spark.deploy.yarn.ExecutorLauncher.main(ApplicationMaster.scala) > Failing this attempt. Failing the application. > at > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.waitForApplication(YarnClientSchedulerBackend.scala:95) > at > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:62) > at > org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:185) > at org.apache.spark.SparkContext.(SparkContext.scala:509) > at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2466) > at > org.apache.spark.sql.SparkSession$Builder.$anonfun$getOrCreate$5(SparkSession.scala:948) > at scala.Option.getOrElse(Option.scala:138) > at > org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:939) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLEnv$.init(SparkSQLEnv.scala:48) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.(SparkSQLCLIDriver.scala:315) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:166) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Meth
[jira] [Commented] (SPARK-28634) Failed to start SparkSession with Keytab file
[ https://issues.apache.org/jira/browse/SPARK-28634?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16902238#comment-16902238 ] Marcelo Vanzin commented on SPARK-28634: Ah. If you use {{--principal}} and {{--keytab}} this works. The config name has changed in master and you're using the deprecated ones; the YARN client code removes them from the config in client mode, but only the new names: https://github.com/apache/spark/blob/master/resource-managers/yarn/src/main/scala/org/apache/spark/deploy/yarn/Client.scala#L769 For proper backwards compatibility it needs to remove the old names too. (Or make a change in the AM instead to ignore the keytab when running in client mode, which avoids the above hack.) > Failed to start SparkSession with Keytab file > -- > > Key: SPARK-28634 > URL: https://issues.apache.org/jira/browse/SPARK-28634 > Project: Spark > Issue Type: Bug > Components: YARN >Affects Versions: 3.0.0 >Reporter: Yuming Wang >Priority: Major > > {noformat} > [user-etl@hermesdevour002-700165 spark-3.0.0-SNAPSHOT-bin-2.7.4]$ > bin/spark-sql --master yarn --conf > spark.yarn.keytab=/apache/spark-2.3.0-bin-2.7.3/conf/user-etl.keytab --conf > spark.yarn.principal=user-...@prod.example.com > log4j:WARN No such property [maxFileSize] in > org.apache.log4j.rolling.RollingFileAppender. > log4j:WARN No such property [maxBackupIndex] in > org.apache.log4j.rolling.RollingFileAppender. > Exception in thread "main" org.apache.spark.SparkException: Application > application_1564558112805_1794 failed 2 times due to AM Container for > appattempt_1564558112805_1794_02 exited with exitCode: 1 > For more detailed output, check the application tracking page: > https://0.0.0.0:8190/applicationhistory/app/application_1564558112805_1794 > Then click on links to logs of each attempt. > Diagnostics: Exception from container-launch. > Container id: container_e1987_1564558112805_1794_02_01 > Exit code: 1 > Shell output: main : command provided 1 > main : run as user is user-etl > main : requested yarn user is user-etl > Getting exit code file... > Creating script paths... > Writing pid file... > Writing to tmp file > /hadoop/2/yarn/local/nmPrivate/application_1564558112805_1794/container_e1987_1564558112805_1794_02_01/container_e1987_1564558112805_1794_02_01.pid.tmp > Writing to cgroup task files... > Creating local dirs... > Launching container... > Getting exit code file... > Creating script paths... > Container exited with a non-zero exit code 1. Error file: prelaunch.err. > Last 4096 bytes of prelaunch.err : > Last 4096 bytes of stderr : > log4j:WARN No such property [maxFileSize] in > org.apache.log4j.rolling.RollingFileAppender. > log4j:WARN No such property [maxBackupIndex] in > org.apache.log4j.rolling.RollingFileAppender. > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in > [jar:file:/hadoop/2/yarn/local/usercache/user-etl/filecache/58/__spark_libs__4358879230136591830.zip/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > [jar:file:/apache/releases/hbase-1.1.2.2.6.4.1/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > [jar:file:/apache/releases/hadoop-2.7.3.2.6.4.3/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > explanation. > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] > Exception in thread "main" org.apache.spark.SparkException: Keytab file: > /apache/spark-2.3.0-bin-2.7.3/conf/user-etl.keytab does not exist > at > org.apache.spark.deploy.SparkHadoopUtil.loginUserFromKeytab(SparkHadoopUtil.scala:131) > at > org.apache.spark.deploy.yarn.ApplicationMaster$.main(ApplicationMaster.scala:846) > at > org.apache.spark.deploy.yarn.ExecutorLauncher$.main(ApplicationMaster.scala:889) > at > org.apache.spark.deploy.yarn.ExecutorLauncher.main(ApplicationMaster.scala) > Failing this attempt. Failing the application. > at > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.waitForApplication(YarnClientSchedulerBackend.scala:95) > at > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:62) > at > org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:185) > at org.apache.spark.SparkContext.(SparkContext.scala:509) > at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2466) > at > org.apache.spark.sql.SparkSession$Builder.$anonfun$getOrCreate$5(SparkSession.scala:948) > at scala.Option.getOrElse(Option.scala:138) > at > org.apache.spark.sql
[jira] [Commented] (SPARK-28634) Failed to start SparkSession with Keytab file
[ https://issues.apache.org/jira/browse/SPARK-28634?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16900769#comment-16900769 ] Yuming Wang commented on SPARK-28634: - cc [~vanzin] I will provide more information if you need it. > Failed to start SparkSession with Keytab file > -- > > Key: SPARK-28634 > URL: https://issues.apache.org/jira/browse/SPARK-28634 > Project: Spark > Issue Type: Bug > Components: YARN >Affects Versions: 3.0.0 >Reporter: Yuming Wang >Priority: Major > > {noformat} > [user-etl@hermesdevour002-700165 spark-3.0.0-SNAPSHOT-bin-2.7.4]$ > bin/spark-sql --master yarn --conf > spark.yarn.keytab=/apache/spark-2.3.0-bin-2.7.3/conf/user-etl.keytab --conf > spark.yarn.principal=user-...@prod.example.com > log4j:WARN No such property [maxFileSize] in > org.apache.log4j.rolling.RollingFileAppender. > log4j:WARN No such property [maxBackupIndex] in > org.apache.log4j.rolling.RollingFileAppender. > Exception in thread "main" org.apache.spark.SparkException: Application > application_1564558112805_1794 failed 2 times due to AM Container for > appattempt_1564558112805_1794_02 exited with exitCode: 1 > For more detailed output, check the application tracking page: > https://0.0.0.0:8190/applicationhistory/app/application_1564558112805_1794 > Then click on links to logs of each attempt. > Diagnostics: Exception from container-launch. > Container id: container_e1987_1564558112805_1794_02_01 > Exit code: 1 > Shell output: main : command provided 1 > main : run as user is user-etl > main : requested yarn user is user-etl > Getting exit code file... > Creating script paths... > Writing pid file... > Writing to tmp file > /hadoop/2/yarn/local/nmPrivate/application_1564558112805_1794/container_e1987_1564558112805_1794_02_01/container_e1987_1564558112805_1794_02_01.pid.tmp > Writing to cgroup task files... > Creating local dirs... > Launching container... > Getting exit code file... > Creating script paths... > Container exited with a non-zero exit code 1. Error file: prelaunch.err. > Last 4096 bytes of prelaunch.err : > Last 4096 bytes of stderr : > log4j:WARN No such property [maxFileSize] in > org.apache.log4j.rolling.RollingFileAppender. > log4j:WARN No such property [maxBackupIndex] in > org.apache.log4j.rolling.RollingFileAppender. > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in > [jar:file:/hadoop/2/yarn/local/usercache/user-etl/filecache/58/__spark_libs__4358879230136591830.zip/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > [jar:file:/apache/releases/hbase-1.1.2.2.6.4.1/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > [jar:file:/apache/releases/hadoop-2.7.3.2.6.4.3/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > explanation. > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] > Exception in thread "main" org.apache.spark.SparkException: Keytab file: > /apache/spark-2.3.0-bin-2.7.3/conf/user-etl.keytab does not exist > at > org.apache.spark.deploy.SparkHadoopUtil.loginUserFromKeytab(SparkHadoopUtil.scala:131) > at > org.apache.spark.deploy.yarn.ApplicationMaster$.main(ApplicationMaster.scala:846) > at > org.apache.spark.deploy.yarn.ExecutorLauncher$.main(ApplicationMaster.scala:889) > at > org.apache.spark.deploy.yarn.ExecutorLauncher.main(ApplicationMaster.scala) > Failing this attempt. Failing the application. > at > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.waitForApplication(YarnClientSchedulerBackend.scala:95) > at > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:62) > at > org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:185) > at org.apache.spark.SparkContext.(SparkContext.scala:509) > at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2466) > at > org.apache.spark.sql.SparkSession$Builder.$anonfun$getOrCreate$5(SparkSession.scala:948) > at scala.Option.getOrElse(Option.scala:138) > at > org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:939) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLEnv$.init(SparkSQLEnv.scala:48) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.(SparkSQLCLIDriver.scala:315) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:166) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala) > at sun.reflect.NativeMethodA