First, thanks to everyone for their assistance and recommendations. @Marcelo
I applied the patch that you recommended and am now able to get into the shell, thank you worked great after I realized that the pom was pointing to the 1.3.0-SNAPSHOT for parent, need to bump that down to 1.2.1. @Zhan Need to apply this patch next. I tried to start the spark-thriftserver but and it starts, then fails with like this: I have the entries in my spark-default.conf, but not the patch applied. ./sbin/start-thriftserver.sh --master yarn --executor-memory 1024m --hiveconf hive.server2.thrift.port=10001 5/03/06 12:34:17 INFO ui.SparkUI: Started SparkUI at http://hadoopdev01.opsdatastore.com:404015/03/06 12:34:18 INFO impl.TimelineClientImpl: Timeline service address: http://hadoopdev02.opsdatastore.com:8188/ws/v1/timeline/15/03/06 12:34:18 INFO client.RMProxy: Connecting to ResourceManager at hadoopdev02.opsdatastore.com/192.168.15.154:805015/03/06 12:34:18 INFO yarn.Client: Requesting a new application from cluster with 4 NodeManagers15/03/06 12:34:18 INFO yarn.Client: Verifying our application has not requested more than the maximum memory capability of the cluster (8192 MB per container)15/03/06 12:34:18 INFO yarn.Client: Will allocate AM container, with 896 MB memory including 384 MB overhead15/03/06 12:34:18 INFO yarn.Client: Setting up container launch context for our AM15/03/06 12:34:18 INFO yarn.Client: Preparing resources for our AM container15/03/06 12:34:19 WARN shortcircuit.DomainSocketFactory: The short-circuit local reads feature cannot be used because libhadoop cannot be loaded.15/03/06 12:34:19 INFO yarn.Client: Uploading resource file:/root/spark-1.2.1-bin-hadoop2.6/lib/spark-assembly-1.2.1-hadoop2.6.0.jar -> hdfs://hadoopdev01.opsdatastore.com:8020/user/root/.sparkStaging/application_1425078697953_0018/spark-assembly-1.2.1-hadoop2.6.0.jar15/03/06 12:34:21 INFO yarn.Client: Setting up the launch environment for our AM container15/03/06 12:34:21 INFO spark.SecurityManager: Changing view acls to: root15/03/06 12:34:21 INFO spark.SecurityManager: Changing modify acls to: root15/03/06 12:34:21 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); users with modify permissions: Set(root)15/03/06 12:34:21 INFO yarn.Client: Submitting application 18 to ResourceManager15/03/06 12:34:21 INFO impl.YarnClientImpl: Submitted application application_1425078697953_001815/03/06 12:34:22 INFO yarn.Client: Application report for application_1425078697953_0018 (state: ACCEPTED)15/03/06 12:34:22 INFO yarn.Client: client token: N/A diagnostics: N/A ApplicationMaster host: N/A ApplicationMaster RPC port: -1 queue: default start time: 1425663261755 final status: UNDEFINED tracking URL: http://hadoopdev02.opsdatastore.com:8088/proxy/application_1425078697953_0018/ user: root15/03/06 12:34:23 INFO yarn.Client: Application report for application_1425078697953_0018 (state: ACCEPTED)15/03/06 12:34:24 INFO yarn.Client: Application report for application_1425078697953_0018 (state: ACCEPTED)15/03/06 12:34:25 INFO yarn.Client: Application report for application_1425078697953_0018 (state: ACCEPTED)15/03/06 12:34:26 INFO yarn.Client: Application report for application_1425078697953_0018 (state: ACCEPTED)15/03/06 12:34:27 INFO cluster.YarnClientSchedulerBackend: ApplicationMaster registered as Actor[akka.tcp://sparkyar...@hadoopdev08.opsdatastore.com:40201/user/YarnAM#-557112763]15/03/06 12:34:27 INFO cluster.YarnClientSchedulerBackend: Add WebUI Filter. org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, Map(PROXY_HOSTS -> hadoopdev02.opsdatastore.com, PROXY_URI_BASES -> http://hadoopdev02.opsdatastore.com:8088/proxy/application_1425078697953_0018), /proxy/application_1425078697953_001815/03/06 12:34:27 INFO ui.JettyUtils: Adding filter: org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter15/03/06 12:34:27 INFO yarn.Client: Application report for application_1425078697953_0018 (state: RUNNING)15/03/06 12:34:27 INFO yarn.Client: client token: N/A diagnostics: N/A ApplicationMaster host: hadoopdev08.opsdatastore.com ApplicationMaster RPC port: 0 queue: default start time: 1425663261755 final status: UNDEFINED tracking URL: http://hadoopdev02.opsdatastore.com:8088/proxy/application_1425078697953_0018/ user: root15/03/06 12:34:27 INFO cluster.YarnClientSchedulerBackend: Application application_1425078697953_0018 has started running.15/03/06 12:34:28 INFO netty.NettyBlockTransferService: Server created on 4612415/03/06 12:34:28 INFO storage.BlockManagerMaster: Trying to register BlockManager15/03/06 12:34:28 INFO storage.BlockManagerMasterActor: Registering block manager hadoopdev01.opsdatastore.com:46124 with 265.4 MB RAM, BlockManagerId(<driver>, hadoopdev01.opsdatastore.com, 46124)15/03/06 12:34:28 INFO storage.BlockManagerMaster: Registered BlockManager15/03/06 12:34:47 INFO cluster.YarnClientSchedulerBackend: SchedulerBackend is ready for scheduling beginning after waiting maxRegisteredResourcesWaitingTime: 30000(ms)15/03/06 12:34:48 INFO hive.metastore: Trying to connect to metastore with URI thrift://hadoopdev03.opsdatastore.com:908315/03/06 12:34:48 INFO hive.metastore: Connected to metastore.15/03/06 12:34:49 INFO session.SessionState: No Tez session required at this point. hive.execution.engine=mr.15/03/06 12:34:49 INFO service.AbstractService: HiveServer2: Async execution pool size 10015/03/06 12:34:49 INFO service.AbstractService: Service:OperationManager is inited.15/03/06 12:34:49 INFO service.AbstractService: Service: SessionManager is inited.15/03/06 12:34:49 INFO service.AbstractService: Service: CLIService is inited.15/03/06 12:34:49 INFO service.AbstractService: Service:ThriftBinaryCLIService is inited.15/03/06 12:34:49 INFO service.AbstractService: Service: HiveServer2 is inited.15/03/06 12:34:49 INFO service.AbstractService: Service:OperationManager is started.15/03/06 12:34:49 INFO service.AbstractService: Service:SessionManager is started.15/03/06 12:34:49 INFO service.AbstractService: Service:CLIService is started.15/03/06 12:34:49 INFO hive.metastore: Trying to connect to metastore with URI thrift://hadoopdev03.opsdatastore.com:908315/03/06 12:34:49 INFO hive.metastore: Connected to metastore.15/03/06 12:34:49 INFO service.AbstractService: Service:ThriftBinaryCLIService is started.15/03/06 12:34:49 INFO service.AbstractService: Service:HiveServer2 is started.15/03/06 12:34:49 INFO thriftserver.HiveThriftServer2: HiveThriftServer2 started15/03/06 12:34:49 INFO thrift.ThriftCLIService: ThriftBinaryCLIService listening on 0.0.0.0/0.0.0.0:1000115/03/06 12:34:58 WARN remote.ReliableDeliverySupervisor: Association with remote system [akka.tcp://sparkyar...@hadoopdev08.opsdatastore.com:40201] has failed, address is now gated for [5000] ms. Reason is: [Disassociated].15/03/06 12:35:02 INFO cluster.YarnClientSchedulerBackend: ApplicationMaster registered as Actor[akka.tcp://sparkyar...@hadoopdev08.opsdatastore.com:53176/user/YarnAM#-1793579186]15/03/06 12:35:02 INFO cluster.YarnClientSchedulerBackend: Add WebUI Filter. org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, Map(PROXY_HOSTS -> hadoopdev02.opsdatastore.com, PROXY_URI_BASES -> http://hadoopdev02.opsdatastore.com:8088/proxy/application_1425078697953_0018), /proxy/application_1425078697953_001815/03/06 12:35:02 INFO ui.JettyUtils: Adding filter: org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter 15/03/06 12:35:38 WARN remote.ReliableDeliverySupervisor: Association with remote system [akka.tcp://sparkyar...@hadoopdev08.opsdatastore.com:53176] has failed, address is now gated for [5000] ms. Reason is: [Disassociated].15/03/06 12:35:39 ERROR cluster.YarnClientSchedulerBackend: Yarn application has already exited with state FINISHED!15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/metrics/json,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/stages/stage/kill,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/static,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/executors/threadDump/json,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/executors/threadDump,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/executors/json,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/executors,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/environment/json,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/environment,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/storage/rdd/json,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/storage/rdd,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/storage/json,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/storage,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/stages/pool/json,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/stages/pool,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/stages/stage/json,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/stages/stage,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/stages/json,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/stages,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/jobs/job/json,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/jobs/job,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/jobs/json,null}15/03/06 12:35:39 INFO handler.ContextHandler: stopped o.e.j.s.ServletContextHandler{/jobs,null}15/03/06 12:35:39 INFO ui.SparkUI: Stopped Spark web UI at http://hadoopdev01.opsdatastore.com:404015/03/06 12:35:39 INFO scheduler.DAGScheduler: Stopping DAGScheduler15/03/06 12:35:39 INFO cluster.YarnClientSchedulerBackend: Shutting down all executors15/03/06 12:35:39 INFO cluster.YarnClientSchedulerBackend: Asking each executor to shut down15/03/06 12:35:39 INFO cluster.YarnClientSchedulerBackend: Stopped15/03/06 12:35:40 INFO spark.MapOutputTrackerMasterActor: MapOutputTrackerActor stopped!15/03/06 12:35:40 INFO storage.MemoryStore: MemoryStore cleared15/03/06 12:35:40 INFO storage.BlockManager: BlockManager stopped15/03/06 12:35:40 INFO storage.BlockManagerMaster: BlockManagerMaster stopped15/03/06 12:35:40 INFO spark.SparkContext: Successfully stopped SparkContext15/03/06 12:35:40 INFO remote.RemoteActorRefProvider$RemotingTerminator: Shutting down remote daemon.15/03/06 12:35:40 INFO remote.RemoteActorRefProvider$RemotingTerminator: Remote daemon shut down; proceeding with flushing remote transports.15/03/06 12:35:40 INFO remote.RemoteActorRefProvider$RemotingTerminator: Remoting shut down. Thanks again for the help. -Todd On Thu, Mar 5, 2015 at 7:06 PM, Zhan Zhang <zzh...@hortonworks.com> wrote: > In addition, you may need following patch if it is not in 1.2.1 to solve > some system property issue if you use HDP 2.2. > > https://github.com/apache/spark/pull/3409 > > You can follow the following link to set hdp.version for java options. > > http://hortonworks.com/hadoop-tutorial/using-apache-spark-hdp/ > > Thanks. > > Zhan Zhang > > On Mar 5, 2015, at 11:09 AM, Marcelo Vanzin <van...@cloudera.com> wrote: > > It seems from the excerpt below that your cluster is set up to use the > Yarn ATS, and the code is failing in that path. I think you'll need to > apply the following patch to your Spark sources if you want this to > work: > > https://github.com/apache/spark/pull/3938 > > On Thu, Mar 5, 2015 at 10:04 AM, Todd Nist <tsind...@gmail.com> wrote: > > > org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.serviceInit(YarnClientImpl.java:166) > at > org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) > at > org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:65) > at > > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:57) > at > > org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:140) > at org.apache.spark.SparkContext.<init>(SparkContext.scala:348) > > > -- > Marcelo > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > > >