First, thanks to everyone for their assistance and recommendations.

@Marcelo

I applied the patch that you recommended and am now able to get into the
shell, thank you worked great after I realized that the pom was pointing to
the 1.3.0-SNAPSHOT for parent, need to bump that down to 1.2.1.

@Zhan

Need to apply this patch next.  I tried to start the spark-thriftserver but
and it starts, then fails with like this:  I have the entries in my
spark-default.conf, but not the patch applied.

./sbin/start-thriftserver.sh --master yarn --executor-memory 1024m
--hiveconf hive.server2.thrift.port=10001

5/03/06 12:34:17 INFO ui.SparkUI: Started SparkUI at
http://hadoopdev01.opsdatastore.com:404015/03/06 12:34:18 INFO
impl.TimelineClientImpl: Timeline service address:
http://hadoopdev02.opsdatastore.com:8188/ws/v1/timeline/15/03/06
12:34:18 INFO client.RMProxy: Connecting to ResourceManager at
hadoopdev02.opsdatastore.com/192.168.15.154:805015/03/06 12:34:18 INFO
yarn.Client: Requesting a new application from cluster with 4
NodeManagers15/03/06 12:34:18 INFO yarn.Client: Verifying our
application has not requested more than the maximum memory capability
of the cluster (8192 MB per container)15/03/06 12:34:18 INFO
yarn.Client: Will allocate AM container, with 896 MB memory including
384 MB overhead15/03/06 12:34:18 INFO yarn.Client: Setting up
container launch context for our AM15/03/06 12:34:18 INFO yarn.Client:
Preparing resources for our AM container15/03/06 12:34:19 WARN
shortcircuit.DomainSocketFactory: The short-circuit local reads
feature cannot be used because libhadoop cannot be loaded.15/03/06
12:34:19 INFO yarn.Client: Uploading resource
file:/root/spark-1.2.1-bin-hadoop2.6/lib/spark-assembly-1.2.1-hadoop2.6.0.jar
-> 
hdfs://hadoopdev01.opsdatastore.com:8020/user/root/.sparkStaging/application_1425078697953_0018/spark-assembly-1.2.1-hadoop2.6.0.jar15/03/06
12:34:21 INFO yarn.Client: Setting up the launch environment for our
AM container15/03/06 12:34:21 INFO spark.SecurityManager: Changing
view acls to: root15/03/06 12:34:21 INFO spark.SecurityManager:
Changing modify acls to: root15/03/06 12:34:21 INFO
spark.SecurityManager: SecurityManager: authentication disabled; ui
acls disabled; users with view permissions: Set(root); users with
modify permissions: Set(root)15/03/06 12:34:21 INFO yarn.Client:
Submitting application 18 to ResourceManager15/03/06 12:34:21 INFO
impl.YarnClientImpl: Submitted application
application_1425078697953_001815/03/06 12:34:22 INFO yarn.Client:
Application report for application_1425078697953_0018 (state:
ACCEPTED)15/03/06 12:34:22 INFO yarn.Client:
     client token: N/A
     diagnostics: N/A
     ApplicationMaster host: N/A
     ApplicationMaster RPC port: -1
     queue: default
     start time: 1425663261755
     final status: UNDEFINED
     tracking URL:
http://hadoopdev02.opsdatastore.com:8088/proxy/application_1425078697953_0018/
     user: root15/03/06 12:34:23 INFO yarn.Client: Application report
for application_1425078697953_0018 (state: ACCEPTED)15/03/06 12:34:24
INFO yarn.Client: Application report for
application_1425078697953_0018 (state: ACCEPTED)15/03/06 12:34:25 INFO
yarn.Client: Application report for application_1425078697953_0018
(state: ACCEPTED)15/03/06 12:34:26 INFO yarn.Client: Application
report for application_1425078697953_0018 (state: ACCEPTED)15/03/06
12:34:27 INFO cluster.YarnClientSchedulerBackend: ApplicationMaster
registered as 
Actor[akka.tcp://sparkyar...@hadoopdev08.opsdatastore.com:40201/user/YarnAM#-557112763]15/03/06
12:34:27 INFO cluster.YarnClientSchedulerBackend: Add WebUI Filter.
org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter,
Map(PROXY_HOSTS -> hadoopdev02.opsdatastore.com, PROXY_URI_BASES ->
http://hadoopdev02.opsdatastore.com:8088/proxy/application_1425078697953_0018),
/proxy/application_1425078697953_001815/03/06 12:34:27 INFO
ui.JettyUtils: Adding filter:
org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter15/03/06
12:34:27 INFO yarn.Client: Application report for
application_1425078697953_0018 (state: RUNNING)15/03/06 12:34:27 INFO
yarn.Client:
     client token: N/A
     diagnostics: N/A
     ApplicationMaster host: hadoopdev08.opsdatastore.com
     ApplicationMaster RPC port: 0
     queue: default
     start time: 1425663261755
     final status: UNDEFINED
     tracking URL:
http://hadoopdev02.opsdatastore.com:8088/proxy/application_1425078697953_0018/
     user: root15/03/06 12:34:27 INFO
cluster.YarnClientSchedulerBackend: Application
application_1425078697953_0018 has started running.15/03/06 12:34:28
INFO netty.NettyBlockTransferService: Server created on 4612415/03/06
12:34:28 INFO storage.BlockManagerMaster: Trying to register
BlockManager15/03/06 12:34:28 INFO storage.BlockManagerMasterActor:
Registering block manager hadoopdev01.opsdatastore.com:46124 with
265.4 MB RAM, BlockManagerId(<driver>, hadoopdev01.opsdatastore.com,
46124)15/03/06 12:34:28 INFO storage.BlockManagerMaster: Registered
BlockManager15/03/06 12:34:47 INFO cluster.YarnClientSchedulerBackend:
SchedulerBackend is ready for scheduling beginning after waiting
maxRegisteredResourcesWaitingTime: 30000(ms)15/03/06 12:34:48 INFO
hive.metastore: Trying to connect to metastore with URI
thrift://hadoopdev03.opsdatastore.com:908315/03/06 12:34:48 INFO
hive.metastore: Connected to metastore.15/03/06 12:34:49 INFO
session.SessionState: No Tez session required at this point.
hive.execution.engine=mr.15/03/06 12:34:49 INFO
service.AbstractService: HiveServer2: Async execution pool size
10015/03/06 12:34:49 INFO service.AbstractService:
Service:OperationManager is inited.15/03/06 12:34:49 INFO
service.AbstractService: Service: SessionManager is inited.15/03/06
12:34:49 INFO service.AbstractService: Service: CLIService is
inited.15/03/06 12:34:49 INFO service.AbstractService:
Service:ThriftBinaryCLIService is inited.15/03/06 12:34:49 INFO
service.AbstractService: Service: HiveServer2 is inited.15/03/06
12:34:49 INFO service.AbstractService: Service:OperationManager is
started.15/03/06 12:34:49 INFO service.AbstractService:
Service:SessionManager is started.15/03/06 12:34:49 INFO
service.AbstractService: Service:CLIService is started.15/03/06
12:34:49 INFO hive.metastore: Trying to connect to metastore with URI
thrift://hadoopdev03.opsdatastore.com:908315/03/06 12:34:49 INFO
hive.metastore: Connected to metastore.15/03/06 12:34:49 INFO
service.AbstractService: Service:ThriftBinaryCLIService is
started.15/03/06 12:34:49 INFO service.AbstractService:
Service:HiveServer2 is started.15/03/06 12:34:49 INFO
thriftserver.HiveThriftServer2: HiveThriftServer2 started15/03/06
12:34:49 INFO thrift.ThriftCLIService: ThriftBinaryCLIService
listening on 0.0.0.0/0.0.0.0:1000115/03/06 12:34:58 WARN
remote.ReliableDeliverySupervisor: Association with remote system
[akka.tcp://sparkyar...@hadoopdev08.opsdatastore.com:40201] has
failed, address is now gated for [5000] ms. Reason is:
[Disassociated].15/03/06 12:35:02 INFO
cluster.YarnClientSchedulerBackend: ApplicationMaster registered as
Actor[akka.tcp://sparkyar...@hadoopdev08.opsdatastore.com:53176/user/YarnAM#-1793579186]15/03/06
12:35:02 INFO cluster.YarnClientSchedulerBackend: Add WebUI Filter.
org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter,
Map(PROXY_HOSTS -> hadoopdev02.opsdatastore.com, PROXY_URI_BASES ->
http://hadoopdev02.opsdatastore.com:8088/proxy/application_1425078697953_0018),
/proxy/application_1425078697953_001815/03/06 12:35:02 INFO
ui.JettyUtils: Adding filter:
org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter
15/03/06 12:35:38 WARN remote.ReliableDeliverySupervisor: Association
with remote system
[akka.tcp://sparkyar...@hadoopdev08.opsdatastore.com:53176] has
failed, address is now gated for [5000] ms. Reason is:
[Disassociated].15/03/06 12:35:39 ERROR
cluster.YarnClientSchedulerBackend: Yarn application has already
exited with state FINISHED!15/03/06 12:35:39 INFO
handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/metrics/json,null}15/03/06 12:35:39
INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/stages/stage/kill,null}15/03/06
12:35:39 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/,null}15/03/06 12:35:39 INFO
handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/static,null}15/03/06 12:35:39 INFO
handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/executors/threadDump/json,null}15/03/06
12:35:39 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/executors/threadDump,null}15/03/06
12:35:39 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/executors/json,null}15/03/06 12:35:39
INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/executors,null}15/03/06 12:35:39 INFO
handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/environment/json,null}15/03/06 12:35:39
INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/environment,null}15/03/06 12:35:39 INFO
handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/storage/rdd/json,null}15/03/06 12:35:39
INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/storage/rdd,null}15/03/06 12:35:39 INFO
handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/storage/json,null}15/03/06 12:35:39
INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/storage,null}15/03/06 12:35:39 INFO
handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/stages/pool/json,null}15/03/06 12:35:39
INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/stages/pool,null}15/03/06 12:35:39 INFO
handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/stages/stage/json,null}15/03/06
12:35:39 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/stages/stage,null}15/03/06 12:35:39
INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/stages/json,null}15/03/06 12:35:39 INFO
handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/stages,null}15/03/06 12:35:39 INFO
handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/jobs/job/json,null}15/03/06 12:35:39
INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/jobs/job,null}15/03/06 12:35:39 INFO
handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/jobs/json,null}15/03/06 12:35:39 INFO
handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/jobs,null}15/03/06 12:35:39 INFO
ui.SparkUI: Stopped Spark web UI at
http://hadoopdev01.opsdatastore.com:404015/03/06 12:35:39 INFO
scheduler.DAGScheduler: Stopping DAGScheduler15/03/06 12:35:39 INFO
cluster.YarnClientSchedulerBackend: Shutting down all
executors15/03/06 12:35:39 INFO cluster.YarnClientSchedulerBackend:
Asking each executor to shut down15/03/06 12:35:39 INFO
cluster.YarnClientSchedulerBackend: Stopped15/03/06 12:35:40 INFO
spark.MapOutputTrackerMasterActor: MapOutputTrackerActor
stopped!15/03/06 12:35:40 INFO storage.MemoryStore: MemoryStore
cleared15/03/06 12:35:40 INFO storage.BlockManager: BlockManager
stopped15/03/06 12:35:40 INFO storage.BlockManagerMaster:
BlockManagerMaster stopped15/03/06 12:35:40 INFO spark.SparkContext:
Successfully stopped SparkContext15/03/06 12:35:40 INFO
remote.RemoteActorRefProvider$RemotingTerminator: Shutting down remote
daemon.15/03/06 12:35:40 INFO
remote.RemoteActorRefProvider$RemotingTerminator: Remote daemon shut
down; proceeding with flushing remote transports.15/03/06 12:35:40
INFO remote.RemoteActorRefProvider$RemotingTerminator: Remoting shut
down.



Thanks again for the help.

-Todd

On Thu, Mar 5, 2015 at 7:06 PM, Zhan Zhang <zzh...@hortonworks.com> wrote:

>  In addition, you may need following patch if it is not in 1.2.1 to solve
> some system property issue if you use HDP 2.2.
>
>  https://github.com/apache/spark/pull/3409
>
>  You can follow the following link to set hdp.version for java options.
>
>  http://hortonworks.com/hadoop-tutorial/using-apache-spark-hdp/
>
>  Thanks.
>
>  Zhan Zhang
>
>  On Mar 5, 2015, at 11:09 AM, Marcelo Vanzin <van...@cloudera.com> wrote:
>
> It seems from the excerpt below that your cluster is set up to use the
> Yarn ATS, and the code is failing in that path. I think you'll need to
> apply the following patch to your Spark sources if you want this to
> work:
>
> https://github.com/apache/spark/pull/3938
>
> On Thu, Mar 5, 2015 at 10:04 AM, Todd Nist <tsind...@gmail.com> wrote:
>
>
> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.serviceInit(YarnClientImpl.java:166)
>    at
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
>    at
> org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:65)
>    at
>
> org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:57)
>    at
>
> org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:140)
>    at org.apache.spark.SparkContext.<init>(SparkContext.scala:348)
>
>
> --
> Marcelo
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>
>

Reply via email to