timyuer commented on PR #984:
URL: https://github.com/apache/bigtop/pull/984#issuecomment-1217757517

   Spark Thrift Server Check
   
![spark_check_thriftserver_20220817172633](https://user-images.githubusercontent.com/16263438/185084632-97a5b070-580d-4d57-9b02-c9d0f12bcc6b.png)
   
   Spark submit Check
   ```bash
   [spark@ambari-server spark]$ ./bin/spark-submit --class 
org.apache.spark.examples.SparkPi     --master yarn     --deploy-mode client    
 --queue default     examples/jars/spark-examples.jar     10
   SLF4J: Class path contains multiple SLF4J bindings.
   SLF4J: Found binding in 
[jar:file:/usr/lib/spark/jars/slf4j-log4j12-1.7.30.jar!/org/slf4j/impl/StaticLoggerBinder.class]
   SLF4J: Found binding in 
[jar:file:/usr/lib/hadoop/lib/slf4j-reload4j-1.7.36.jar!/org/slf4j/impl/StaticLoggerBinder.class]
   SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
explanation.
   SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
   22/08/17 09:08:03 INFO SparkContext: Running Spark version 3.2.1
   22/08/17 09:08:03 WARN NativeCodeLoader: Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
   22/08/17 09:08:03 INFO ResourceUtils: 
==============================================================
   22/08/17 09:08:03 INFO ResourceUtils: No custom resources configured for 
spark.driver.
   22/08/17 09:08:03 INFO ResourceUtils: 
==============================================================
   22/08/17 09:08:03 INFO SparkContext: Submitted application: Spark Pi
   22/08/17 09:08:03 INFO ResourceProfile: Default ResourceProfile created, 
executor resources: Map(cores -> name: cores, amount: 1, script: , vendor: , 
memory -> name: memory, amount: 1024, script: , vendor: , offHeap -> name: 
offHeap, amount: 0, script: , vendor: ), task resources: Map(cpus -> name: 
cpus, amount: 1.0)
   22/08/17 09:08:03 INFO ResourceProfile: Limiting resource is cpus at 1 tasks 
per executor
   22/08/17 09:08:03 INFO ResourceProfileManager: Added ResourceProfile id: 0
   22/08/17 09:08:03 INFO SecurityManager: Changing view acls to: spark
   22/08/17 09:08:03 INFO SecurityManager: Changing modify acls to: spark
   22/08/17 09:08:03 INFO SecurityManager: Changing view acls groups to: 
   22/08/17 09:08:03 INFO SecurityManager: Changing modify acls groups to: 
   22/08/17 09:08:03 INFO SecurityManager: SecurityManager: authentication 
disabled; ui acls disabled; users  with view permissions: Set(spark); groups 
with view permissions: Set(); users  with modify permissions: Set(spark); 
groups with modify permissions: Set()
   22/08/17 09:08:04 INFO Utils: Successfully started service 'sparkDriver' on 
port 41738.
   22/08/17 09:08:04 INFO SparkEnv: Registering MapOutputTracker
   22/08/17 09:08:04 INFO SparkEnv: Registering BlockManagerMaster
   22/08/17 09:08:04 INFO BlockManagerMasterEndpoint: Using 
org.apache.spark.storage.DefaultTopologyMapper for getting topology information
   22/08/17 09:08:04 INFO BlockManagerMasterEndpoint: 
BlockManagerMasterEndpoint up
   22/08/17 09:08:04 INFO SparkEnv: Registering BlockManagerMasterHeartbeat
   22/08/17 09:08:04 INFO DiskBlockManager: Created local directory at 
/tmp/blockmgr-89afb528-910d-47a3-92a9-d3c329250a5a
   22/08/17 09:08:04 INFO MemoryStore: MemoryStore started with capacity 366.3 
MiB
   22/08/17 09:08:04 INFO SparkEnv: Registering OutputCommitCoordinator
   22/08/17 09:08:04 INFO log: Logging initialized @3291ms to 
org.sparkproject.jetty.util.log.Slf4jLog
   22/08/17 09:08:04 INFO Server: jetty-9.4.43.v20210629; built: 
2021-06-30T11:07:22.254Z; git: 526006ecfa3af7f1a27ef3a288e2bef7ea9dd7e8; jvm 
1.8.0_332-b09
   22/08/17 09:08:04 INFO Server: Started @3449ms
   22/08/17 09:08:04 WARN Utils: Service 'SparkUI' could not bind on port 4040. 
Attempting port 4041.
   22/08/17 09:08:04 INFO AbstractConnector: Started 
ServerConnector@39aa45a1{HTTP/1.1, (http/1.1)}{0.0.0.0:4041}
   22/08/17 09:08:04 INFO Utils: Successfully started service 'SparkUI' on port 
4041.
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@14faa38c{/jobs,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@759d81f3{/jobs/json,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@5a4c638d{/jobs/job,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@1f12e153{/jobs/job/json,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@5a101b1c{/stages,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@29f0802c{/stages/json,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@60f2e0bd{/stages/stage,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@203dd56b{/stages/stage/json,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@6d64b553{/stages/pool,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@1d3e6d34{/stages/pool/json,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@26a94fa5{/storage,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@2873d672{/storage/json,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@577f9109{/storage/rdd,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@757529a4{/storage/rdd/json,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@5c41d037{/environment,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@5ec77191{/environment/json,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@1450078a{/executors,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@69c6161d{/executors/json,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@2e1792e7{/executors/threadDump,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@3eb631b8{/executors/threadDump/json,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@6bff19ff{/static,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@3b0ca5e1{/,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@54dcbb9f{/api,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@5bb8f9e2{/jobs/job/kill,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@5f78de22{/stages/stage/kill,null,AVAILABLE,@Spark}
   22/08/17 09:08:04 INFO SparkUI: Bound SparkUI to 0.0.0.0, and started at 
http://ambari-server:4041
   22/08/17 09:08:04 INFO SparkContext: Added JAR 
file:/usr/lib/spark/examples/jars/spark-examples_2.12-3.2.1.jar at 
spark://ambari-server:41738/jars/spark-examples_2.12-3.2.1.jar with timestamp 
1660727283432
   22/08/17 09:08:05 INFO FairSchedulableBuilder: Creating Fair Scheduler pools 
from file:////etc/spark/conf/spark-thrift-fairscheduler.xml
   22/08/17 09:08:05 INFO FairSchedulableBuilder: Created pool: default, 
schedulingMode: FAIR, minShare: 2, weight: 1
   22/08/17 09:08:05 INFO DefaultNoHARMFailoverProxyProvider: Connecting to 
ResourceManager at ambari-agent-02/172.18.0.4:8050
   22/08/17 09:08:06 INFO AHSProxy: Connecting to Application History server at 
localhost/127.0.0.1:10200
   22/08/17 09:08:06 INFO Client: Requesting a new application from cluster 
with 2 NodeManagers
   22/08/17 09:08:07 WARN DomainSocketFactory: The short-circuit local reads 
feature cannot be used because libhadoop cannot be loaded.
   22/08/17 09:08:07 INFO Configuration: resource-types.xml not found
   22/08/17 09:08:07 INFO ResourceUtils: Unable to find 'resource-types.xml'.
   22/08/17 09:08:07 INFO Client: Verifying our application has not requested 
more than the maximum memory capability of the cluster (15360 MB per container)
   22/08/17 09:08:07 INFO Client: Will allocate AM container, with 896 MB 
memory including 384 MB overhead
   22/08/17 09:08:07 INFO Client: Setting up container launch context for our AM
   22/08/17 09:08:07 INFO Client: Setting up the launch environment for our AM 
container
   22/08/17 09:08:07 INFO Client: Preparing resources for our AM container
   22/08/17 09:08:07 WARN Client: Neither spark.yarn.jars nor 
spark.yarn.archive is set, falling back to uploading libraries under SPARK_HOME.
   22/08/17 09:08:12 INFO Client: Uploading resource 
file:/tmp/spark-9f1844ca-32b6-49b1-a545-cfe3e9e12f40/__spark_libs__3058775863316389856.zip
 -> 
hdfs://ambari-agent-01:8020/user/spark/.sparkStaging/application_1660722608926_0006/__spark_libs__3058775863316389856.zip
   22/08/17 09:08:15 INFO Client: Uploading resource 
file:/tmp/spark-9f1844ca-32b6-49b1-a545-cfe3e9e12f40/__spark_conf__3875795998828899735.zip
 -> 
hdfs://ambari-agent-01:8020/user/spark/.sparkStaging/application_1660722608926_0006/__spark_conf__.zip
   22/08/17 09:08:16 INFO SecurityManager: Changing view acls to: spark
   22/08/17 09:08:16 INFO SecurityManager: Changing modify acls to: spark
   22/08/17 09:08:16 INFO SecurityManager: Changing view acls groups to: 
   22/08/17 09:08:16 INFO SecurityManager: Changing modify acls groups to: 
   22/08/17 09:08:16 INFO SecurityManager: SecurityManager: authentication 
disabled; ui acls disabled; users  with view permissions: Set(spark); groups 
with view permissions: Set(); users  with modify permissions: Set(spark); 
groups with modify permissions: Set()
   22/08/17 09:08:16 INFO Client: Submitting application 
application_1660722608926_0006 to ResourceManager
   22/08/17 09:08:16 INFO YarnClientImpl: Submitted application 
application_1660722608926_0006
   22/08/17 09:08:17 INFO Client: Application report for 
application_1660722608926_0006 (state: ACCEPTED)
   22/08/17 09:08:17 INFO Client: 
         client token: N/A
         diagnostics: AM container is launched, waiting for AM container to 
Register with RM
         ApplicationMaster host: N/A
         ApplicationMaster RPC port: -1
         queue: default
         start time: 1660727296266
         final status: UNDEFINED
         tracking URL: 
http://ambari-agent-02:8088/proxy/application_1660722608926_0006/
         user: spark
   22/08/17 09:08:18 INFO Client: Application report for 
application_1660722608926_0006 (state: ACCEPTED)
   22/08/17 09:08:19 INFO Client: Application report for 
application_1660722608926_0006 (state: ACCEPTED)
   22/08/17 09:08:20 INFO Client: Application report for 
application_1660722608926_0006 (state: ACCEPTED)
   22/08/17 09:08:21 INFO Client: Application report for 
application_1660722608926_0006 (state: ACCEPTED)
   22/08/17 09:08:22 INFO Client: Application report for 
application_1660722608926_0006 (state: ACCEPTED)
   22/08/17 09:08:23 INFO Client: Application report for 
application_1660722608926_0006 (state: RUNNING)
   22/08/17 09:08:23 INFO Client: 
         client token: N/A
         diagnostics: N/A
         ApplicationMaster host: 172.18.0.4
         ApplicationMaster RPC port: -1
         queue: default
         start time: 1660727296266
         final status: UNDEFINED
         tracking URL: 
http://ambari-agent-02:8088/proxy/application_1660722608926_0006/
         user: spark
   22/08/17 09:08:23 INFO YarnClientSchedulerBackend: Application 
application_1660722608926_0006 has started running.
   22/08/17 09:08:23 INFO Utils: Successfully started service 
'org.apache.spark.network.netty.NettyBlockTransferService' on port 35127.
   22/08/17 09:08:23 INFO NettyBlockTransferService: Server created on 
ambari-server:35127
   22/08/17 09:08:23 INFO BlockManager: Using 
org.apache.spark.storage.RandomBlockReplicationPolicy for block replication 
policy
   22/08/17 09:08:23 INFO BlockManagerMaster: Registering BlockManager 
BlockManagerId(driver, ambari-server, 35127, None)
   22/08/17 09:08:23 INFO BlockManagerMasterEndpoint: Registering block manager 
ambari-server:35127 with 366.3 MiB RAM, BlockManagerId(driver, ambari-server, 
35127, None)
   22/08/17 09:08:23 INFO BlockManagerMaster: Registered BlockManager 
BlockManagerId(driver, ambari-server, 35127, None)
   22/08/17 09:08:23 INFO BlockManager: Initialized BlockManager: 
BlockManagerId(driver, ambari-server, 35127, None)
   22/08/17 09:08:23 INFO ContextHandler: Started 
o.s.j.s.ServletContextHandler@388623ad{/metrics/json,null,AVAILABLE,@Spark}
   22/08/17 09:08:23 INFO SingleEventLogFileWriter: Logging events to 
hdfs:/spark-history/application_1660722608926_0006.inprogress
   22/08/17 09:08:23 INFO YarnClientSchedulerBackend: Add WebUI Filter. 
org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, Map(PROXY_HOSTS -> 
ambari-agent-02, PROXY_URI_BASES -> 
http://ambari-agent-02:8088/proxy/application_1660722608926_0006), 
/proxy/application_1660722608926_0006
   22/08/17 09:08:24 INFO YarnSchedulerBackend$YarnSchedulerEndpoint: 
ApplicationMaster registered as NettyRpcEndpointRef(spark-client://YarnAM)
   22/08/17 09:08:33 INFO YarnSchedulerBackend$YarnDriverEndpoint: Registered 
executor NettyRpcEndpointRef(spark-client://Executor) (172.18.0.2:38336) with 
ID 1,  ResourceProfileId 0
   22/08/17 09:08:33 INFO YarnSchedulerBackend$YarnDriverEndpoint: Registered 
executor NettyRpcEndpointRef(spark-client://Executor) (172.18.0.2:38334) with 
ID 2,  ResourceProfileId 0
   22/08/17 09:08:33 INFO YarnClientSchedulerBackend: SchedulerBackend is ready 
for scheduling beginning after reached minRegisteredResourcesRatio: 0.8
   22/08/17 09:08:33 INFO BlockManagerMasterEndpoint: Registering block manager 
ambari-server:38665 with 366.3 MiB RAM, BlockManagerId(1, ambari-server, 38665, 
None)
   22/08/17 09:08:33 INFO BlockManagerMasterEndpoint: Registering block manager 
ambari-server:42731 with 366.3 MiB RAM, BlockManagerId(2, ambari-server, 42731, 
None)
   22/08/17 09:08:34 INFO SparkContext: Starting job: reduce at SparkPi.scala:38
   22/08/17 09:08:34 INFO DAGScheduler: Got job 0 (reduce at SparkPi.scala:38) 
with 10 output partitions
   22/08/17 09:08:34 INFO DAGScheduler: Final stage: ResultStage 0 (reduce at 
SparkPi.scala:38)
   22/08/17 09:08:34 INFO DAGScheduler: Parents of final stage: List()
   22/08/17 09:08:34 INFO DAGScheduler: Missing parents: List()
   22/08/17 09:08:34 INFO DAGScheduler: Submitting ResultStage 0 
(MapPartitionsRDD[1] at map at SparkPi.scala:34), which has no missing parents
   22/08/17 09:08:34 INFO MemoryStore: Block broadcast_0 stored as values in 
memory (estimated size 4.0 KiB, free 366.3 MiB)
   22/08/17 09:08:34 INFO MemoryStore: Block broadcast_0_piece0 stored as bytes 
in memory (estimated size 2.3 KiB, free 366.3 MiB)
   22/08/17 09:08:34 INFO BlockManagerInfo: Added broadcast_0_piece0 in memory 
on ambari-server:35127 (size: 2.3 KiB, free: 366.3 MiB)
   22/08/17 09:08:34 INFO SparkContext: Created broadcast 0 from broadcast at 
DAGScheduler.scala:1478
   22/08/17 09:08:34 INFO DAGScheduler: Submitting 10 missing tasks from 
ResultStage 0 (MapPartitionsRDD[1] at map at SparkPi.scala:34) (first 15 tasks 
are for partitions Vector(0, 1, 2, 3, 4, 5, 6, 7, 8, 9))
   22/08/17 09:08:34 INFO YarnScheduler: Adding task set 0.0 with 10 tasks 
resource profile 0
   22/08/17 09:08:34 INFO FairSchedulableBuilder: Added task set TaskSet_0.0 
tasks to pool default
   22/08/17 09:08:34 INFO TaskSetManager: Starting task 0.0 in stage 0.0 (TID 
0) (ambari-server, executor 1, partition 0, PROCESS_LOCAL, 4589 bytes) 
taskResourceAssignments Map()
   22/08/17 09:08:34 INFO TaskSetManager: Starting task 1.0 in stage 0.0 (TID 
1) (ambari-server, executor 2, partition 1, PROCESS_LOCAL, 4589 bytes) 
taskResourceAssignments Map()
   22/08/17 09:08:34 INFO BlockManagerInfo: Added broadcast_0_piece0 in memory 
on ambari-server:38665 (size: 2.3 KiB, free: 366.3 MiB)
   22/08/17 09:08:34 INFO BlockManagerInfo: Added broadcast_0_piece0 in memory 
on ambari-server:42731 (size: 2.3 KiB, free: 366.3 MiB)
   22/08/17 09:08:35 INFO TaskSetManager: Starting task 2.0 in stage 0.0 (TID 
2) (ambari-server, executor 2, partition 2, PROCESS_LOCAL, 4589 bytes) 
taskResourceAssignments Map()
   22/08/17 09:08:35 INFO TaskSetManager: Finished task 1.0 in stage 0.0 (TID 
1) in 965 ms on ambari-server (executor 2) (1/10)
   22/08/17 09:08:35 INFO TaskSetManager: Starting task 3.0 in stage 0.0 (TID 
3) (ambari-server, executor 1, partition 3, PROCESS_LOCAL, 4589 bytes) 
taskResourceAssignments Map()
   22/08/17 09:08:35 INFO TaskSetManager: Finished task 0.0 in stage 0.0 (TID 
0) in 1007 ms on ambari-server (executor 1) (2/10)
   22/08/17 09:08:35 INFO TaskSetManager: Starting task 4.0 in stage 0.0 (TID 
4) (ambari-server, executor 2, partition 4, PROCESS_LOCAL, 4589 bytes) 
taskResourceAssignments Map()
   22/08/17 09:08:35 INFO TaskSetManager: Finished task 2.0 in stage 0.0 (TID 
2) in 37 ms on ambari-server (executor 2) (3/10)
   22/08/17 09:08:35 INFO TaskSetManager: Starting task 5.0 in stage 0.0 (TID 
5) (ambari-server, executor 1, partition 5, PROCESS_LOCAL, 4589 bytes) 
taskResourceAssignments Map()
   22/08/17 09:08:35 INFO TaskSetManager: Starting task 6.0 in stage 0.0 (TID 
6) (ambari-server, executor 2, partition 6, PROCESS_LOCAL, 4589 bytes) 
taskResourceAssignments Map()
   22/08/17 09:08:35 INFO TaskSetManager: Finished task 3.0 in stage 0.0 (TID 
3) in 48 ms on ambari-server (executor 1) (4/10)
   22/08/17 09:08:35 INFO TaskSetManager: Finished task 4.0 in stage 0.0 (TID 
4) in 32 ms on ambari-server (executor 2) (5/10)
   22/08/17 09:08:35 INFO TaskSetManager: Starting task 7.0 in stage 0.0 (TID 
7) (ambari-server, executor 1, partition 7, PROCESS_LOCAL, 4589 bytes) 
taskResourceAssignments Map()
   22/08/17 09:08:35 INFO TaskSetManager: Finished task 5.0 in stage 0.0 (TID 
5) in 29 ms on ambari-server (executor 1) (6/10)
   22/08/17 09:08:35 INFO TaskSetManager: Starting task 8.0 in stage 0.0 (TID 
8) (ambari-server, executor 2, partition 8, PROCESS_LOCAL, 4589 bytes) 
taskResourceAssignments Map()
   22/08/17 09:08:35 INFO TaskSetManager: Finished task 6.0 in stage 0.0 (TID 
6) in 29 ms on ambari-server (executor 2) (7/10)
   22/08/17 09:08:35 INFO TaskSetManager: Starting task 9.0 in stage 0.0 (TID 
9) (ambari-server, executor 2, partition 9, PROCESS_LOCAL, 4589 bytes) 
taskResourceAssignments Map()
   22/08/17 09:08:35 INFO TaskSetManager: Finished task 8.0 in stage 0.0 (TID 
8) in 26 ms on ambari-server (executor 2) (8/10)
   22/08/17 09:08:35 INFO TaskSetManager: Finished task 7.0 in stage 0.0 (TID 
7) in 41 ms on ambari-server (executor 1) (9/10)
   22/08/17 09:08:35 INFO TaskSetManager: Finished task 9.0 in stage 0.0 (TID 
9) in 25 ms on ambari-server (executor 2) (10/10)
   22/08/17 09:08:35 INFO YarnScheduler: Removed TaskSet 0.0, whose tasks have 
all completed, from pool default
   22/08/17 09:08:35 INFO DAGScheduler: ResultStage 0 (reduce at 
SparkPi.scala:38) finished in 1.348 s
   22/08/17 09:08:35 INFO DAGScheduler: Job 0 is finished. Cancelling potential 
speculative or zombie tasks for this job
   22/08/17 09:08:35 INFO YarnScheduler: Killing all running tasks in stage 0: 
Stage finished
   22/08/17 09:08:35 INFO DAGScheduler: Job 0 finished: reduce at 
SparkPi.scala:38, took 1.423045 s
   Pi is roughly 3.1396031396031394
   22/08/17 09:08:35 INFO AbstractConnector: Stopped Spark@39aa45a1{HTTP/1.1, 
(http/1.1)}{0.0.0.0:4041}
   22/08/17 09:08:35 INFO SparkUI: Stopped Spark web UI at 
http://ambari-server:4041
   22/08/17 09:08:35 INFO YarnClientSchedulerBackend: Interrupting monitor 
thread
   22/08/17 09:08:35 INFO YarnClientSchedulerBackend: Shutting down all 
executors
   22/08/17 09:08:35 INFO YarnSchedulerBackend$YarnDriverEndpoint: Asking each 
executor to shut down
   22/08/17 09:08:35 INFO YarnClientSchedulerBackend: YARN client scheduler 
backend Stopped
   22/08/17 09:08:35 INFO MapOutputTrackerMasterEndpoint: 
MapOutputTrackerMasterEndpoint stopped!
   22/08/17 09:08:35 INFO MemoryStore: MemoryStore cleared
   22/08/17 09:08:35 INFO BlockManager: BlockManager stopped
   22/08/17 09:08:35 INFO BlockManagerMaster: BlockManagerMaster stopped
   22/08/17 09:08:35 INFO 
OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: 
OutputCommitCoordinator stopped!
   22/08/17 09:08:35 INFO SparkContext: Successfully stopped SparkContext
   22/08/17 09:08:35 INFO ShutdownHookManager: Shutdown hook called
   22/08/17 09:08:35 INFO ShutdownHookManager: Deleting directory 
/tmp/spark-2e1ee283-8ac9-4187-9cb4-3e8b50e79617
   22/08/17 09:08:35 INFO ShutdownHookManager: Deleting directory 
/tmp/spark-9f1844ca-32b6-49b1-a545-cfe3e9e12f40
   ```
   
   Spark Sql Check
   
![spark_check_sparksql_20220817173113](https://user-images.githubusercontent.com/16263438/185085522-228a3b3b-a5c5-4c9f-9b42-20c861eb5109.png)
   
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to