How to verify that the worker is connected to master in CDH5.4

2015-07-07 Thread Ashish Dutt
Hi,
I have CDH 5.4 installed on a linux server. It has 1 cluster in which spark
is deployed as a history server.
I am trying to connect my laptop to the spark history server.
When I run spark-shell master ip: port number I get the following output
How can I verify that the worker is connected to the master?

Thanks,
Ashish


Re: How to verify that the worker is connected to master in CDH5.4

2015-07-07 Thread Guru Medasani
Hi Ashish,

Are you running Spark-on-YARN on the cluster with an instance of Spark History 
server? 

Also if you are using Cloudera Manager and using Spark on YARN, spark on yarn 
service has a link for the history server web UI. 

Can you paste the command and the output you are seeing in the thread?

Guru Medasani
gdm...@gmail.com



 On Jul 7, 2015, at 10:42 PM, Ashish Dutt ashish.du...@gmail.com wrote:
 
 Hi,
 I have CDH 5.4 installed on a linux server. It has 1 cluster in which spark 
 is deployed as a history server.
 I am trying to connect my laptop to the spark history server.
 When I run spark-shell master ip: port number I get the following output
 How can I verify that the worker is connected to the master?
 
 Thanks,
 Ashish
  


-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org



Re: How to verify that the worker is connected to master in CDH5.4

2015-07-07 Thread Ashish Dutt
Thank you Ayan for your response.. But I have just realised that the Spark
is configured to be a history server.
Please, can somebody suggest to me how can I convert Spark history server
to be a Master server?

Thank you

Sincerely,
Ashish Dutt


On Wed, Jul 8, 2015 at 12:28 PM, ayan guha guha.a...@gmail.com wrote:

 On UI?

 Master: http://masterip:8080
 Worker: http://workerIp:8081

 On Wed, Jul 8, 2015 at 1:42 PM, Ashish Dutt ashish.du...@gmail.com
 wrote:

 Hi,
 I have CDH 5.4 installed on a linux server. It has 1 cluster in which
 spark is deployed as a history server.
 I am trying to connect my laptop to the spark history server.
 When I run spark-shell master ip: port number I get the following output
 How can I verify that the worker is connected to the master?

 Thanks,
 Ashish





 --
 Best Regards,
 Ayan Guha



Re: How to verify that the worker is connected to master in CDH5.4

2015-07-07 Thread Ashish Dutt
Hello Guru,
Thank you for your quick response.
 This is what i get when I try executing spark-shell master ip:port number

C:\spark-1.4.0\binspark-shell master IP:18088
log4j:WARN No appenders could be found for logger
(org.apache.hadoop.metrics2.lib.MutableMetricsFactory).
log4j:WARN Please initialize the log4j system properly.
log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for
more info.
Using Spark's default log4j profile:
org/apache/spark/log4j-defaults.properties
15/07/08 11:28:35 INFO SecurityManager: Changing view acls to: Ashish Dutt
15/07/08 11:28:35 INFO SecurityManager: Changing modify acls to: Ashish Dutt
15/07/08 11:28:35 INFO SecurityManager: SecurityManager: authentication
disabled; ui acls disabled; users with view permissions: Set
(Ashish Dutt); users with modify permissions: Set(Ashish Dutt)
15/07/08 11:28:35 INFO HttpServer: Starting HTTP Server
15/07/08 11:28:35 INFO Utils: Successfully started service 'HTTP class
server' on port 52767.
Welcome to
    __
 / __/__  ___ _/ /__
_\ \/ _ \/ _ `/ __/  '_/
   /___/ .__/\_,_/_/ /_/\_\   version 1.4.0
  /_/

Using Scala version 2.10.4 (Java HotSpot(TM) 64-Bit Server VM, Java
1.7.0_79)
Type in expressions to have them evaluated.
Type :help for more information.
15/07/08 11:28:39 INFO SparkContext: Running Spark version 1.4.0
15/07/08 11:28:39 INFO SecurityManager: Changing view acls to: Ashish Dutt
15/07/08 11:28:39 INFO SecurityManager: Changing modify acls to: Ashish Dutt
15/07/08 11:28:39 INFO SecurityManager: SecurityManager: authentication
disabled; ui acls disabled; users with view permissions: Set
(Ashish Dutt); users with modify permissions: Set(Ashish Dutt)
15/07/08 11:28:40 INFO Slf4jLogger: Slf4jLogger started
15/07/08 11:28:40 INFO Remoting: Starting remoting
15/07/08 11:28:40 INFO Remoting: Remoting started; listening on addresses
:[akka.tcp://sparkDriver@10.228.208.74:52780]
15/07/08 11:28:40 INFO Utils: Successfully started service 'sparkDriver' on
port 52780.
15/07/08 11:28:40 INFO SparkEnv: Registering MapOutputTracker
15/07/08 11:28:40 INFO SparkEnv: Registering BlockManagerMaster
15/07/08 11:28:40 INFO DiskBlockManager: Created local directory at
C:\Users\Ashish Dutt\AppData\Local\Temp\spark-80c4f1fe-37de-4aef
-9063-cae29c488382\blockmgr-a967422b-05e8-4fc1-b60b-facc7dbd4414
15/07/08 11:28:40 INFO MemoryStore: MemoryStore started with capacity 265.4
MB
15/07/08 11:28:40 INFO HttpFileServer: HTTP File server directory is
C:\Users\Ashish Dutt\AppData\Local\Temp\spark-80c4f1fe-37de-4ae
f-9063-cae29c488382\httpd-928f4485-ea08-4749-a478-59708db0fefa
15/07/08 11:28:40 INFO HttpServer: Starting HTTP Server
15/07/08 11:28:40 INFO Utils: Successfully started service 'HTTP file
server' on port 52781.
15/07/08 11:28:40 INFO SparkEnv: Registering OutputCommitCoordinator
15/07/08 11:28:40 INFO Utils: Successfully started service 'SparkUI' on
port 4040.
15/07/08 11:28:40 INFO SparkUI: Started SparkUI at http://10.228.208.74:4040
15/07/08 11:28:40 INFO Executor: Starting executor ID driver on host
localhost
15/07/08 11:28:41 INFO Executor: Using REPL class URI:
http://10.228.208.74:52767
15/07/08 11:28:41 INFO Utils: Successfully started service
'org.apache.spark.network.netty.NettyBlockTransferService' on port 52800.

15/07/08 11:28:41 INFO NettyBlockTransferService: Server created on 52800
15/07/08 11:28:41 INFO BlockManagerMaster: Trying to register BlockManager
15/07/08 11:28:41 INFO BlockManagerMasterEndpoint: Registering block
manager localhost:52800 with 265.4 MB RAM, BlockManagerId(drive
r, localhost, 52800)
15/07/08 11:28:41 INFO BlockManagerMaster: Registered BlockManager

15/07/08 11:28:41 INFO SparkILoop: Created spark context..
Spark context available as sc.
15/07/08 11:28:41 INFO HiveContext: Initializing execution hive, version
0.13.1
15/07/08 11:28:42 INFO HiveMetaStore: 0: Opening raw store with
implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
15/07/08 11:28:42 INFO ObjectStore: ObjectStore, initialize called
15/07/08 11:28:42 INFO Persistence: Property datanucleus.cache.level2
unknown - will be ignored
15/07/08 11:28:42 INFO Persistence: Property
hive.metastore.integral.jdo.pushdown unknown - will be ignored
15/07/08 11:28:42 WARN Connection: BoneCP specified but not present in
CLASSPATH (or one of dependencies)
15/07/08 11:28:42 WARN Connection: BoneCP specified but not present in
CLASSPATH (or one of dependencies)
15/07/08 11:28:52 INFO ObjectStore: Setting MetaStore object pin classes
with hive.metastore.cache.pinobjtypes=Table,StorageDescrip
tor,SerDeInfo,Partition,Database,Type,FieldSchema,Order
15/07/08 11:28:52 INFO MetaStoreDirectSql: MySQL check failed, assuming we
are not on mysql: Lexical error at line 1, column 5.  Enc
ountered: @ (64), after : .
15/07/08 11:28:53 INFO Datastore: The class
org.apache.hadoop.hive.metastore.model.MFieldSchema is tagged as
embedded-only so do
es not have its own datastore table.
15/07/08 11:28:53 INFO 

Re: How to verify that the worker is connected to master in CDH5.4

2015-07-07 Thread Guru Medasani
Hi Ashish,

If you are not using Spark on YARN and instead using Spark Standalone, you 
don’t need Spark history server. More on the Web Interfaces is provided in the 
following link. Since are using standalone mode, you should be able to access 
the web UI for the master and workers at ports that Ayan provided in early 
email.

Master: http://masterip:8080 
Worker: http://workerIp:8081

https://spark.apache.org/docs/latest/monitoring.html 
https://spark.apache.org/docs/latest/monitoring.html

If you are using Spark on YARN, spark history server is configured to run on 
port 18080 by default on the server where Spark history server is running.

Guru Medasani
gdm...@gmail.com



 On Jul 8, 2015, at 12:01 AM, Ashish Dutt ashish.du...@gmail.com wrote:
 
 Hello Guru,
 Thank you for your quick response. 
  This is what i get when I try executing spark-shell master ip:port number
 
 C:\spark-1.4.0\binspark-shell master IP:18088
 log4j:WARN No appenders could be found for logger 
 (org.apache.hadoop.metrics2.lib.MutableMetricsFactory).
 log4j:WARN Please initialize the log4j system properly.
 log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig 
 http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info.
 Using Spark's default log4j profile: 
 org/apache/spark/log4j-defaults.properties
 15/07/08 11:28:35 INFO SecurityManager: Changing view acls to: Ashish Dutt
 15/07/08 11:28:35 INFO SecurityManager: Changing modify acls to: Ashish Dutt
 15/07/08 11:28:35 INFO SecurityManager: SecurityManager: authentication 
 disabled; ui acls disabled; users with view permissions: Set
 (Ashish Dutt); users with modify permissions: Set(Ashish Dutt)
 15/07/08 11:28:35 INFO HttpServer: Starting HTTP Server
 15/07/08 11:28:35 INFO Utils: Successfully started service 'HTTP class 
 server' on port 52767.
 Welcome to
     __
  / __/__  ___ _/ /__
 _\ \/ _ \/ _ `/ __/  '_/
/___/ .__/\_,_/_/ /_/\_\   version 1.4.0
   /_/
 
 Using Scala version 2.10.4 (Java HotSpot(TM) 64-Bit Server VM, Java 1.7.0_79)
 Type in expressions to have them evaluated.
 Type :help for more information.
 15/07/08 11:28:39 INFO SparkContext: Running Spark version 1.4.0
 15/07/08 11:28:39 INFO SecurityManager: Changing view acls to: Ashish Dutt
 15/07/08 11:28:39 INFO SecurityManager: Changing modify acls to: Ashish Dutt
 15/07/08 11:28:39 INFO SecurityManager: SecurityManager: authentication 
 disabled; ui acls disabled; users with view permissions: Set
 (Ashish Dutt); users with modify permissions: Set(Ashish Dutt)
 15/07/08 11:28:40 INFO Slf4jLogger: Slf4jLogger started
 15/07/08 11:28:40 INFO Remoting: Starting remoting
 15/07/08 11:28:40 INFO Remoting: Remoting started; listening on addresses 
 :[akka.tcp://sparkDriver@10.228.208.74:52780 
 http://sparkDriver@10.228.208.74:52780/]
 15/07/08 11:28:40 INFO Utils: Successfully started service 'sparkDriver' on 
 port 52780.
 15/07/08 11:28:40 INFO SparkEnv: Registering MapOutputTracker
 15/07/08 11:28:40 INFO SparkEnv: Registering BlockManagerMaster
 15/07/08 11:28:40 INFO DiskBlockManager: Created local directory at 
 C:\Users\Ashish Dutt\AppData\Local\Temp\spark-80c4f1fe-37de-4aef
 -9063-cae29c488382\blockmgr-a967422b-05e8-4fc1-b60b-facc7dbd4414
 15/07/08 11:28:40 INFO MemoryStore: MemoryStore started with capacity 265.4 MB
 15/07/08 11:28:40 INFO HttpFileServer: HTTP File server directory is 
 C:\Users\Ashish Dutt\AppData\Local\Temp\spark-80c4f1fe-37de-4ae
 f-9063-cae29c488382\httpd-928f4485-ea08-4749-a478-59708db0fefa
 15/07/08 11:28:40 INFO HttpServer: Starting HTTP Server
 15/07/08 11:28:40 INFO Utils: Successfully started service 'HTTP file server' 
 on port 52781.
 15/07/08 11:28:40 INFO SparkEnv: Registering OutputCommitCoordinator
 15/07/08 11:28:40 INFO Utils: Successfully started service 'SparkUI' on port 
 4040.
 15/07/08 11:28:40 INFO SparkUI: Started SparkUI at http://10.228.208.74:4040 
 http://10.228.208.74:4040/
 15/07/08 11:28:40 INFO Executor: Starting executor ID driver on host localhost
 15/07/08 11:28:41 INFO Executor: Using REPL class URI: 
 http://10.228.208.74:52767 http://10.228.208.74:52767/
 15/07/08 11:28:41 INFO Utils: Successfully started service 
 'org.apache.spark.network.netty.NettyBlockTransferService' on port 52800.
 
 15/07/08 11:28:41 INFO NettyBlockTransferService: Server created on 52800
 15/07/08 11:28:41 INFO BlockManagerMaster: Trying to register BlockManager
 15/07/08 11:28:41 INFO BlockManagerMasterEndpoint: Registering block manager 
 localhost:52800 with 265.4 MB RAM, BlockManagerId(drive
 r, localhost, 52800)
 15/07/08 11:28:41 INFO BlockManagerMaster: Registered BlockManager
 
 15/07/08 11:28:41 INFO SparkILoop: Created spark context..
 Spark context available as sc.
 15/07/08 11:28:41 INFO HiveContext: Initializing execution hive, version 
 0.13.1
 15/07/08 11:28:42 INFO HiveMetaStore: 0: Opening raw store with implemenation 
 class:org.apache.hadoop.hive.metastore.ObjectStore
 15/07/08 11:28:42 

Re: How to verify that the worker is connected to master in CDH5.4

2015-07-07 Thread ayan guha
On UI?

Master: http://masterip:8080
Worker: http://workerIp:8081

On Wed, Jul 8, 2015 at 1:42 PM, Ashish Dutt ashish.du...@gmail.com wrote:

 Hi,
 I have CDH 5.4 installed on a linux server. It has 1 cluster in which
 spark is deployed as a history server.
 I am trying to connect my laptop to the spark history server.
 When I run spark-shell master ip: port number I get the following output
 How can I verify that the worker is connected to the master?

 Thanks,
 Ashish





-- 
Best Regards,
Ayan Guha


Re: How to verify that the worker is connected to master in CDH5.4

2015-07-07 Thread Ashish Dutt
Hello Guru,

Many thanks for your reply.
I am new to this who thing. So pardon me for my naiivety at times.
I am not sure if I am using Spark standalone or Spark on Yarn because when
I check the port number of Spark it shows it as 18088 and like you have
mentioned maybe it is then Spark on Yarn.
All I want for now is how to connect my laptop to the spark cluster machine
using either pyspark or SparkR. (I have python 2.7)
On my laptop I am using winutils in place of hadoop and have spark 1.4
installed
Thank you

Sincerely,
Ashish Dutt
PhD Candidate
Department of Information Systems
University of Malaya, Lembah Pantai,
50603 Kuala Lumpur, Malaysia

On Wed, Jul 8, 2015 at 1:13 PM, Guru Medasani gdm...@gmail.com wrote:

 Hi Ashish,

 If you are not using Spark on YARN and instead using Spark Standalone, you
 don’t need Spark history server. More on the Web Interfaces is provided in
 the following link. Since are using standalone mode, you should be able to
 access the web UI for the master and workers at ports that Ayan provided in
 early email.

 Master: http://masterip:8080
 Worker: http://workerIp:8081

 https://spark.apache.org/docs/latest/monitoring.html

 If you are using Spark on YARN, spark history server is configured to run
 on port 18080 by default on the server where Spark history server is
 running.

 Guru Medasani
 gdm...@gmail.com



 On Jul 8, 2015, at 12:01 AM, Ashish Dutt ashish.du...@gmail.com wrote:

 Hello Guru,
 Thank you for your quick response.
  This is what i get when I try executing spark-shell master ip:port
 number

 C:\spark-1.4.0\binspark-shell master IP:18088
 log4j:WARN No appenders could be found for logger
 (org.apache.hadoop.metrics2.lib.MutableMetricsFactory).
 log4j:WARN Please initialize the log4j system properly.
 log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for
 more info.
 Using Spark's default log4j profile:
 org/apache/spark/log4j-defaults.properties
 15/07/08 11:28:35 INFO SecurityManager: Changing view acls to: Ashish Dutt
 15/07/08 11:28:35 INFO SecurityManager: Changing modify acls to: Ashish
 Dutt
 15/07/08 11:28:35 INFO SecurityManager: SecurityManager: authentication
 disabled; ui acls disabled; users with view permissions: Set
 (Ashish Dutt); users with modify permissions: Set(Ashish Dutt)
 15/07/08 11:28:35 INFO HttpServer: Starting HTTP Server
 15/07/08 11:28:35 INFO Utils: Successfully started service 'HTTP class
 server' on port 52767.
 Welcome to
     __
  / __/__  ___ _/ /__
 _\ \/ _ \/ _ `/ __/  '_/
/___/ .__/\_,_/_/ /_/\_\   version 1.4.0
   /_/

 Using Scala version 2.10.4 (Java HotSpot(TM) 64-Bit Server VM, Java
 1.7.0_79)
 Type in expressions to have them evaluated.
 Type :help for more information.
 15/07/08 11:28:39 INFO SparkContext: Running Spark version 1.4.0
 15/07/08 11:28:39 INFO SecurityManager: Changing view acls to: Ashish Dutt
 15/07/08 11:28:39 INFO SecurityManager: Changing modify acls to: Ashish
 Dutt
 15/07/08 11:28:39 INFO SecurityManager: SecurityManager: authentication
 disabled; ui acls disabled; users with view permissions: Set
 (Ashish Dutt); users with modify permissions: Set(Ashish Dutt)
 15/07/08 11:28:40 INFO Slf4jLogger: Slf4jLogger started
 15/07/08 11:28:40 INFO Remoting: Starting remoting
 15/07/08 11:28:40 INFO Remoting: Remoting started; listening on addresses
 :[akka.tcp://sparkDriver@10.228.208.74:52780]
 15/07/08 11:28:40 INFO Utils: Successfully started service 'sparkDriver'
 on port 52780.
 15/07/08 11:28:40 INFO SparkEnv: Registering MapOutputTracker
 15/07/08 11:28:40 INFO SparkEnv: Registering BlockManagerMaster
 15/07/08 11:28:40 INFO DiskBlockManager: Created local directory at
 C:\Users\Ashish Dutt\AppData\Local\Temp\spark-80c4f1fe-37de-4aef
 -9063-cae29c488382\blockmgr-a967422b-05e8-4fc1-b60b-facc7dbd4414
 15/07/08 11:28:40 INFO MemoryStore: MemoryStore started with capacity
 265.4 MB
 15/07/08 11:28:40 INFO HttpFileServer: HTTP File server directory is
 C:\Users\Ashish Dutt\AppData\Local\Temp\spark-80c4f1fe-37de-4ae
 f-9063-cae29c488382\httpd-928f4485-ea08-4749-a478-59708db0fefa
 15/07/08 11:28:40 INFO HttpServer: Starting HTTP Server
 15/07/08 11:28:40 INFO Utils: Successfully started service 'HTTP file
 server' on port 52781.
 15/07/08 11:28:40 INFO SparkEnv: Registering OutputCommitCoordinator
 15/07/08 11:28:40 INFO Utils: Successfully started service 'SparkUI' on
 port 4040.
 15/07/08 11:28:40 INFO SparkUI: Started SparkUI at
 http://10.228.208.74:4040
 15/07/08 11:28:40 INFO Executor: Starting executor ID driver on host
 localhost
 15/07/08 11:28:41 INFO Executor: Using REPL class URI:
 http://10.228.208.74:52767
 15/07/08 11:28:41 INFO Utils: Successfully started service
 'org.apache.spark.network.netty.NettyBlockTransferService' on port 52800.

 15/07/08 11:28:41 INFO NettyBlockTransferService: Server created on 52800
 15/07/08 11:28:41 INFO BlockManagerMaster: Trying to register BlockManager
 15/07/08 11:28:41 INFO