Hi Nick,
there's the Spark master defined in conf/spark-defaults.conf and the -h
option that you can provide to sbin/start-master.sh script.
Did you try:
sbin/start-master.sh -h xxx.xxx.xxx.xxx
and then use the IP when you start the slaves:
sbin/start-slave.sh spark://xxx.xxx.xxx.xxx.7077
?
Regards
JB
On 10/16/2015 06:01 PM, Nicholas Chammas wrote:
I'd look into tracing a possible bug here, but I'm not sure where to
look. Searching the codebase for `SPARK_MASTER_IP`, amazingly, does not
show it being used in any place directly by Spark
<https://github.com/apache/spark/search?utf8=%E2%9C%93&q=SPARK_MASTER_IP>.
Clearly, Spark is using this environment variable (otherwise I wouldn't
see the behavior described in my first email), but I can't see where.
Can someone give me a pointer?
Nick
On Thu, Oct 15, 2015 at 12:37 AM Ted Yu <yuzhih...@gmail.com
<mailto:yuzhih...@gmail.com>> wrote:
Some old bits:
http://stackoverflow.com/questions/28162991/cant-run-spark-1-2-in-standalone-mode-on-mac
http://stackoverflow.com/questions/29412157/passing-hostname-to-netty
FYI
On Wed, Oct 14, 2015 at 7:10 PM, Nicholas Chammas
<nicholas.cham...@gmail.com <mailto:nicholas.cham...@gmail.com>> wrote:
I’m setting the Spark master address via the |SPARK_MASTER_IP|
environment variable in |spark-env.sh|, like spark-ec2 does
<https://github.com/amplab/spark-ec2/blob/a990752575cd8b0ab25731d7820a55c714798ec3/templates/root/spark/conf/spark-env.sh#L13>.
The funny thing is that Spark seems to accept this only if the
value of |SPARK_MASTER_IP| is a DNS name and not an IP address.
When I provide an IP address, I get errors in the log when
starting the master:
|15/10/15 01:47:31 ERROR NettyTransport: failed to bind to
/54.210.XX.XX:7077, shutting down Netty transport |
(XX is my redaction of the full IP address.)
Am I misunderstanding something about how to use this
environment variable?
The spark-env.sh template indicates that either an IP address or
a hostname should work
<https://github.com/apache/spark/blob/4ace4f8a9c91beb21a0077e12b75637a4560a542/conf/spark-env.sh.template#L49>,
but my testing shows that only hostnames work.
Nick
--
Jean-Baptiste Onofré
jbono...@apache.org
http://blog.nanthrax.net
Talend - http://www.talend.com
---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
For additional commands, e-mail: dev-h...@spark.apache.org