Try using the --ip parameter
<http://spark.apache.org/docs/latest/spark-standalone.html#starting-a-cluster-manually>
while starting the worker. like:

spark-1.0.1/bin/spark-class org.apache.spark.deploy.worker.Worker --ip
1.2.3.4 spark://1.2.3.4:7077

Thanks
Best Regards

On Fri, Oct 24, 2014 at 12:34 PM, Theodore Si <sjyz...@gmail.com> wrote:

>  Hi all,
>
> I have two network interface card on one node, one is a Eithernet card,
> the other Infiniband HCA.
> The master has two IP addresses, lets say 1.2.3.4 (for Eithernet card) and
> 2.3.4.5 (for HCA).
> I can start the master by
> export SPARK_MASTER_IP='1.2.3.4';sbin/start-master.sh
> to let master listen on 1.2.3.4:7077
>
> But when I connect the worker to the master by using
> spark-1.0.1/bin/spark-class org.apache.spark.deploy.worker.Worker spark://
> 1.2.3.4:7077
>
> I will get errors, since it is using its HCA card. How can I let the
> worker use its Eithernet card?
>
> Thanks
>

Reply via email to