There is no binding issue here. Spark picks the right ip 10.211.55.3 for you. The printed message is just an indication. However I have no idea why spark-shell hangs or stops.
发自我的 iPhone > 在 2015年1月14日,上午5:10,Akhil Das <ak...@sigmoidanalytics.com> 写道: > > It just a binding issue with the hostnames in your /etc/hosts file. You can > set SPARK_LOCAL_IP and SPARK_MASTER_IP in your conf/spark-env.sh file and > restart your cluster. (in that case the spark://myworkstation:7077 will > change to the ip address that you provided eg: spark://10.211.55.3). > > Thanks > Best Regards > >> On Tue, Jan 13, 2015 at 11:15 PM, jeremy p <athomewithagroove...@gmail.com> >> wrote: >> Hello all, >> >> I wrote some Java code that uses Spark, but for some reason I can't run it >> from the command line. I am running Spark on a single node (my >> workstation). The program stops running after this line is executed : >> >> SparkContext sparkContext = new SparkContext("spark://myworkstation:7077", >> "sparkbase"); >> >> When that line is executed, this is printed to the screen : >> 15/01/12 15:56:19 WARN util.Utils: Your hostname, myworkstation resolves to >> a loopback address: 127.0.1.1; using 10.211.55.3 instead (on interface eth0) >> 15/01/12 15:56:19 WARN util.Utils: Set SPARK_LOCAL_IP if you need to bind to >> another address >> 15/01/12 15:56:19 INFO spark.SecurityManager: Changing view acls to: >> myusername >> 15/01/12 15:56:19 INFO spark.SecurityManager: Changing modify acls to: >> myusername >> 15/01/12 15:56:19 INFO spark.SecurityManager: SecurityManager: >> authentication disabled; ui acls disabled; users with view permissions: >> Set(myusername); users with modify permissions: Set(myusername) >> >> After it writes this to the screen, the program stops executing without >> reporting an exception. >> >> What's odd is that when I run this code from Eclipse, the same lines are >> printed to the screen, but the program keeps executing. >> >> Don't know if it matters, but I'm using the maven assembly plugin, which >> includes the dependencies in the JAR. >> >> Here are the versions I'm using : >> Cloudera : 2.5.0-cdh5.2.1 >> Hadoop : 2.5.0-cdh5.2.1 >> HBase : HBase 0.98.6-cdh5.2.1 >> Java : 1.7.0_65 >> Ubuntu : 14.04.1 LTS >> Spark : 1.2 >