Hmm, I can’t see why using ~ would be problematic, especially if you confirm that echo ~/path/to/pem expands to the correct path to your identity file.
If you have a simple reproduction of the problem, please send it over. I’d love to look into this. When I pass paths with ~ to spark-ec2 on my system, it works fine. I’m using bash, but zsh handles tilde expansion the same as bash. Nick On Wed Jan 28 2015 at 3:30:08 PM Charles Feduke <charles.fed...@gmail.com> wrote: > It was only hanging when I specified the path with ~ I never tried > relative. > > Hanging on the waiting for ssh to be ready on all hosts. I let it sit for > about 10 minutes then I found the StackOverflow answer that suggested > specifying an absolute path, cancelled, and re-run with --resume and the > absolute path and all slaves were up in a couple minutes. > > (I've stood up 4 integration clusters and 2 production clusters on EC2 > since with no problems.) > > On Wed Jan 28 2015 at 12:05:43 PM Nicholas Chammas < > nicholas.cham...@gmail.com> wrote: > >> Ey-chih, >> >> That makes more sense. This is a known issue that will be fixed as part >> of SPARK-5242 <https://issues.apache.org/jira/browse/SPARK-5242>. >> >> Charles, >> >> Thanks for the info. In your case, when does spark-ec2 hang? Only when >> the specified path to the identity file doesn't exist? Or also when you >> specify the path as a relative path or with ~? >> >> Nick >> >> >> On Wed Jan 28 2015 at 9:29:34 AM ey-chih chow <eyc...@hotmail.com> wrote: >> >>> We found the problem and already fixed it. Basically, spark-ec2 >>> requires ec2 instances to have external ip addresses. You need to specify >>> this in the ASW console. >>> ------------------------------ >>> From: nicholas.cham...@gmail.com >>> Date: Tue, 27 Jan 2015 17:19:21 +0000 >>> Subject: Re: spark 1.2 ec2 launch script hang >>> To: charles.fed...@gmail.com; pzybr...@gmail.com; eyc...@hotmail.com >>> CC: user@spark.apache.org >>> >>> >>> For those who found that absolute vs. relative path for the pem file >>> mattered, what OS and shell are you using? What version of Spark are you >>> using? >>> >>> ~/ vs. absolute path shouldn’t matter. Your shell will expand the ~/ to >>> the absolute path before sending it to spark-ec2. (i.e. tilde >>> expansion.) >>> >>> Absolute vs. relative path (e.g. ../../path/to/pem) also shouldn’t >>> matter, since we fixed that for Spark 1.2.0 >>> <https://issues.apache.org/jira/browse/SPARK-4137>. Maybe there’s some >>> case that we missed? >>> >>> Nick >>> >>> On Tue Jan 27 2015 at 10:10:29 AM Charles Feduke < >>> charles.fed...@gmail.com> wrote: >>> >>> >>> Absolute path means no ~ and also verify that you have the path to the >>> file correct. For some reason the Python code does not validate that the >>> file exists and will hang (this is the same reason why ~ hangs). >>> On Mon, Jan 26, 2015 at 10:08 PM Pete Zybrick <pzybr...@gmail.com> >>> wrote: >>> >>> Try using an absolute path to the pem file >>> >>> >>> >>> > On Jan 26, 2015, at 8:57 PM, ey-chih chow <eyc...@hotmail.com> wrote: >>> > >>> > Hi, >>> > >>> > I used the spark-ec2 script of spark 1.2 to launch a cluster. I have >>> > modified the script according to >>> > >>> > https://github.com/grzegorz-dubicki/spark/commit/5dd8458d2ab >>> 9753aae939b3bb33be953e2c13a70 >>> > >>> > But the script was still hung at the following message: >>> > >>> > Waiting for cluster to enter 'ssh-ready' >>> > state............................................. >>> > >>> > Any additional thing I should do to make it succeed? Thanks. >>> > >>> > >>> > Ey-Chih Chow >>> > >>> > >>> > >>> > -- >>> > View this message in context: http://apache-spark-user-list. >>> 1001560.n3.nabble.com/spark-1-2-ec2-launch-script-hang-tp21381.html >>> > Sent from the Apache Spark User List mailing list archive at >>> Nabble.com. >>> > >>> > --------------------------------------------------------------------- >>> > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>> > For additional commands, e-mail: user-h...@spark.apache.org >>> > >>> >>> --------------------------------------------------------------------- >>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>> For additional commands, e-mail: user-h...@spark.apache.org >>> >>> >>> >>> >>