I'm trying to launch Spark with this command on AWS: *./spark-ec2 -k keypair_name -i keypair.pem -s 5 -t c1.xlarge -r us-west-2 --hadoop-major-version=2.4.0 launch spark_cluster*
This script is erroring out with this message: *ssh: connect to host <hostname> port 22: Connection refused Error executing remote command, retrying after 30 seconds*: Command '['ssh', '-o', 'StrictHostKeyChecking=no', '-i', 'keypair.pem', '-t', '-t', u'root@<hostname>', "\n [ -f ~/.ssh/id_rsa ] ||\n (ssh-keygen -q -t rsa -N '' -f ~/.ssh/id_rsa &&\n cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys)\n "]' returned non-zero exit status 255 Strange this is, I can manually ssh to master node as "root" using this command: *ssh root@<hostname> -i keypair.pem* Does anyone know what is going on here? Any help is appreciated. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/ssh-connection-refused-tp10818.html Sent from the Apache Spark User List mailing list archive at Nabble.com.