I fixed my issue - two IP addresses had the same hostname.
Regards



On Fri, Feb 7, 2014 at 12:59 PM, Soumya Simanta <[email protected]>wrote:

> I see similar logs but only when I try to run a standalone Scala program.
> The whole setup works just fine if I'm using the spark-shell/REPL.
>
>
>
>
> On Fri, Feb 7, 2014 at 3:05 PM, mohankreddy <[email protected]>wrote:
>
>> Here's more information. I have the master up but when I try to get the
>> workers up I am getting the following error.
>>
>> log4j:WARN No appenders could be found for logger
>> (akka.event.slf4j.Slf4jLogger).
>> log4j:WARN Please initialize the log4j system properly.
>> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for
>> more info.
>> 14/02/07 15:01:17 INFO Worker: Using Spark's default log4j profile:
>> org/apache/spark/log4j-defaults.properties
>> 14/02/07 15:01:17 INFO Worker: Starting Spark worker yyyyyyy:58020 with 16
>> cores, 67.0 GB RAM
>> 14/02/07 15:01:17 INFO Worker: Spark home: /opt/spark
>> 14/02/07 15:01:17 INFO WorkerWebUI: Started Worker web UI at
>> http://yyyyyyyyy:8081
>> 14/02/07 15:01:17 INFO Worker: Connecting to master spark://xxxxx/:7077...
>> 14/02/07 15:01:17 INFO RemoteActorRefProvider$RemoteDeadLetterActorRef:
>> Message [org.apache.spark.deploy.DeployMessages$RegisterWorker] from
>> Actor[akka://sparkWorker/user/Worker#2037095035] to
>> Actor[akka://sparkWorker/deadLetters] was not delivered. [1] dead letters
>> encountered. This logging can be turned off or adjusted with configuration
>> settings 'akka.log-dead-letters' and
>> 'akka.log-dead-letters-during-shutdown'.
>> 14/02/07 15:01:37 INFO Worker: Connecting to master spark://xxxxx/:7077...
>> 14/02/07 15:01:37 INFO RemoteActorRefProvider$RemoteDeadLetterActorRef:
>> Message [org.apache.spark.deploy.DeployMessages$RegisterWorker] from
>> Actor[akka://sparkWorker/user/Worker#2037095035] to
>> Actor[akka://sparkWorker/deadLetters] was not delivered. [2] dead letters
>> encountered. This logging can be turned off or adjusted with configuration
>> settings 'akka.log-dead-letters' and
>> 'akka.log-dead-letters-during-shutdown'.
>> 14/02/07 15:01:57 INFO Worker: Connecting to master spark://xxxx/:7077...
>> 14/02/07 15:01:57 INFO RemoteActorRefProvider$RemoteDeadLetterActorRef:
>> Message [org.apache.spark.deploy.DeployMessages$RegisterWorker] from
>> Actor[akka://sparkWorker/user/Worker#2037095035] to
>> Actor[akka://sparkWorker/deadLetters] was not delivered. [3] dead letters
>> encountered. This logging can be turned off or adjusted with configuration
>> settings 'akka.log-dead-letters' and
>> 'akka.log-dead-letters-during-shutdown'.
>> 14/02/07 15:02:17 ERROR Worker: All masters are unresponsive! Giving up.
>>
>>
>>
>> PS: I masked the IPs
>>
>>
>>
>> --
>> View this message in context:
>> http://apache-spark-user-list.1001560.n3.nabble.com/Akka-Connection-refused-standalone-cluster-using-spark-0-9-0-tp1297p1311.html
>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>
>
>

Reply via email to