[ 
https://issues.apache.org/jira/browse/SPARK-5820?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322022#comment-14322022
 ] 

Eric O. LEBIGOT (EOL) edited comment on SPARK-5820 at 2/15/15 3:27 PM:
-----------------------------------------------------------------------

Thank you for investigating this.

The proxy server (localhost) does not require authentication.

The example (SparkPi) does work when no SOCKS proxy is used (in the OS X 
Network Preferences).

I did not mean to say that my proxy doesn't work for any JVM-based process (I'm 
not sure how I could test this). :) If you were referring to the three JVM 
settings that I tried (with the local SOCKS proxy server active), they are 
simply those of the OS X Java Preferences (Apple>System 
Preferences>Java>Network Settings). For all JVM proxy settings, 
./bin/run-example SparkPi fails in the same way (as in the original report).

Here is how to reproduce the problem:
1) Creation of the local SOCKS proxy: ssh -g -N -D 12345 login@remotehost
2) Activation of the proxy in OS X: Apple>System 
Preferences>Network>Advanced>Proxies, select SOCKS Proxy, set SOCKS Proxy 
Server to localhost, port 12345, make sure that Bypassed hosts contains 
localhost (and 127.0.0.1, for good measure, maybe).

Now, connections go through the proxy (this should show in 
http://www.whatismyip.com/, for example).


was (Author: lebigot):
Thank you for investigating this.

The proxy server (localhost) does not require authentication.

The example (SparkPi) does work when no SOCKS proxy is used (in the OS X 
Network Preferences).

I did not mean to say that my proxy doesn't work for any JVM-based process (I'm 
not sure how I could test this). :) If you were referring to the three JVM 
settings that I tried (with the local SOCKS proxy server active), they are 
simply those of the OS X Java Preferences (Apple>System 
Preferences>Java>Network Settings). For all JVM proxy settings, 
./bin/run-example SparkPi fails in the same way (as in the original report).

Here is how to reproduce the problem:
1) Creation of the local SOCKS proxy: ssh -g -N -D 12345 login@remotehost
2) Activation of the proxy in OS X: Apple>System 
Preferences>Network>Advanced>Proxies, select SOCKS Proxy, set SOCKS Proxy 
Server to localhost, port 12345.

Now, browser connections go through the proxy (this should show in 
http://www.whatismyip.com/, for example).

> Example does not work when using SOCKS proxy
> --------------------------------------------
>
>                 Key: SPARK-5820
>                 URL: https://issues.apache.org/jira/browse/SPARK-5820
>             Project: Spark
>          Issue Type: Bug
>    Affects Versions: 1.2.1
>            Reporter: Eric O. LEBIGOT (EOL)
>
> When using a SOCKS proxy (on OS X 10.10.2), running even the basic example 
> ./bin/run-example SparkPi 10 fails.
> -- Partial log --
> 15/02/14 23:23:00 ERROR TaskSetManager: Task 0 in stage 0.0 failed 1 times; 
> aborting job
> 15/02/14 23:23:00 INFO TaskSchedulerImpl: Cancelling stage 0
> 15/02/14 23:23:00 INFO Executor: Executor is trying to kill task 1.0 in stage 
> 0.0 (TID 1)
> 15/02/14 23:23:00 INFO TaskSchedulerImpl: Stage 0 was cancelled
> 15/02/14 23:23:00 INFO DAGScheduler: Job 0 failed: reduce at 
> SparkPi.scala:35, took 1.920223 s
> Exception in thread "main" org.apache.spark.SparkException: Job aborted due 
> to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: 
> Lost task 0.0 in stage 0.0 (TID 0, localhost): java.net.SocketException: 
> Malformed reply from SOCKS server
>         at java.net.SocksSocketImpl.readSocksReply(SocksSocketImpl.java:129)
>         at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:503)
>         at java.net.Socket.connect(Socket.java:579)
>         at sun.net.NetworkClient.doConnect(NetworkClient.java:175)
>         at sun.net.www.http.HttpClient.openServer(HttpClient.java:432)
>         at sun.net.www.http.HttpClient.openServer(HttpClient.java:527)
>         at sun.net.www.http.HttpClient.<init>(HttpClient.java:211)
>         at sun.net.www.http.HttpClient.New(HttpClient.java:308)
>         at 
> sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1003)
>         at 
> sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:951)
>         at 
> sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:850)
>         at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:582)
>         at org.apache.spark.util.Utils$.fetchFile(Utils.scala:433)
>         at 
> org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$6.apply(Executor.scala:356)
>         at 
> org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$6.apply(Executor.scala:353)
>         at 
> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:778)
>         at 
> scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:99)
>         at 
> scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:99)
>         at 
> scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230)
>         at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
>         at scala.collection.mutable.HashMap.foreach(HashMap.scala:99)
>         at 
> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:777)
>         at 
> org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:353)
>         at 
> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:181)
>         at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>         at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>         at java.lang.Thread.run(Thread.java:724)
> Driver stacktrace:
>         at 
> org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1214)
>         at 
> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1203)
>         at 
> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1202)
>         at 
> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
>         at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
>         at 
> org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1202)
>         at 
> org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:696)
>         at 
> org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:696)
>         at scala.Option.foreach(Option.scala:245)
>         at 
> org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:696)
>         at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessActor$$anonfun$receive$2.applyOrElse(DAGScheduler.scala:1420)
>         at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
>         at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessActor.aroundReceive(DAGScheduler.scala:1375)
>         at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
>         at akka.actor.ActorCell.invoke(ActorCell.scala:487)
>         at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:238)
>         at akka.dispatch.Mailbox.run(Mailbox.scala:220)
>         at 
> akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:393)
>         at 
> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>         at 
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>         at 
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>         at 
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to