+1 On Wed, Dec 16, 2015 at 6:15 PM, Ted Yu <[email protected]> wrote:
> Ran test suite (minus docker-integration-tests) > All passed > > +1 > > [INFO] Spark Project External ZeroMQ ...................... SUCCESS [ > 13.647 s] > [INFO] Spark Project External Kafka ....................... SUCCESS [ > 45.424 s] > [INFO] Spark Project Examples ............................. SUCCESS [02:06 > min] > [INFO] Spark Project External Kafka Assembly .............. SUCCESS [ > 11.280 s] > [INFO] > ------------------------------------------------------------------------ > [INFO] BUILD SUCCESS > [INFO] > ------------------------------------------------------------------------ > [INFO] Total time: 01:49 h > [INFO] Finished at: 2015-12-16T17:06:58-08:00 > > On Wed, Dec 16, 2015 at 4:37 PM, Andrew Or <[email protected]> wrote: > >> +1 >> >> Mesos cluster mode regression in RC2 is now fixed (SPARK-12345 >> <https://issues.apache.org/jira/browse/SPARK-12345> / PR10332 >> <https://github.com/apache/spark/pull/10332>). >> >> Also tested on standalone client and cluster mode. No problems. >> >> 2015-12-16 15:16 GMT-08:00 Rad Gruchalski <[email protected]>: >> >>> I also noticed that spark.replClassServer.host and >>> spark.replClassServer.port aren’t used anymore. The transport now happens >>> over the main RpcEnv. >>> >>> Kind regards, >>> Radek Gruchalski >>> [email protected] <[email protected]> >>> de.linkedin.com/in/radgruchalski/ >>> >>> >>> *Confidentiality:*This communication is intended for the above-named >>> person and may be confidential and/or legally privileged. >>> If it has come to you in error you must take no action based on it, nor >>> must you copy or show it to anyone; please delete/destroy and inform the >>> sender immediately. >>> >>> On Wednesday, 16 December 2015 at 23:43, Marcelo Vanzin wrote: >>> >>> I was going to say that spark.executor.port is not used anymore in >>> 1.6, but damn, there's still that akka backend hanging around there >>> even when netty is being used... we should fix this, should be a >>> simple one-liner. >>> >>> On Wed, Dec 16, 2015 at 2:35 PM, singinpirate <[email protected]> >>> wrote: >>> >>> -0 (non-binding) >>> >>> I have observed that when we set spark.executor.port in 1.6, we get >>> thrown a >>> NPE in SparkEnv$.create(SparkEnv.scala:259). It used to work in 1.5.2. Is >>> anyone else seeing this? >>> >>> >>> -- >>> Marcelo >>> >>> --------------------------------------------------------------------- >>> To unsubscribe, e-mail: [email protected] >>> For additional commands, e-mail: [email protected] >>> >>> >>> >> >
