For my part I'm +1 on this, though Sean it would be great separately to fix the test environment.
For those who voted on rc2, this is almost identical, so feel free to +1 unless you think there are issues with the two minor bug fixes. On Sun, Aug 31, 2014 at 10:18 AM, Sean Owen <so...@cloudera.com> wrote: > Fantastic. As it happens, I just fixed up Mahout's tests for Java 8 > and observed a lot of the same type of failure. > > I'm about to submit PRs for the two issues I identified. AFAICT these > 3 then cover the failures I mentioned: > > https://issues.apache.org/jira/browse/SPARK-3329 > https://issues.apache.org/jira/browse/SPARK-3330 > https://issues.apache.org/jira/browse/SPARK-3331 > > I'd argue that none necessarily block a release, since they just > represent a problem with test-only code in Java 8, with the test-only > context of Jenkins and multiple profiles, and with a trivial > configuration in a style check for Python. Should be fixed but none > indicate a bug in the release. > > On Sun, Aug 31, 2014 at 6:11 PM, Will Benton <wi...@redhat.com> wrote: >> ----- Original Message ----- >> >>> dev/run-tests fails two tests (1 Hive, 1 Kafka Streaming) for me >>> locally on 1.1.0-rc3. Does anyone else see that? It may be my env. >>> Although I still see the Hive failure on Debian too: >>> >>> [info] - SET commands semantics for a HiveContext *** FAILED *** >>> [info] Expected Array("spark.sql.key.usedfortestonly=test.val.0", >>> "spark.sql.key.usedfortestonlyspark.sql.key.usedfortestonly=test.val.0test.val.0"), >>> but got >>> Array("spark.sql.key.usedfortestonlyspark.sql.key.usedfortestonly=test.val.0test.val.0", >>> "spark.sql.key.usedfortestonly=test.val.0") (HiveQuerySuite.scala:541) >> >> I've seen this error before. (In particular, I've seen it on my OS X >> machine using Oracle JDK 8 but not on Fedora using OpenJDK.) I've also seen >> similar errors in topic branches (but not on master) that seem to indicate >> that tests depend on sets of pairs arriving from Hive in a particular order; >> it seems that this isn't a safe assumption. >> >> I just submitted a (trivial) PR to fix this spurious failure: >> https://github.com/apache/spark/pull/2220 >> >> >> best, >> wb --------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org For additional commands, e-mail: dev-h...@spark.apache.org