Have not seen this happen, this is the Frequent PAttern code that was resurrected at the last minute for .9 release; but haven't seen this failure.
Does it fail consistently? On Thursday, March 6, 2014 11:41 PM, Dmitriy Lyubimov <[email protected]> wrote: I have this test failure during due diligence run -- is it a known problem ? Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 2.226 sec <<< FAILURE! - in org.apache.mahout.fpm.pfpgrowth.TransactionTreeTest testTransactionTree(org.apache.mahout.fpm.pfpgrowth.TransactionTreeTest) Time elapsed: 1.902 sec <<< FAILURE! org.junit.ComparisonFailure: expected:<...4)([7, 11, 13, 24, 2[5]],19)([7, 11, 13, 15...> but was:<...4)([7, 11, 13, 24, 2[7]],19)([7, 11, 13, 15...> at __randomizedtesting.SeedInfo.seed([2CCB549C54CC2941:939C3E240470B52D]:0) at org.junit.Assert.assertEquals(Assert.java:115) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.mahout.fpm.pfpgrowth.TransactionTreeTest.testTransactionTree(TransactionTreeTest.java:105) On Thu, Mar 6, 2014 at 8:22 PM, Dmitriy Lyubimov <[email protected]> wrote: > Just a heads up... if you haven't tracked jira... for this > > This is a big commit that, aside from its primary content, brings in Spark > 0.9 (in a separate module only), updates scala to 2.10.3 , scalatest to 2.0 > . Spark module also has CDH-compatible maven profile (which I think still > pulls CDH4 hdfs dependencies). > > Exact hadoop dependency is fairly irrelevant since one can compile and > setup spark to work with whatever version, and all mahout-spark binary > therefore will be compatible with whatever hdfs without recompilation. > > -d >
