Tried doing a PR to your repo and you asked for it to go to apache HEAD. I certainly didn’t want it to get into the master yet.
Happy to delete it but isn’t the Apache git OK for WIP branches? On Jun 5, 2014, at 9:18 AM, Dmitriy Lyubimov <[email protected]> wrote: * don't think we should be pushing this to apache git, I'd suggest to keep individual issue branches strictly on github. I'd suggest to drop this branch from apache git.* On Wed, Jun 4, 2014 at 6:44 PM, <[email protected]> wrote: > Repository: mahout > Updated Branches: > refs/heads/mahout-1541 8a4b4347d -> 2f87f5433 > > > MAHOUT-1541 still working on this, some refactoring in the DSL for > abstracting away Spark has moved access to rddsno Jira is closed yet > > > Project: http://git-wip-us.apache.org/repos/asf/mahout/repo > Commit: http://git-wip-us.apache.org/repos/asf/mahout/commit/2f87f543 > Tree: http://git-wip-us.apache.org/repos/asf/mahout/tree/2f87f543 > Diff: http://git-wip-us.apache.org/repos/asf/mahout/diff/2f87f543 > > Branch: refs/heads/mahout-1541 > Commit: 2f87f5433f90fa2c49ef386ca245943e1fc73beb > Parents: 8a4b434 > Author: pferrel <[email protected]> > Authored: Wed Jun 4 18:44:16 2014 -0700 > Committer: pferrel <[email protected]> > Committed: Wed Jun 4 18:44:16 2014 -0700 > > ---------------------------------------------------------------------- > .../src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala | 4 ++++ > 1 file changed, 4 insertions(+) > ---------------------------------------------------------------------- > > > > http://git-wip-us.apache.org/repos/asf/mahout/blob/2f87f543/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala > ---------------------------------------------------------------------- > diff --git > a/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala > b/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala > index 1179eef..9201c81 100644 > --- a/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala > +++ b/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala > @@ -149,6 +149,10 @@ trait TDIndexedDatasetWriter extends > Writer[IndexedDataset]{ > val matrix: DrmLike[Int] = indexedDataset.matrix > val rowIDDictionary: BiMap[String, Int] = indexedDataset.rowIDs > val columnIDDictionary: BiMap[String, Int] = > indexedDataset.columnIDs > + // below doesn't compile because the rdd is not in a > CheckpointedDrmSpark also I don't know how to turn a > + // CheckpointedDrmSpark[Int] into a DrmLike[Int], which I need to > pass in the CooccurrenceAnalysis#cooccurrence > + // This seems to be about the refacotring to abstract away from > Spark but the Read and Write are Spark specific > + // and the non-specific DrmLike is no longer attached to a > CheckpointedDrmSpark, could be missing something though > matrix.rdd.map({ case (rowID, itemVector) => > var line: String = rowIDDictionary.inverse.get(rowID) + outDelim1 > for (item <- itemVector.nonZeroes()) { > >
