I probably meant to do a PR to github's "apache/mahout" MASTER, not push to it to git-wip-us.
On Thu, Jun 5, 2014 at 9:42 AM, Pat Ferrel <[email protected]> wrote: > Tried doing a PR to your repo and you asked for it to go to apache HEAD. I > certainly didn’t want it to get into the master yet. > > Happy to delete it but isn’t the Apache git OK for WIP branches? > > On Jun 5, 2014, at 9:18 AM, Dmitriy Lyubimov <[email protected]> wrote: > > * don't think we should be pushing this to apache git, I'd suggest to keep > individual issue branches strictly on github. I'd suggest to drop this > branch from apache git.* > > > On Wed, Jun 4, 2014 at 6:44 PM, <[email protected]> wrote: > > > Repository: mahout > > Updated Branches: > > refs/heads/mahout-1541 8a4b4347d -> 2f87f5433 > > > > > > MAHOUT-1541 still working on this, some refactoring in the DSL for > > abstracting away Spark has moved access to rddsno Jira is closed yet > > > > > > Project: http://git-wip-us.apache.org/repos/asf/mahout/repo > > Commit: http://git-wip-us.apache.org/repos/asf/mahout/commit/2f87f543 > > Tree: http://git-wip-us.apache.org/repos/asf/mahout/tree/2f87f543 > > Diff: http://git-wip-us.apache.org/repos/asf/mahout/diff/2f87f543 > > > > Branch: refs/heads/mahout-1541 > > Commit: 2f87f5433f90fa2c49ef386ca245943e1fc73beb > > Parents: 8a4b434 > > Author: pferrel <[email protected]> > > Authored: Wed Jun 4 18:44:16 2014 -0700 > > Committer: pferrel <[email protected]> > > Committed: Wed Jun 4 18:44:16 2014 -0700 > > > > ---------------------------------------------------------------------- > > .../src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala | 4 ++++ > > 1 file changed, 4 insertions(+) > > ---------------------------------------------------------------------- > > > > > > > > > http://git-wip-us.apache.org/repos/asf/mahout/blob/2f87f543/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala > > ---------------------------------------------------------------------- > > diff --git > > a/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala > > b/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala > > index 1179eef..9201c81 100644 > > --- a/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala > > +++ b/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala > > @@ -149,6 +149,10 @@ trait TDIndexedDatasetWriter extends > > Writer[IndexedDataset]{ > > val matrix: DrmLike[Int] = indexedDataset.matrix > > val rowIDDictionary: BiMap[String, Int] = indexedDataset.rowIDs > > val columnIDDictionary: BiMap[String, Int] = > > indexedDataset.columnIDs > > + // below doesn't compile because the rdd is not in a > > CheckpointedDrmSpark also I don't know how to turn a > > + // CheckpointedDrmSpark[Int] into a DrmLike[Int], which I need to > > pass in the CooccurrenceAnalysis#cooccurrence > > + // This seems to be about the refacotring to abstract away from > > Spark but the Read and Write are Spark specific > > + // and the non-specific DrmLike is no longer attached to a > > CheckpointedDrmSpark, could be missing something though > > matrix.rdd.map({ case (rowID, itemVector) => > > var line: String = rowIDDictionary.inverse.get(rowID) + outDelim1 > > for (item <- itemVector.nonZeroes()) { > > > > > >
