sorry for confusion
On Thu, Jun 5, 2014 at 9:59 AM, Dmitriy Lyubimov <[email protected]> wrote: > I probably meant to do a PR to github's "apache/mahout" MASTER, not push > to it to git-wip-us. > > > On Thu, Jun 5, 2014 at 9:42 AM, Pat Ferrel <[email protected]> wrote: > >> Tried doing a PR to your repo and you asked for it to go to apache HEAD. >> I certainly didn’t want it to get into the master yet. >> >> Happy to delete it but isn’t the Apache git OK for WIP branches? >> >> On Jun 5, 2014, at 9:18 AM, Dmitriy Lyubimov <[email protected]> wrote: >> >> * don't think we should be pushing this to apache git, I'd suggest to keep >> individual issue branches strictly on github. I'd suggest to drop this >> branch from apache git.* >> >> >> On Wed, Jun 4, 2014 at 6:44 PM, <[email protected]> wrote: >> >> > Repository: mahout >> > Updated Branches: >> > refs/heads/mahout-1541 8a4b4347d -> 2f87f5433 >> > >> > >> > MAHOUT-1541 still working on this, some refactoring in the DSL for >> > abstracting away Spark has moved access to rddsno Jira is closed yet >> > >> > >> > Project: http://git-wip-us.apache.org/repos/asf/mahout/repo >> > Commit: http://git-wip-us.apache.org/repos/asf/mahout/commit/2f87f543 >> > Tree: http://git-wip-us.apache.org/repos/asf/mahout/tree/2f87f543 >> > Diff: http://git-wip-us.apache.org/repos/asf/mahout/diff/2f87f543 >> > >> > Branch: refs/heads/mahout-1541 >> > Commit: 2f87f5433f90fa2c49ef386ca245943e1fc73beb >> > Parents: 8a4b434 >> > Author: pferrel <[email protected]> >> > Authored: Wed Jun 4 18:44:16 2014 -0700 >> > Committer: pferrel <[email protected]> >> > Committed: Wed Jun 4 18:44:16 2014 -0700 >> > >> > ---------------------------------------------------------------------- >> > .../src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala | 4 >> ++++ >> > 1 file changed, 4 insertions(+) >> > ---------------------------------------------------------------------- >> > >> > >> > >> > >> http://git-wip-us.apache.org/repos/asf/mahout/blob/2f87f543/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala >> > ---------------------------------------------------------------------- >> > diff --git >> > a/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala >> > b/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala >> > index 1179eef..9201c81 100644 >> > --- a/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala >> > +++ b/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala >> > @@ -149,6 +149,10 @@ trait TDIndexedDatasetWriter extends >> > Writer[IndexedDataset]{ >> > val matrix: DrmLike[Int] = indexedDataset.matrix >> > val rowIDDictionary: BiMap[String, Int] = indexedDataset.rowIDs >> > val columnIDDictionary: BiMap[String, Int] = >> > indexedDataset.columnIDs >> > + // below doesn't compile because the rdd is not in a >> > CheckpointedDrmSpark also I don't know how to turn a >> > + // CheckpointedDrmSpark[Int] into a DrmLike[Int], which I need to >> > pass in the CooccurrenceAnalysis#cooccurrence >> > + // This seems to be about the refacotring to abstract away from >> > Spark but the Read and Write are Spark specific >> > + // and the non-specific DrmLike is no longer attached to a >> > CheckpointedDrmSpark, could be missing something though >> > matrix.rdd.map({ case (rowID, itemVector) => >> > var line: String = rowIDDictionary.inverse.get(rowID) + >> outDelim1 >> > for (item <- itemVector.nonZeroes()) { >> > >> > >> >> >
