you need to do a PR of pferrel/mahout branch "MAHOUT-1541" for apache/mahout "master"
On Thu, Jun 5, 2014 at 10:10 AM, Pat Ferrel <[email protected]> wrote: > I deleted the apache.git version of the branch but it did not get mirrored > to github. > > 'git remote show apache’ shows it removed but the github UI still has it, > maybe it will disappear later. > > I’ll do a pr to github/apache/mahout, but it will by default try to merge > with master. I can’t pick the branch the PR is targeted for. That makes me > nervous but if you say so... > > > On Jun 5, 2014, at 10:02 AM, Dmitriy Lyubimov <[email protected]> wrote: > > sorry for confusion > > > On Thu, Jun 5, 2014 at 9:59 AM, Dmitriy Lyubimov <[email protected]> > wrote: > > > I probably meant to do a PR to github's "apache/mahout" MASTER, not push > > to it to git-wip-us. > > > > > > On Thu, Jun 5, 2014 at 9:42 AM, Pat Ferrel <[email protected]> wrote: > > > >> Tried doing a PR to your repo and you asked for it to go to apache HEAD. > >> I certainly didn’t want it to get into the master yet. > >> > >> Happy to delete it but isn’t the Apache git OK for WIP branches? > >> > >> On Jun 5, 2014, at 9:18 AM, Dmitriy Lyubimov <[email protected]> wrote: > >> > >> * don't think we should be pushing this to apache git, I'd suggest to > keep > >> individual issue branches strictly on github. I'd suggest to drop this > >> branch from apache git.* > >> > >> > >> On Wed, Jun 4, 2014 at 6:44 PM, <[email protected]> wrote: > >> > >>> Repository: mahout > >>> Updated Branches: > >>> refs/heads/mahout-1541 8a4b4347d -> 2f87f5433 > >>> > >>> > >>> MAHOUT-1541 still working on this, some refactoring in the DSL for > >>> abstracting away Spark has moved access to rddsno Jira is closed yet > >>> > >>> > >>> Project: http://git-wip-us.apache.org/repos/asf/mahout/repo > >>> Commit: http://git-wip-us.apache.org/repos/asf/mahout/commit/2f87f543 > >>> Tree: http://git-wip-us.apache.org/repos/asf/mahout/tree/2f87f543 > >>> Diff: http://git-wip-us.apache.org/repos/asf/mahout/diff/2f87f543 > >>> > >>> Branch: refs/heads/mahout-1541 > >>> Commit: 2f87f5433f90fa2c49ef386ca245943e1fc73beb > >>> Parents: 8a4b434 > >>> Author: pferrel <[email protected]> > >>> Authored: Wed Jun 4 18:44:16 2014 -0700 > >>> Committer: pferrel <[email protected]> > >>> Committed: Wed Jun 4 18:44:16 2014 -0700 > >>> > >>> ---------------------------------------------------------------------- > >>> .../src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala | 4 > >> ++++ > >>> 1 file changed, 4 insertions(+) > >>> ---------------------------------------------------------------------- > >>> > >>> > >>> > >>> > >> > http://git-wip-us.apache.org/repos/asf/mahout/blob/2f87f543/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala > >>> ---------------------------------------------------------------------- > >>> diff --git > >>> a/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala > >>> b/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala > >>> index 1179eef..9201c81 100644 > >>> --- a/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala > >>> +++ b/spark/src/main/scala/org/apache/mahout/drivers/ReaderWriter.scala > >>> @@ -149,6 +149,10 @@ trait TDIndexedDatasetWriter extends > >>> Writer[IndexedDataset]{ > >>> val matrix: DrmLike[Int] = indexedDataset.matrix > >>> val rowIDDictionary: BiMap[String, Int] = indexedDataset.rowIDs > >>> val columnIDDictionary: BiMap[String, Int] = > >>> indexedDataset.columnIDs > >>> + // below doesn't compile because the rdd is not in a > >>> CheckpointedDrmSpark also I don't know how to turn a > >>> + // CheckpointedDrmSpark[Int] into a DrmLike[Int], which I need > to > >>> pass in the CooccurrenceAnalysis#cooccurrence > >>> + // This seems to be about the refacotring to abstract away from > >>> Spark but the Read and Write are Spark specific > >>> + // and the non-specific DrmLike is no longer attached to a > >>> CheckpointedDrmSpark, could be missing something though > >>> matrix.rdd.map({ case (rowID, itemVector) => > >>> var line: String = rowIDDictionary.inverse.get(rowID) + > >> outDelim1 > >>> for (item <- itemVector.nonZeroes()) { > >>> > >>> > >> > >> > > > >
