[ 
https://issues.apache.org/jira/browse/MAHOUT-1750?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15233804#comment-15233804
 ] 

ASF GitHub Bot commented on MAHOUT-1750:
----------------------------------------

GitHub user andrewpalumbo opened a pull request:

    https://github.com/apache/mahout/pull/215

    MAHOUT-1750 - For Review.  FlinkOpABt too many operations to implement?

    When (partially) Implementing `FlinkOpABt`  so far, It seems that there may 
be too many Flink map/reduce/group/etc operations to finish using this method. 
**NOTE**: this is unfinished and is not numerically correct.
    
    Currently, when testing am getting kyro stack overflow exceptions, which as 
I understand are often caused by a long string of operations: 
    ```
    Job execution failed.
    org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
        at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply$mcV$sp(JobManager.scala:716)
        at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:662)
        at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:662)
        at 
scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24)
        at 
scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24)
        at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:41)
        at 
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:401)
        at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
        at 
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
        at 
scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
        at 
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
    Caused by: java.lang.StackOverflowError
        at 
com.esotericsoftware.kryo.util.DefaultClassResolver.writeClass(DefaultClassResolver.java:74)
        at com.esotericsoftware.kryo.Kryo.writeClass(Kryo.java:488)
        at 
com.esotericsoftware.kryo.serializers.ObjectField.write(ObjectField.java:57)
        at 
com.esotericsoftware.kryo.serializers.FieldSerializer.write(FieldSerializer.java:495)
        at com.esotericsoftware.kryo.Kryo.writeObject(Kryo.java:523)
    {...}```
    
    Any comments are appreciated.

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/andrewpalumbo/mahout MAHOUT-1750

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/mahout/pull/215.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #215
    
----
commit fab6bb86fa1af0a7efce2300dcada45c32b35677
Author: Andrew Palumbo <[email protected]>
Date:   2016-03-28T01:34:34Z

    start Stealing DL's Spark code

commit 7e1a8bbaaca71036e9c0247f07a40368432f592b
Author: Andrew Palumbo <[email protected]>
Date:   2016-03-28T01:55:01Z

    wip: Spark combineByKey => Flink groupReduce

commit 31a644b36ed92c5aae3b80d5e8fbca7dd8a1d87a
Author: Andrew Palumbo <[email protected]>
Date:   2016-03-28T22:38:31Z

    implement first step of the combiner

commit 7dc73890ac5eeaad1ba38ab799b19b989c6ff918
Author: Andrew Palumbo <[email protected]>
Date:   2016-03-28T22:53:09Z

    Add in type information for maps, etc.

commit 5c60167f92a32b4c0751f6bf8dc12a838e1cde78
Author: Andrew Palumbo <[email protected]>
Date:   2016-03-28T23:50:12Z

    Merge branch 'flink-binding' of 
https://git-wip-us.apache.org/repos/asf/mahout into MAHOUT-1750

commit 9b01a4a498c080e6063387608253890a13e77ad4
Author: Andrew Palumbo <[email protected]>
Date:   2016-03-29T02:48:43Z

    remove all unnecessary type info still not builing with 
GroupCombineFunctions

commit 37391a5a51b6803fdf893349a45a61903a6928e5
Author: Andrew Palumbo <[email protected]>
Date:   2016-03-30T00:07:26Z

    wip

commit d65e23b91f3c055c80de432dc9699c62512dd34e
Author: Andrew Palumbo <[email protected]>
Date:   2016-03-30T00:07:40Z

    Merge branch 'flink-binding' of 
https://git-wip-us.apache.org/repos/asf/mahout into MAHOUT-1750

commit 3127d3218d9fa0fba35c9c240db63e71b00c8a32
Author: Andrew Palumbo <[email protected]>
Date:   2016-03-30T01:01:15Z

    wip: still problems with  error: object creation impossible, since method 
mapPartition in class RichMapPartitionFunction of type (x: Iterable[(Array[K1], 
org.apache.mahout.math.Matrix)], x: org.apache.flink.util.Collector[(Int, 
Array[K1], org.apache.mahout.math.Matrix)])Unit is not defined

commit c19493b058e6d92f10fadeb32be9bafb14e7c671
Author: Andrew Palumbo <[email protected]>
Date:   2016-04-08T05:17:03Z

    WIP

commit b9ed381c2178392e785c4835adfac53b8d0ec5fb
Author: Andrew Palumbo <[email protected]>
Date:   2016-04-08T08:41:58Z

    Revert "WIP"
    
    This reverts commit c19493b058e6d92f10fadeb32be9bafb14e7c671.

commit 3311e89fbef3751476dbf113f680330b61f64160
Author: Andrew Palumbo <[email protected]>
Date:   2016-04-08T08:51:20Z

    Merge branch 'flink-binding' of 
https://git-wip-us.apache.org/repos/asf/mahout into MAHOUT-1750

commit db21e7b12a0b4e05ff27fa9031fd8bf3e5f22bda
Author: Andrew Palumbo <[email protected]>
Date:   2016-04-08T17:02:05Z

    Use java.lang.Iterable instead of scala.Iterable- this fixes the object 
creation issue.

commit d3d60fadf50ceb12f559d4c94c9402735c8a7d30
Author: Andrew Palumbo <[email protected]>
Date:   2016-04-08T21:45:03Z

    wip

commit 0a83e122c82eb0584a00510f7fb571432e834064
Author: Andrew Palumbo <[email protected]>
Date:   2016-04-08T22:11:26Z

    wip: error is
    
    Error:(180, 23) overriding method reduce in trait ReduceFunction of type 
(x: (Array[K], org.apache.mahout.math.Matrix), x: (Array[K], 
org.apache.mahout.math.Matrix))(Array[K], org.apache.mahout.math.Matrix);
     method reduce has incompatible type
                      def reduce(mx1: (Array[K], Matrix), mx2: (Array[K], 
Matrix)) {
                          ^

commit 97a6aea7c4fd0abc719605991322ac4acb758ea8
Author: Andrew Palumbo <[email protected]>
Date:   2016-04-08T22:20:56Z

    wip: compiling now

commit 088d925ba3cc5e93e25ad37f8e2fcf4d4698595f
Author: Andrew Palumbo <[email protected]>
Date:   2016-04-09T00:44:22Z

    not returning any partitions.  failing at dimensionality count in 
CheckpointedDrmFlink

commit 82ae321b85b84e478152f208059f9c8146d814c8
Author: Andrew Palumbo <[email protected]>
Date:   2016-04-09T01:01:05Z

    wip: need to make combiner into a merger

commit 3c7c2ff937206b0882f6395058a8646ed1c0cdaa
Author: Andrew Palumbo <[email protected]>
Date:   2016-04-09T14:29:38Z

    wip: comments, combiner

commit cd2b59ac96b63c93ae18619cdf22b6c2742e93d3
Author: Andrew Palumbo <[email protected]>
Date:   2016-04-09T19:58:29Z

    wip: kryo error

commit d313ebc1618ae5cb44b5161b492bb03baba8216d
Author: Andrew Palumbo <[email protected]>
Date:   2016-04-10T00:58:13Z

    Kryo stackOverflow

----


> Mahout DSL for Flink: Implement ABt
> -----------------------------------
>
>                 Key: MAHOUT-1750
>                 URL: https://issues.apache.org/jira/browse/MAHOUT-1750
>             Project: Mahout
>          Issue Type: Task
>          Components: Math
>    Affects Versions: 0.10.2
>            Reporter: Alexey Grigorev
>            Assignee: Andrew Palumbo
>            Priority: Minor
>             Fix For: 0.12.0
>
>
> Now ABt is expressed through AtB, which is not optimal, and we need to have a 
> special implementation for ABt



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to