[ https://issues.apache.org/jira/browse/FLINK-1745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14942703#comment-14942703 ]
ASF GitHub Bot commented on FLINK-1745: --------------------------------------- Github user danielblazevski commented on the pull request: https://github.com/apache/flink/pull/1220#issuecomment-145364401 Thanks @chiwanpark for the very useful comments. I have made changes to the comments, which can be found here: https://github.com/danielblazevski/flink/tree/FLINK-1745/flink-staging/flink-ml/src/main/scala/org/apache/flink/ml/nn I also changed the testing of KNN + QuadTree, which can be found here: https://github.com/danielblazevski/flink/tree/FLINK-1745/flink-staging/flink-ml/src/test/scala/org/apache/flink/ml/nn Since useQuadTree is now a parameter, I did not need KNNQuadTreeSuite anymore and I removed it. I did not address comment 6 yet. I need to have the training set before I can define a non-user specified useQuadTree, so any main if(useQuadTree) should come within ` val crossed = trainingSet.cross(inputSplit).mapPartition {` About your last "P.S" comment, Creating the quadtree after the cross operation is likely more efficient -- each CPU/Node will form their own quadtree, which is what is suggested for the R-tree here: https://www.cs.utah.edu/~lifeifei/papers/mrknnj.pdf This will result less communication overhead than creating a more global quadtree, if that is what you were referring to. > Add exact k-nearest-neighbours algorithm to machine learning library > -------------------------------------------------------------------- > > Key: FLINK-1745 > URL: https://issues.apache.org/jira/browse/FLINK-1745 > Project: Flink > Issue Type: New Feature > Components: Machine Learning Library > Reporter: Till Rohrmann > Assignee: Daniel Blazevski > Labels: ML, Starter > > Even though the k-nearest-neighbours (kNN) [1,2] algorithm is quite trivial > it is still used as a mean to classify data and to do regression. This issue > focuses on the implementation of an exact kNN (H-BNLJ, H-BRJ) algorithm as > proposed in [2]. > Could be a starter task. > Resources: > [1] [http://en.wikipedia.org/wiki/K-nearest_neighbors_algorithm] > [2] [https://www.cs.utah.edu/~lifeifei/papers/mrknnj.pdf] -- This message was sent by Atlassian JIRA (v6.3.4#6332)