[
https://issues.apache.org/jira/browse/FLINK-2116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14571009#comment-14571009
]
ASF GitHub Bot commented on FLINK-2116:
---------------------------------------
Github user sachingoel0101 commented on the pull request:
https://github.com/apache/flink/pull/772#issuecomment-108482112
Great. This is exactly what I had in mind.
There is perhaps another feature we could incorporate. Every algorithm has
some performance measure to so it can be evaluated on a test data set. We could
incorporate this as a parameter in the model. As soon as evaluate gets called,
this parameter is set to the performance value. It could be squared-error for
MLR, or F-score and accuracy, etc. for SVM, and so on.
User accesses this performance measure with a simple instance.get and (most
likely) prints it, so we don't need to make it of the same type across
different algorithms. Every Predictor can have its own performance object.
> Make pipeline extension require less coding
> -------------------------------------------
>
> Key: FLINK-2116
> URL: https://issues.apache.org/jira/browse/FLINK-2116
> Project: Flink
> Issue Type: Improvement
> Components: Machine Learning Library
> Reporter: Mikio Braun
> Assignee: Till Rohrmann
> Priority: Minor
>
> Right now, implementing methods from the pipelines for new types, or even
> adding new methods to pipelines requires many steps:
> 1) implementing methods for new types
> implement implicit of the corresponding class encapsulating the operation
> in the companion object
> 2) adding methods to the pipeline
> - adding a method
> - adding a trait for the operation
> - implement implicit in the companion object
> These are all objects which contain many generic parameters, so reducing the
> work would be great.
> The goal should be that you can really focus on the code to add, and have as
> little boilerplate code as possible.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)