[
https://issues.apache.org/jira/browse/MAHOUT-1557?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14007235#comment-14007235
]
Sebastian Schelter commented on MAHOUT-1557:
--------------------------------------------
Karol, your patch contains some errors, e.g. the variable position is set but
never read in RunMultilayerPerceptron.
Furthermore, NeuralNetwork converts the input to a DenseVector internally in
getOutput(), so you also have to modify that code.
> Add support for sparse training vectors in MLP
> ----------------------------------------------
>
> Key: MAHOUT-1557
> URL: https://issues.apache.org/jira/browse/MAHOUT-1557
> Project: Mahout
> Issue Type: Improvement
> Components: Classification
> Reporter: Karol Grzegorczyk
> Priority: Minor
> Labels: mlp
> Fix For: 1.0
>
> Attachments: mlp_sparse.diff
>
>
> When the number of input units of MLP is big, it is likely that input vector
> will be sparse. It should be possible to read input files in a sparse format.
--
This message was sent by Atlassian JIRA
(v6.2#6252)