Github user debasish83 commented on the pull request:

    https://github.com/apache/spark/pull/1290#issuecomment-82057716
  
    @avulanov could you please point me to a stable branch that I can 
experiment with..I am focused on collaborative filtering and implemented 
various matrix factorization formulations (quadratic and nonlinear forms):
    https://issues.apache.org/jira/browse/SPARK-2426
    https://issues.apache.org/jira/browse/SPARK-6323
    Sparse Coding and PLSA are very useful for feature extraction but there are 
paper where neural nets (autoencoder) have beaten both of them.
    
    I want to start experimenting with the autoencoder variants of neural net. 
Specifically I will be focused on these 3 aspects:
    
    1. Distributing the neural net gradient calculation (I think you have 
already distributed the gradient calculation which is exactly what I want). If 
not we should do it on graphx similar to LDA architecture.
    
    2. A block coordinate descent solver running on Master with BFGS/OWLQN as 
the inner solver.
    
    3. Use L1 regularization in place of drop out heuristic to automatically 
select interesting features to make the model sparse


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to