[jira] [Commented] (SPARK-5133) Feature Importance for Decision Tree (Ensembles)

2015-07-29 Thread Parv Oberoi (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-5133?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14646493#comment-14646493
 ] 

Parv Oberoi commented on SPARK-5133:


[~josephkb]: is the plan to still include this in SPARK 1.5 considering that 
the code cutoff is this week?

> Feature Importance for Decision Tree (Ensembles)
> 
>
> Key: SPARK-5133
> URL: https://issues.apache.org/jira/browse/SPARK-5133
> Project: Spark
>  Issue Type: New Feature
>  Components: ML, MLlib
>Reporter: Peter Prettenhofer
>   Original Estimate: 168h
>  Remaining Estimate: 168h
>
> Add feature importance to decision tree model and tree ensemble models.
> If people are interested in this feature I could implement it given a mentor 
> (API decisions, etc). Please find a description of the feature below:
> Decision trees intrinsically perform feature selection by selecting 
> appropriate split points. This information can be used to assess the relative 
> importance of a feature. 
> Relative feature importance gives valuable insight into a decision tree or 
> tree ensemble and can even be used for feature selection.
> More information on feature importance (via decrease in impurity) can be 
> found in ESLII (10.13.1) or here [1].
> R's randomForest package uses a different technique for assessing variable 
> importance that is based on permutation tests.
> All necessary information to create relative importance scores should be 
> available in the tree representation (class Node; split, impurity gain, 
> (weighted) nr of samples?).
> [1] 
> http://scikit-learn.org/stable/modules/ensemble.html#feature-importance-evaluation



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-5133) Feature Importance for Decision Tree (Ensembles)

2015-04-15 Thread Parv Oberoi (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-5133?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14496625#comment-14496625
 ] 

Parv Oberoi commented on SPARK-5133:


this would be a really useful feature to have in MLLib Ensemble Tree Training. 
Exporting Feature Importance is a highly useful during feature engineering.

> Feature Importance for Decision Tree (Ensembles)
> 
>
> Key: SPARK-5133
> URL: https://issues.apache.org/jira/browse/SPARK-5133
> Project: Spark
>  Issue Type: New Feature
>  Components: ML, MLlib
>Reporter: Peter Prettenhofer
>Priority: Minor
>
> Add feature importance to decision tree model and tree ensemble models.
> If people are interested in this feature I could implement it given a mentor 
> (API decisions, etc). Please find a description of the feature below:
> Decision trees intrinsically perform feature selection by selecting 
> appropriate split points. This information can be used to assess the relative 
> importance of a feature. 
> Relative feature importance gives valuable insight into a decision tree or 
> tree ensemble and can even be used for feature selection.
> More information on feature importance (via decrease in impurity) can be 
> found in ESLII (10.13.1) or here [1].
> R's randomForest package uses a different technique for assessing variable 
> importance that is based on permutation tests.
> All necessary information to create relative importance scores should be 
> available in the tree representation (class Node; split, impurity gain, 
> (weighted) nr of samples?).
> [1] 
> http://scikit-learn.org/stable/modules/ensemble.html#feature-importance-evaluation



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org