To maximize accuracy, n_estimators should ideally be as high as possible,
yet we would like to use a reasonable value to limit training and
prediction times. The new warm_start option is a nice way to incrementally
add more trees until you reach a satisfying accuracy.
Warm start in linear models is more general. You can also use it for
computing a regularization path (computing the solutions for many different
alphas or Cs) or for refitting an existing model with more or less data.
M.
On Wed, Jul 1, 2015 at 4:42 AM, Andreas Mueller <t3k...@gmail.com> wrote:
> It does the second.
> You need to feed it the same data.
>
>
>
> On 06/25/2015 11:59 AM, Rafael Calsaverini wrote:
>
> I saw the new parameter warm_start on the RandomForestClassifier class
> and was curious about what is its most common use. I can see two uses for
> it: (1) instead of fitting a huge forest in one go, fit it many times in
> the same data set and check the improvement in cross validation, (2) fit
> the data in a whole new data set and so doing some kind of "online
> mini-batch Random Forest", which might be useful for some of my use cases.
>
> The first use is very straightforward and it seems obvious that a forest
> trained two times in the same dataset with warm_start=True will have
> similar performance with a forest trained only "once" with double
> n_estimators.
>
> The second case though seem to break a few assumptions of the statistical
> ideas behind Random Forests. Have anyone tried it successfully?
>
> Thanks,
> Rafael Calsaverini
>
>
> ------------------------------------------------------------------------------
> Monitor 25 network devices or servers for free with OpManager!
> OpManager is web-based network management software that monitors
> network devices and physical & virtual servers, alerts via email & sms
> for fault. Monitor 25 devices for free with no restriction. Download
> nowhttp://ad.doubleclick.net/ddm/clk/292181274;119417398;o
>
>
>
> _______________________________________________
> Scikit-learn-general mailing
> listScikit-learn-general@lists.sourceforge.nethttps://lists.sourceforge.net/lists/listinfo/scikit-learn-general
>
>
>
>
> ------------------------------------------------------------------------------
> Don't Limit Your Business. Reach for the Cloud.
> GigeNET's Cloud Solutions provide you with the tools and support that
> you need to offload your IT needs and focus on growing your business.
> Configured For All Businesses. Start Your Cloud Today.
> https://www.gigenetcloud.com/
> _______________________________________________
> Scikit-learn-general mailing list
> Scikit-learn-general@lists.sourceforge.net
> https://lists.sourceforge.net/lists/listinfo/scikit-learn-general
>
>
------------------------------------------------------------------------------
Don't Limit Your Business. Reach for the Cloud.
GigeNET's Cloud Solutions provide you with the tools and support that
you need to offload your IT needs and focus on growing your business.
Configured For All Businesses. Start Your Cloud Today.
https://www.gigenetcloud.com/
_______________________________________________
Scikit-learn-general mailing list
Scikit-learn-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/scikit-learn-general