Do you have any plan to get rid of the memory limitation in Random Forest? Regards,
Xiaobo Gu On Thu, Jul 7, 2011 at 11:48 PM, Ted Dunning <[email protected]> wrote: > The summary of the reason is that this was a summer project and > parallelizing the random forest algorithm at all was a big enough project. > > Writing a single pass on-line algorithm was considered a bit much for the > project size. Figuring out how to make multiple passes through an input > split was similarly out of scope. > > If you have a good alternative, this would be of substantial interest > because it could improve the currently limited scalability of the decision > forest code. > > On Thu, Jul 7, 2011 at 8:20 AM, Xiaobo Gu <[email protected]> wrote: > >> Why can't a tree be built against a dataset resides on the disk as >> long as we can read it ? >> >
