Re: [scikit-learn] Incremental learning in scikit-learn

2020-11-15 Thread Danny Sullivan
Hi Farzana, The chunks do not have to be the same size, you just need to call partial_fit to update the model. Hope that helps. Danny El El dom, nov. 15, 2020 a la(s) 11:39 a. m., Farzana Anowar < fad...@uregina.ca> escribió: > Hello everyone, > > Currently, I am working with incremental learni

[scikit-learn] Incremental learning in scikit-learn

2020-11-15 Thread Farzana Anowar
Hello everyone, Currently, I am working with incremental learning. I know that scikit-learn allows using incremental learning for some classifiers i. e. SGD. In incremental learning, data is not available all together rather the data become available chunk by chunk over the time. Now, my que

Re: [scikit-learn] Incremental learning in scikit-learn

2019-09-09 Thread Farzana Anowar
On 2019-09-09 17:53, Daniel Sullivan wrote: Hey Farzana, The algorithm only keeps one batch in memory at a time. Between processing over each batch, SGD keeps a set of weights that it alters with each iteration of a data point or instance within a batch. This set of weights functions as the pers

Re: [scikit-learn] Incremental learning in scikit-learn

2019-09-09 Thread Daniel Sullivan
Hey Farzana, The algorithm only keeps one batch in memory at a time. Between processing over each batch, SGD keeps a set of weights that it alters with each iteration of a data point or instance within a batch. This set of weights functions as the persisted state between calls of partial_fit. That

[scikit-learn] Incremental learning in scikit-learn

2019-09-09 Thread Farzana Anowar
Hello Sir/Madam, I am going through the incremental learning algorithm in Scikit-learn. SGD in sci-kit learn is such a kind of algorithm that allows learning incrementally by passing chunks/batches. Now my question is: does sci-kit learn keeps all the batches for training data in memory? Or it