ahh.. makes sense.. but would have hoped i could parelize it as i have so many cores to run on.. > Den 16. mar. 2017 kl. 18.05 skrev Julio Antonio Soto de Vicente > <[email protected]>: > > Your code is perfectly fine. > > You are training 10 networks in parallel (since you have n_jobs=10), so each > network started training in its own, and outputing its progress independently. > > Given enough amount of time, you will see that all 10 networks will > eventually get to epoch number 2, and 10 messages of epoch #2 will be printed > out. > > -- > Julio > > El 16 mar 2017, a las 17:59, Carlton Banks <[email protected] > <mailto:[email protected]>> escribió: > >> I haven’t a verbosity level in the code?… but set it to 3 as suggested by >> Julio… It did not seem to work.. >> >> https://www.dropbox.com/s/nr5rattzts0wuvd/Screenshot%20from%202017-03-16%2017%3A56%3A26.png?dl=0 >> >> <https://www.dropbox.com/s/nr5rattzts0wuvd/Screenshot%20from%202017-03-16%2017%3A56%3A26.png?dl=0> >> >>> Den 16. mar. 2017 kl. 17.51 skrev Carlton Banks <[email protected] >>> <mailto:[email protected]>>: >>> >>> Ohh.. actually the data size cannot be wrong.. >>> input_train and output_train are both lists… which i then only take a part >>> of … and then make then to a np.array… >>> >>> So that should not be incorrect. >>> >>>> Den 16. mar. 2017 kl. 17.33 skrev Carlton Banks <[email protected] >>>> <mailto:[email protected]>>: >>>> >>>> I am running this on a super computer, so yes I am running a few training >>>> sessions. >>>> I guess i will look at the verbose, and the adjust the training data size. >>>> >>>>> Den 16. mar. 2017 kl. 17.30 skrev Julio Antonio Soto de Vicente >>>>> <[email protected] <mailto:[email protected]>>: >>>>> >>>>> IMO this has nothing to do with GridSearchCV itself... >>>>> >>>>> It rather looks like different (verbose) keras models are being trained >>>>> simultaneously, and therefore "collapsing" your stdout. >>>>> >>>>> I recommend setting Keras verbosity level to 3, in order to avoid >>>>> printing the progress bars during GridSearchCV (which can be misleading). >>>>> >>>>> -- >>>>> Julio >>>>> >>>>> El 16 mar 2017, a las 16:50, Carlton Banks <[email protected] >>>>> <mailto:[email protected]>> escribió: >>>>> >>>>>> I am currently using grid search to optimize my keras model… >>>>>> >>>>>> Something seemed a bit off during the training? >>>>>> >>>>>> https://www.dropbox.com/s/da0ztv2kqtkrfpu/Screenshot%20from%202017-03-16%2016%3A43%3A42.png?dl=0 >>>>>> >>>>>> <https://www.dropbox.com/s/da0ztv2kqtkrfpu/Screenshot%20from%202017-03-16%2016:43:42.png?dl=0> >>>>>> >>>>>> For some reason is the training for each epoch not done for all >>>>>> datapoints?… >>>>>> >>>>>> What could be wrong? >>>>>> >>>>>> Here is the code: >>>>>> >>>>>> http://pastebin.com/raw/itJFm5a6 <http://pastebin.com/raw/itJFm5a6> >>>>>> >>>>>> Anything that seems off? >>>>>> _______________________________________________ >>>>>> scikit-learn mailing list >>>>>> [email protected] <mailto:[email protected]> >>>>>> https://mail.python.org/mailman/listinfo/scikit-learn >>>>>> <https://mail.python.org/mailman/listinfo/scikit-learn> >>>>> _______________________________________________ >>>>> scikit-learn mailing list >>>>> [email protected] <mailto:[email protected]> >>>>> https://mail.python.org/mailman/listinfo/scikit-learn >>>>> <https://mail.python.org/mailman/listinfo/scikit-learn> >>>> >>> >> >> _______________________________________________ >> scikit-learn mailing list >> [email protected] <mailto:[email protected]> >> https://mail.python.org/mailman/listinfo/scikit-learn >> <https://mail.python.org/mailman/listinfo/scikit-learn> > _______________________________________________ > scikit-learn mailing list > [email protected] > https://mail.python.org/mailman/listinfo/scikit-learn
_______________________________________________ scikit-learn mailing list [email protected] https://mail.python.org/mailman/listinfo/scikit-learn
