I'm comparing a few algorithms, and trying to have them run using the same
random datasets.
Each algorithm is a separate python process and I provide a file with a list of
integers, generated using numpy.random.randint. It is a small sequence of
random integers between 0 and 10,000,000.
Every time a new run of a certain experiment is made, random_state is set to a
number in the seed sequence and fed into train_test_split() function.
Since I don't know about the internals of random_state behavior, I would like
to know whether this makes sense or not.
So far, I seem to be getting reasonable results, but it'd be great to have a
second opinion.
Thank you,
------------------------------------------------------------------------------
Download BIRT iHub F-Type - The Free Enterprise-Grade BIRT Server
from Actuate! Instantly Supercharge Your Business Reports and Dashboards
with Interactivity, Sharing, Native Excel Exports, App Integration & more
Get technology previously reserved for billion-dollar corporations, FREE
http://pubads.g.doubleclick.net/gampad/clk?id=190641631&iu=/4140/ostg.clktrk
_______________________________________________
Scikit-learn-general mailing list
Scikit-learn-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/scikit-learn-general