What if you split the data pairwise(i.e. X_success, X_fail, etc) with subjects 
matched by row index, then run train_test_split on each one with the same 
random_state?

Naoya Kanai

Sent from
https://polymail.io/

On Mon, Sep 26, 2016 at 2:06 PM Afarin Famili

<
mailto:Afarin Famili <afarin.fam...@utsouthwestern.edu>
> wrote:

a, pre, code, a:link, body { word-wrap: break-word !important; }

Hi David,

When applying Train_test_split to the sample space, we have a single row per 
subject. I am looking for some other function like Train_test_split that can 
deal with pairs of rows (for each subject), which does not lead to a biased 
accuracy. We are studying memory and have a row of features for successful 
memory encoding, and a second row for unsuccessful memory encoding in each of 
the subjects. Our target space being 1 for successful and 0 for unsuccessful 
encoding respectively.

How do you recommend me to split this set of data in order to get a 
reasonable/unbiased accuracy?

Thanks,

Afarin

________________________________________

From: scikit-learn
mailto:utsouthwestern....@python.org
_______________________________________________
scikit-learn mailing list
scikit-learn@python.org
https://mail.python.org/mailman/listinfo/scikit-learn

Reply via email to