I'd phrase that as "breaking the run structure is **SOMETIMES** ok for classification" ...

You should definitely check the partitioning schemes for order effects and be as conservative as possible. For example, you could ensure that samples collected 2 TR apart are not split between the training and testing sets, etc. If you have signs that your results are very sensitive to partitioning scheme (e.g. vastly different accuracies when shifting the partitioning by one sample) you'll need to look closer at dependencies. Basically, use extreme caution.

I'm not current enough on pymvpa to give any advice there ... I do strongly suggest pre-planning the partitioning schemes (i.e. which examples in which sets on which folds and replications) prior to starting to ensure balance and to enable sensitivity testing.

Jo



On 8/10/2012 1:53 PM, Edmund Chong wrote:
Thanks Jo!

For me, I don't have that many runs, so partitioning on groups of runs
is not really a good option. So I'd rather try doing "leave-n-samples
out" instead of "leave-n-runs out" -- I looked at your paper and indeed
it seems that breaking the run structure is ok for classification
However, do you know of any way to do that with pre-written pymvpa
functions, or do I have to manually partition the runs myself?

Thanks!
-Edmund

On Wed, Aug 8, 2012 at 3:37 PM, J.A. Etzel <[email protected]
<mailto:[email protected]>> wrote:

    What you describe is one option. I talked about those types of
    schemes and when they can be ok (in my opinion!) in
    http://dx.doi.org/10.1016/j.__neuroimage.2010.08.050
    <http://dx.doi.org/10.1016/j.neuroimage.2010.08.050>

    As general advice, it seems best to try to partition so that the
    number of examples of each case in each cross-validation fold is
    roughly equal. Sometimes that's just plain not possible. For
    example, I have a dataset with a large number of runs, but only
    trials the person gets correct are analyzed, so the number of
    examples in some runs for some people varies drastically. What we
    did in this case was to partition on groups of runs, so one fold is
    to leave runs 1,2,3, and 4 out. This scheme equalized the number of
    examples somewhat (though I still subsetted examples to make them
    exactly equal), and seemed to help the amount of variation.

    Jo



    On 8/7/2012 10:52 AM, Edmund Chong wrote:

        Hi all,

        I recently asked a question on dealing with unbalanced datasets and
        here's a follow-up question.
        So let's say I have empty runs, or runs where there are zero
        samples for
        one of the conditions. This leads to problems if that run
        happens to be
        the test run on a leave-one-run-out cross-validation procedure.

        My workaround for that was this: if I had one of such runs with
        empty
        conditions, then I would set NFoldPartitioner(cvtype=2),
        together with
        Balancer() so that any combination of two runs would have at
        least one
        sample per condition. But if I had two of such runs with empty
        conditions, then I would set cvtype=3, and so on. However this
        means I
        have less data for the training set on each classification fold.

        Is there any other possible solution for this? In fact, is it
        possible
        to do leave-n-samples-out classification: So on each fold I randomly
        select n samples per condition to test on, and use the remaining
        samples
        (after balancing) for training, disregarding the chunks structure?

        Thanks!
        -Edmund


        _________________________________________________
        Pkg-ExpPsy-PyMVPA mailing list
        Pkg-ExpPsy-PyMVPA@lists.__alioth.debian.org
        <mailto:[email protected]>
        
http://lists.alioth.debian.__org/cgi-bin/mailman/listinfo/__pkg-exppsy-pymvpa
        
<http://lists.alioth.debian.org/cgi-bin/mailman/listinfo/pkg-exppsy-pymvpa>


    --
    Joset A. Etzel, Ph.D.
    Research Analyst
    Cognitive Control & Psychopathology Lab
    Washington University in St. Louis
    http://mvpa.blogspot.com/

    _________________________________________________
    Pkg-ExpPsy-PyMVPA mailing list
    Pkg-ExpPsy-PyMVPA@lists.__alioth.debian.org
    <mailto:[email protected]>
    
http://lists.alioth.debian.__org/cgi-bin/mailman/listinfo/__pkg-exppsy-pymvpa
    <http://lists.alioth.debian.org/cgi-bin/mailman/listinfo/pkg-exppsy-pymvpa>




_______________________________________________
Pkg-ExpPsy-PyMVPA mailing list
[email protected]
http://lists.alioth.debian.org/cgi-bin/mailman/listinfo/pkg-exppsy-pymvpa


--
Joset A. Etzel, Ph.D.
Research Analyst
Cognitive Control & Psychopathology Lab
Washington University in St. Louis
http://mvpa.blogspot.com/

_______________________________________________
Pkg-ExpPsy-PyMVPA mailing list
[email protected]
http://lists.alioth.debian.org/cgi-bin/mailman/listinfo/pkg-exppsy-pymvpa

Reply via email to