[pymvpa] significance

Jonas Kaplan jtkaplan at usc.edu
Thu May 7 23:23:12 UTC 2009


On May 7, 2009, at 3:17 PM, Scott Gorlin wrote:

> Are your 8 chunks just separate scans, or were the conditions  
> different?  If they're just repetitions of the same thing  
> (especially, if chance performance is equal) then there should not  
> be an issue with summing the confusion matrices in each.  In theory  
> the results are independent provided that for each chunk you're  
> training your classifiers on the others.

That's right they are just 8 separate scans, each a repetition of the  
same thing.   Regarding independence, what I was thinking is that the  
results of one cross-validation step are not entirely independent from  
another since the training occurs on 6/7th of the same data points.

> Although, if you can pair up trials across your classes, I prefer  
> the McNemar test to the binomial.  This will also let you handle  
> things like if the baseline chance is expected to be different in  
> each chunk, ie if you use different stimuli.

Thanks, I will look into that.

>
> What do you mean by scrambling the regressors?  the null  
> distribution test only requires that you scramble the labels (ie see  
> how seperable the data are in directions orthogonal to the effect of  
> interest); if you mean GLM regressors then you probably should leave  
> those alone.  It might be interesting to do a ND test by placing the  
> stim regressors randomly throughout the scan but assuming that your  
> data are balanced and there's no bias in the GLM estimation, I don't  
> see there being any benefit to this over simply scrambling the  
> labels.  This would surely be much slower too if you need to go back  
> to SPM/FSL
>

Sorry, I just meant scrambling the labels.


> -Scott
>
> Jonas Kaplan wrote:
>> Hello,
>>
>> I wonder if anyone could help me think through the issue of testing  
>> classifier results for significance and how it relates to cross- 
>> validation.
>>
>> We are running a design with 8 chunks, 27 trials in each chunk  
>> divided into 3 classes.   Let's say we do an eight way (leave one  
>> out) cross-validation.  This results in an accuracy value for each  
>> set of 27 tests... 8 x 27 for a total of 216 trials that were  
>> predicted correctly or incorrectly.
>>
>> Is it wrong to use a binomial test for significance on the total  
>> number of correct predictions out of 216?  Or would that be  
>> inappropriate given that the 8 cross-validation steps are not  
>> really independent from each other and we must test each cross- 
>> validation step separately as a binomial with n=27? This latter  
>> option raises the issue of how to combine across the 8 tests.
>>
>> Alternatively, if we use the Monte Carlo simulation to produce a  
>> null distribution we have the same issue -- we are generating this  
>> null distribution for each cross-validation step -- and therefore  
>> not taking into account the overall success of the cross-validation  
>> routine across all 216 trials.   Would it make sense to generate a  
>> null distribution by scrambling the regressors and generating the  
>> results of an entire cross-validation procedure for scrambled  
>> regressors?  If so, does pymvpa have a routine for doing this?
>>
>> Thanks for any input or corrections of my thought,
>>
>>
>> Jonas
>>
>>
>>
>> P.S. we are using pymvpa for several active projects with much  
>> pleasure and will happily send you posters/papers when our work is  
>> more complete.
>>
>>
>> ----
>> Jonas Kaplan, Ph.D.
>> Research Assistant Professor
>> Brain & Creativity Institute
>> University of Southern California
>>
>>
>>
>>
>>
>>
>>
>> _______________________________________________
>> Pkg-ExpPsy-PyMVPA mailing list
>> Pkg-ExpPsy-PyMVPA at lists.alioth.debian.org
>> http://lists.alioth.debian.org/mailman/listinfo/pkg-exppsy-pymvpa




More information about the Pkg-ExpPsy-PyMVPA mailing list