[pymvpa] Justification for trial averaging?

MS Al-Rawi rawi707 at yahoo.com
Thu Jan 23 22:03:54 UTC 2014



I think a correlation classifier/method was used in Haxby's et al 2001 work, and it gave high classification accuracy using the averages. 
One might argue that, although not sure about this, assigning a volume/exemplar to a single label/condition is problematic, thus, averaging is a good option. 


Using a correlation-based classifier on exemplars/volumes would give less accuracy than the using the averages, but other powerful classifiers, e.g. SVMs, LR, ANN, RIDGE-LR will do well. 

-Rawi



> On Thursday, January 23, 2014 8:06 PM, J.A. Etzel <jetzel at artsci.wustl.edu> wrote:
> > I also agree, and will "toss in" a few more ideas:
> 
>>>  But forming decisions boundaries over features is exactly what a
>>>  classifier is meant to do, so why not just throw all these
>>>  different exemplars into the mix, and let the classifier figure out
>>>  its own notion of prototypicality?
> I think because of power, particularly the lack of it. Our datasets are
> usually massively out of balance (way more dimensions than examples),
> making learning quite difficult. It often just isn't possible to further
> subdivide the data to let the classifier learn the exemplars as well.
> 
>>>  And if you’re going to pre-classify, why pick the average
>>>  response? Why not take some kind of lower-dimensional input; the
>>>  first several eigenvectors or something, or something else?
> Probably the most common technique other than averaging is creating
> "parameter estimate images": taking the beta weights that result from
> fitting a linear model to the inputs, convolved with a hemodynamic
> response function. This can be done in programs such as SPM, and is a
> bit closer to the first level analyses done for mass-univariate analysis.
> 
>>  It seems weird to average the regressor weights, but maybe it
>>  shouldn't. Is that something that's done, or is the averaging 
> process
>>  only used with raw voxel activity?
> It does seem a bit weird, but I've done it. It feels "cleaner" to 
> generate a single set of parameter estimates than to generate one per 
> example (or run) then average, but I don't know if it is actually 
> mathematically different.
> 
> Jo
> 
> 
> -- 
> Joset A. Etzel, Ph.D.
> Research Analyst
> Cognitive Control & Psychopathology Lab
> Washington University in St. Louis
> http://mvpa.blogspot.com/
> 
> _______________________________________________
> Pkg-ExpPsy-PyMVPA mailing list
> Pkg-ExpPsy-PyMVPA at lists.alioth.debian.org
> http://lists.alioth.debian.org/cgi-bin/mailman/listinfo/pkg-exppsy-pymvpa
> 



More information about the Pkg-ExpPsy-PyMVPA mailing list