[pymvpa] redundant voxels

kimberg at mail.med.upenn.edu kimberg at mail.med.upenn.edu
Thu Apr 1 17:51:12 UTC 2010


Yaroslav Halchenko replied:
> questions relevant to multi-variate analysis of neural data are
> always welcome to this list ;)

Thanks, I'll try not to take unfair advantage.

> do you mean smth like
> In [13]: remove_invariant_features?

As I understand it, removeInvariantFeatures just removes features that have no variance across samples.  The voxels I need to remove have different values for each sample, but neighboring voxels may have exactly the same values for each sample (these are actually 0/1 lesion maps from stroke patients, so it often happens that neighboring voxels will be lesioned in exactly the same subset of patients).  In any case, right now I do this filtering outside python for various trivial reasons.

All that said, this:

> I would expect no difference in error, iff (if and only if) you had
> "equivalent
> redundancy" among bogus and relevant features -- then it would indeed
> should not alter most of the classifiers.
[helpful example snipped]

does seem to be the right explanation.  There's obviously at least a numerical difference in the degree of redundancy between the most vs. least informative features, and it doesn't surprise me that it would vary a bit between ROIs.  To some extent, it's reassuring that the differences were so small, but in any case I think it does make more sense to prune redundant features.

Thanks for the quick reply.

dan



More information about the Pkg-ExpPsy-PyMVPA mailing list