[pymvpa] Mac memory issues w/ large datasets
James M. Hughes
james.m.hughes at dartmouth.edu
Sun Oct 17 22:33:36 UTC 2010
Just one more question:
I have a nifti file that has 5 dimensions (it's 10 x 1 x 80 x 80 x 33)
instead of 4. Is there anyway I can force it to ignore that second
dimension or get rid of it later on?
It's messing up searchlight, it seems, because the mapper thinks that
the individual datapoints are in 4 dimensions instead of 3.
On Oct 15, 2010, at 11:34 PM, Yaroslav Halchenko wrote:
> well... everything, sure thing depends on how much RAM you have in
> general ;-) otherwise I am not sure what is the exact 'issue' ;)
> and regardless either mask was applied, you better load with brain
> just to avoid wasting memory/cpu on searchlighting outside (although,
> then you would fail to discover the throat ventral pathway as it was
> done in 0.5 tutorials ;))
> On Fri, 15 Oct 2010, James M. Hughes wrote:
>> I'm having some trouble creating a dataset using a Nifti file that
>> is approximately 1.6 GB in size on Mac OS 10.5 using python 2.5 and
>> pymvpa 0.4.2.
>> Previously I had been importing using a mask which signifianctly
>> reduced the size and so it was feasible, but it seems that Python's
>> memory limits are keeping me from importing the dataset w/ no mask
>> (since I want to do a searchlight).
>> I'm not sure whether a brain/non-brain mask has already been
>> applied, but this seems like one possible solution.
>> Aside from that, has anyone faced this issue before? Does anyone
> =------------------------------ /v\ ----------------------------=
> Keep in touch // \\ (yoh@|www.)onerussian.com
> Yaroslav Halchenko /( )\ ICQ#: 60653192
> Linux User ^^-^^ 
> Pkg-ExpPsy-PyMVPA mailing list
> Pkg-ExpPsy-PyMVPA at lists.alioth.debian.org
More information about the Pkg-ExpPsy-PyMVPA