[pymvpa] Q: Time-offset of category labels in Haxby 2001 dataset. Was this ever properly resolved?

Yaroslav Halchenko debian at onerussian.com
Tue Mar 8 17:12:01 UTC 2011


Hi Raj,

sorry -- I cannot reincarnate in my memory exactly why there is a
mismatch of 121 physically present volumes I got per each run from Jim some
time ago (about 5-6 years) and 120 as paper description suggests (besides that
for the run length of 300sec, if you want to gather activation at t=300sec, you
would need 1 more volume to acquire, since collection for the previous one ends
at 299.whatever).

You can try asking Jim of cause as the ultimate source of information ;)

Nifti files we distribute are the reconstruction of the sequence from a set of
Analyze file pairs and text files listing which Analyze volumes belonged to
which category.  In those text files, each block had only 7 volumes listed per
each condition block (7*2.5 = 17.5sec), so, my blunt guess was that Jim
has discarded 1 volume from each side of the block, so those were appended back
to each block while I was reconstructing the sequence of stimuli hoping to get
closer to the original stimuli sequence which could not be regular in terms of
the volumes anyways due to 24 sec(condition) + 12 sec (rest) = 36 = 14
(volumes) *2.5 sec(TR) + 1sec.  that is why in labels.txt you would see varying
number of volumes for 'rest' condition (6 or 5).

$> uniq -c ~pymvpa/pymvpa/datadb/haxby2001/subj1/labels.txt | head -10
      1 labels chunks
      6 rest 0
      9 scissors 0
      6 rest 0
      9 face 0
      5 rest 0
      9 cat 0
      5 rest 0
      9 shoe 0
      5 rest 0

and first initial rest block in each chunk having 6 volumes (I think).  So, if
we forget for a moment that we got 121 volumes instead of 120, then initial
rest block should have indeed been allocated  5 volumes (although once again
not exactly, due to 12/2.5 = 4.8) instead of 6 volumes.  Thus you could
probably consider indeed that all labels in the dataset are shifted by 2
volumes, and I should have not added 1 volume on both sides of a stimuli block,
but rather 2 on the front, if I was looking for "precise" timing in terms of
volumes.   But then you would obtain 7 rest volumes at the end of each run
(121st volume?) That would be my speculation atm

hope this helps

Cheers,
Yarik

On Sun, 06 Mar 2011, Rajeev Raizada wrote:

> Dear PyMVPA folks,

> A while ago on this list there was some discussion
> of the question of whether the category-labels
> attached to each TR in the Haxby 2001 data-set:
> http://dev.pymvpa.org/datadb/haxby2001.html
> correspond to either:
> 1. The times at which the stimuli were presented to the subjects
> or
> 2. The times at which the subjects' HRFs are showing responses to those stimuli.

> Here's the final element in that discussion thread:
> http://lists.alioth.debian.org/pipermail/pkg-exppsy-pymvpa/2009q1/000321.html

> Was this question ever properly resolved?

> I am planning to submit a paper with some new analyses of this dataset,
> so I want to make extra-sure that I am using the right labels.

> Comparing the dataset with the methods-description
> in the 2001 Science paper itself, here's what I can glean so far:

> The data has a TR of 2.5s,
> with 8 stimulus-blocks of 24s each, and 9 rest periods of 12s each
> (one rest period after each block, and another one at the start of each run).

> That makes each run last 300s:
> Blocks = 8*24 = 192s
> Rest = 9*12 = 108s

> 300secs is 120 TRs:
> 300 / 2.5 = 120

> However, in the dataset at
> http://data.pymvpa.org/datasets/haxby2001
> each run has 121 TRs.

> In each of those runs, the first 6 TRs are labeled as rest,
> which makes a 6*2.5 = 15secs long rest period.
> However, as mentioned above, the 2001 paper says that
> each rest period, including the one at the start of each run, was 12s.
> 12 seconds would actually be 4.8 TRs.

> This suggests to me (but it certainly doesn't guarantee it!)
> that the downloadable data in effect have one rest-period TR
> pre-attached at the beginning of each run,
> making, in effect, a time-offset of 2.5s.

> When taking into account the HRF-delay in assigning category-labels to TRs,
> I personally code it such that a category gets listed as occurring when the
> HRF-convolved stimulus-time-series for that category is above its
> run-mean value.

> For a TR of 2.5s, this ends up in effect being an offset of one TR.

> So, my current guess is that the fact that the downloadable Haxby data
> appear to have one extra label of rest inserted at the start of each run
> means that, in effect, the labels already have the correct offsets
> for corresponding to evoked BOLD responses, rather than raw stimulus-onsets.

> I am not sure about this, though, and would love to hear other
> people's thoughts.
> In particular, one key question is left unaddressed:
> the 2001 Science paper says that each run had 120 TRs in it,
> but the downloadable data has 121 TRs in each run.
> Where did the extra TR's worth of data come from?
> The best-case scenario would be that the first TR in each run,
> which gets assigned the label "rest" and which in effect offsets all
> the label-times,
> is a dummy-scan volume that was collected 2.5s before the
> moment that counted as t=0 for the stimulus-presentation timings.
> However, I have no idea if that is the case or not.

> Any help greatly appreciated.

> Raj

> _______________________________________________
> Pkg-ExpPsy-PyMVPA mailing list
> Pkg-ExpPsy-PyMVPA at lists.alioth.debian.org
> http://lists.alioth.debian.org/mailman/listinfo/pkg-exppsy-pymvpa


-- 
=------------------------------------------------------------------=
Keep in touch                                     www.onerussian.com
Yaroslav Halchenko                 www.ohloh.net/accounts/yarikoptic



More information about the Pkg-ExpPsy-PyMVPA mailing list