Professional Documents
Culture Documents
http://www.fmri4newbies.com/
fMRI Analysis
with emphasis on the General Linear Model
raw time
course
HRF-
convolved
time course
Time (2 s volumes)
Alternation every 4 sec (2 volumes)
• signal amplitude is weakened by HRF because signal doesn’t have enough time
to return to baseline
• not to far from range of breathing frequency (every 4-10 sec) could lead to
respiratory artifacts
• if design is a task manipulation, subject is constantly changing tasks, gets
confused
Block Design: Short Unequal Epochs
raw time
course
HRF-
convolved
time course
Time (2 s volumes)
raw time
course
HRF-
convolved
time course
Time (2 s volumes)
Cardiac Cycle
• every ~1 sec (0.9 Hz)
• pulsing motion, blood changes
Solutions
• gating
• avoiding paradigms at those frequencies
raw time
course
HRF-
convolved
time course
Time (2 s volumes)
Every 16 sec (8 volumes)
• allows enough time for signal to oscillate fully
• not near artifact frequencies
• enough repetitions to see cycles by eye
• a reasonable time for subjects to keep doing the same thing
Block Design: Other Niceties
truncated
too soon
Time (2 s volumes)
• If you start and end with a baseline condition, you’re less likely
to lose information with linear trend removal and you can use
the last epoch in an event related average
Block Design Sequences: Three Conditions
• Suppose you want to add a third condition to act as a
more neutral baseline
• For example, if you wanted to identify visual areas as
well as object-selective areas, you could include resting
fixation as the baseline.
• That would allow two subtractions
– scrambled - fixation visual areas
– intact - scrambled object-selective areas
• That would also help you discriminate differences in
activations from differences in deactivations
As you can imagine, the more conditions you try to shove in a run, the thornier
ordering issues are and the fewer n you have for each condition.
But I have 8 conditions to compare!
• Just don’t.
Visual Stimuli
Baseline (blank screen with fixation point)
TR = 2 s/volume
Duration = 8 min, 44 s = 524 s
#Volumes = 262
Let’s Start with One Voxel in Occipital Cortex
One Occipital Voxel’s Time Course
Activation
(Raw)
Activation
(Raw)
Activation
(Raw)
…
Correlation Between %BSC and SW Predictor
How Can We Do Better? Use HRF
r = .40
16% of variance
p < 10-10
r > .40
GLM: 1 predictor
GLM: 1 predictor
1.2
0.8
0.6
0.4
0.2
0
1 8 15 22 29 36 43 50 57 64 71 78 85 92 99 06 13 20 27 34 41 48 55 62 69 76 83 90 97 04 11 18 25 32 39 46 53 60
1 1 1 1 1 1 1 1 1 1 1 1 1 1 2 2 2 2 2 2 2 2 2
-0.2
-0.4
We have 1 degree of freedom here
• Adjust the height of the predictor function to match the data
4
3
2
1
0
-1
-2
-3
-4
-5
The beta weight is NOT a correlation
• correlations measure goodness of fit regardless of scale
• beta weights are a measure of scale
small ß small ß
large r small r
large ß large ß
large r small r
Brain Voyager’s Model
Brain Voyager’s Output
Our model
accounts for
Our model had
variance of
only 1 df
635
In a model for a
single subject,
total df =
volumes - 1 Our model
accounts for
R2 = 0.92 =
81% of the
variance
635/784
= 81%
Brain Voyager’s Output
F test
F = MSsignal/Msnoise
F = 635/0.576 = 1102
Look up t =
33.2
with 260 df
p<.
000001
se is an estimate of
noise for our beta
t = signal/noise
Remember our 1 df (height of predictor)
= β/se
This is it – our β
= 3.591/0.108
= 33.2
Comparison
Correlation GLM
• Model
– statistical model
• Linear
– things add up sensibly (1+1 = 2)
• note that linearity refers to the predictors in the
model and not necessarily the BOLD signal
• General
– many simpler statistical procedures such as
correlations, t-tests and ANOVAs are subsumed by
the GLM
A More Complex Design
• Actually, we had more conditions
• There were multiple categories of visual stimuli
Houses
Faces
Objects
Bodies
Scrambled Images
Now we have 5 df
• Now we have 5 degrees of freedom
Bodies
Scrambled
Images
Let’s look at another voxel (in PPA)
Our Second Voxel’s Data
Our Second Voxel’s Model
1 x βHouses
-1 x βFaces
0 x βObjects
0 x βBodies
0 x βScrambled Images
Contrast
a vector is just a row (or column) of numbers
Vectors
But are Houses > Faces?
1 -1 0 0 0
Houses – Faces
βHouses – βFaces
= 1.031 – 0.147
= 0.884
Is this Difference Significant?
se = noise estimate
for contrast
t = signal/noise
=0.884/0.109
= 8.075
Look up t (df=260)
= 8.075
p < .000001
Simple Example Experiment: LO Localizer
Lateral Occipital Complex
• responds when Blank
subject views objects
Screen
TIME
Intact Scrambled
Objects (Unit: Volumes) Objects
× 1
= + +
× 2
Intact Scrambled
Predictor Predictor
• SPM represents time as going down
• SPM represents predictors within the design matrix as grayscale plots (where black = low,
white = high) over time
• GLM includes a constant to take care of the average activation level throughout each run
– SPM shows this explicity (BV may not)
We create a GLM with 2 predictors
when 1=2
= + +
when 2=0.5
raw
data
high-
pass
low-
pass
band-
pass
z = -20
• We can also examine whether a single predictor is
significantly greater than another predictor
Contrast Vectors
Faces - Baseline 0 +1 0 0 0
Faces - Houses -1 +1 0 0 0
Faces - Objects 0 +1 -1 0 0
Faces - Bodies 0 +1 0 -1 0
Faces - Scrambled 0 +1 0 0 -1
Balanced Contrasts
1 2 1 1 1
Condition
Unbalanced Balanced
Contrast -1 +1 -1 -1 -1 Σ=-3 Contrast -1 +4 -1 -1 -1 Σ=-0
β 1 2 1 1 1 β 1 2 1 1 1
If you do not balance the contrast, you are If you balance the contrast, you are comparing one
comparing one condition vs. the sum of all the others condition vs. the average of all the others
Problems with Bulk Contrasts
β β
1 2 1 1 1 2 2 2 2 .5
Condition Condition
Faces - Baseline 0 +1 0 0 0
AND
Faces - Houses -1 +1 0 0 0
Faces - Objects 0 +1 -1 0 0
AND
AND
Faces - Bodies 0 +1 0 -1 0
AND
Faces - Scrambled 0 +1 0 0 -1
Superimposed Conjunction
Maps
P Values for Conjunctions
• If the contrasts are independent:
• e.g., [(Faces > Houses) AND (Scrambled > Baseline)]
– pcombined = (psinglecontrast)numberofcontrasts
dfpredictors = # of predictors
dfresidual = dftotal - dfpredictors
dftotal = #volumes - 1
262 volumes (time points)
t = β/se
e.g., tFace = βFace/seFace
tFace = 1.371/0.076 = 18.145
Σ[Contrast x β] = 0 x 0.964
+ 1 x 1.371
+ 0 x 0.979
+ 0 x 1.000
- 1 x 0.687
= 1.371 – 0.687
= 0.684
Dealing with Faulty Assumptions
What’s this #*%&ing reviewer
complaining about?!
1. Correction for multiple comparisons
2. Correction for serial correlations
– only necessary for data from single subjects
– not necessary for group data
Types of Errors
Yes No
Does our stat test indicate
p value:
that the region is active?
• 130,000 voxels
• no correction for
multiple
comparisons
Fishy Headlines
Mega-Multiple Comparisons Problem
Typical 3T Data Set
30 slices x 64 x 64
= 122,880 voxels of (3 mm)3
We can reduce this number by only examining voxels inside the brain
• Drawbacks:
• handicaps small regions (e.g., subcortical foci) more than large
regions
• researcher can test many combinations of p values and k values
and publish the one that looks the best
False Discovery Rate
• “controls the proportion of rejected hypotheses that are falsely rejected”
(Type II errors)
• standard p value (e.g., p < .01) means that a certain proportion of all
voxels will be significant by chance (1%)
• FDR uses q value (e.g., q < .01), meaning that a certain proportion of the
“activated” (colored) voxels will be significant by chance (1%)
• Drawbacks
• very conservative when there is little activation; less conservative when
there is a lot of activation
Gaussian Random Field Theory
• Fundamental to SPM
• If data are very smooth, then the chance of noise points passing
threshold is reduced
• Can correct for the number of “resolvable elements” (“resels”) rather
than number of voxels
Yes No
Does our stat test indicate
that the region is active?
HIT Type I
Yes
Error
Type II Correct
No
Error Rejection
simulated
data
uncorrected
-high Type I
-low Type II
Bonferroni
-low Type I
-high Type II
FDR
-low Type I
-low Type II
• Inductive approach
– Run a few subjects to see if you’re on the right track
– Spend a lot of time exploring the pilot data for
interesting patterns
– “Find the story” in the data
– You may even change the experiment, run additional
subjects, or run a follow-up experiment to chase the
story
Even in a “screen saver scan”, activation in a voxel at one time is correlated with it’s
activation within ~6 sec
time
BEFORE
AFTER
BV Preprocessing Options
Temporal Smoothing of Data
• We have the option in our software to temporally
smooth our data (i.e., remove high temporal
frequencies or “low-pass filter”)
• However, I recommended that you not use this option
images from
O’Reilly et al.,
2012, SCAN
Intact
Objects
Scrambled Objects
Example of ROI Approach
Culham et al., 2003, Experimental Brain Research
Does the Lateral Occipital Complex compute object shape for grasping?
Grasping
Reaching
NS NS
p = .35 p = .31
Example of ROI Approach
Very Simple Stats
% BOLD Signal Then simply do a
Change paired t-test to see
Left Hem. LOC
whether the peaks are
Subject Grasping Reaching significantly different
between conditions
1 0.02 0.03
6 0.16 0.09
7 0.19 0.12
• Instead of using % BOLD Signal Change, you can use beta weights
• You can also do a planned contrast in Brain Voyager using a module
called the ROI GLM
Example: The Danger of ROI Approaches
• Example 1: LOC may be a heterogeneous area with subdivisions; ROI
analyses gloss over this
• Example 2: Some experiments miss important areas (e.g., Kanwisher
et al., 1997 identified one important face processing area -- the fusiform
face area, FFA -- but did not report a second area that is a very
important part of the face processing network -- the occipital face area,
OFA -- because it was less robust and consistent than the FFA.
Pros and Cons: Voxelwise Approach
Benefits
• Require no prior hypotheses about areas involved
• Include entire brain
• May identify subregions of known areas that are implicated in a
function
• Doesn’t require independent data set
Drawbacks
• Requires conservative corrections for multiple comparisons
• vulnerable to Type II errors
• Neglects individual differences in brain regions
• poor for some types of studies (e.g., topographic areas)
• Can lose spatial resolution with intersubject averaging
• Requires speculation about areas involved
Pros and Cons: ROI Approach
Benefits
• Extraction of ROI data can be subjected to simple stats
• Elimination of mega multiple comparisons problem greatly improves
statistical power (e.g., p < .05)
• Hypothesis-driven
• Useful when hypotheses are motivated by other techniques (e.g.,
electrophysiology) in specific brain regions
• ROI is not smeared due to intersubject averaging
• Important for discriminating abutting areas (e.g., V1/V2)
• Easy to analyze and interpret
• Can be useful for dissecting factorial design data in an unbiased manner
Drawbacks
• Neglects other areas that may play a fundamental role
• If multiple ROIs need to be considered, you can spend a lot of scan time
collecting localizer data (thus limiting the time available for experimental
runs)
• Works best for reliable and robust areas with unambiguous definitions
• Sometimes you can’t find an ROI in some subjects
• Selection of ROIs can be highly subjective and error-prone
A Proposed Resolution
• There is no reason not to do BOTH ROI analyses and
voxelwise analyses
– ROI analyses for well-defined key regions
– Voxelwise analyses to see if other regions are also involved
• Ideally, the conclusions will not differ
• If the conclusions do differ, there may be sensible reasons
– Effect in ROI but not voxelwise
• perhaps region is highly variable in stereotaxic location between subjects
• perhaps voxelwise approach is not powerful enough
– Effect in voxelwise but not ROI
• perhaps ROI is not homogenous or is context-specific
The War of Non-Independence
Finding the Obvious
A priori probability of getting JQKA
sequence = (1/13)4 = 1/28,561
Non-independence error
• occurs when statistical tests performed are not independent
from the means used to select the brain region
social exclusion
> inclusion
“Voodoo Correlations”
The original title of the paper
was not well-received by
reviewers so it was changed
even though some people still
use the term
Voodoo
2009
"Notably, 53% of the surveyed studies selected voxels based on a correlation with the
behavioral individual-differences measure and then used those same data to compute a
correlation within that subset of voxels."