Upload
others
View
0
Download
0
Embed Size (px)
Citation preview
Supplementary Materials
Altered Topological Organization of High-Level Visual Networks in Alzheimer’s
Disease and Mild Cognitive Impairment
Yanjia Deng1, Lin Shi2,4, Defeng Wang1,3,#, for the Alzheimer’s Disease Neuroimaging
Initiative*
# Corresponding author: Defeng Wang; Tel: +852- 26975027; Fax: +852- 26493761;
Email: [email protected]
I.ALE analysis for locating “what” and “where” visual cortices
1) Literature selection criteria
We searched the studies that investigated HLV functions in the Web of Pubmed
database (http://www.ncbi.nlm.nih.gov/pubmed). For more specific and effective
search, the MeSH (medical subject headings) entry terms [1] were used, including
"Visual Perception", "Color Perception", "Motion Perception", "Space Perception",
"Depth Perception", "Form Perception", "Size Perception" and "Magnetic Resonance
Imaging".
Until February 2015, this search revealed 3095 published, peer reviewed papers.
Finally, 74 studies matched the following criteria and were included for ALE
analysis: (1) Papers were published in English and in which subjects were healthy
adults; (2) fMRI was adopted as image modality; (3) The whole brain was scanned
and the coordinates of the activation clusters were reported in Talairach and Tournoux
(1988) or Montreal Neurological Institute (MNI) space; (4) Only one visual functions
was tested in one task; (5) Results were obtained results by using low level contrasts
(target vision vs. baseline, noise or scrambled meaningless images).
2) Data analysis
Among the included 74 studies, studies on recognition perception (including face,
object, alphabetic word/ letter, scene, and body vision) were grouped into “what”
vision, and those focused on motion or spatial vision were categorized into “where”
vision (Supplementary Table 1). The meta-analyses were performed using the revised
algorithm of the ALE implemented in the Ginger ALE software, version 2.0
(available at www.brainmap.org/ale). Coordinates reported in the Talairach space
were transformed into the MNI space using the Lancaster transform [2]. Foci from
each individual study were smoothed by a full width half maximum (FWHM) value
scaled by the study’s sample size to model the uncertainty in spatial location of the
activations [3]. The statistical level to determine the null distribution of the ALE value
was at p < 0.01 (multiple comparisons corrected by false discovery rate (FDR pN)
method) [3].
II. Subjects and neuropsychological tests
A total of 162 subjects, including 44 normal controls (CN), 52 early MCI (EMCI), 35
late MCI (LMCI), 31 AD, were included in this study. Briefly, the cognitive level (i.e.
CN, EMCI, LMCI or AD) was judged by combining the memory complaints of
patients or their families, the Wechsler Memory Scale - Logical Memory II
(Wechsler, 1987), the Mini-Mental State Exam (MMSE) scores, and the Clinical
Dementia Rating (CDR) of subjects. Specifically, the major difference of the
diagnostic criteria for EMCI and LMCI is in the score of Logical Memory II subscale.
The score of Logical Memory II subscale for defining EMCI is: a. 9-11 for 16 or more
years of education; b. 5-9 for 8-15 years of education; c. 3-6 for 0-7 years of
education, which is between the scale for CN and LMCI. Therefore, the EMCI and
LMCI refer to the severity of cognitive impairments. The detailed inclusion and
exclusion criteria were found in the ADNI2 and ADNI GO protocol
(http://adni.loni.usc.edu/methods/documents/). For the current study, subjects with
glaucoma or congenital blindness were further excluded. Besides, one AD subjects
was excluded due to excessive head motion during acquisition of rs-fMRI (see data
preprocess section). Finally, there were 44 CN, 52 EMCI, 35 LMCI and 30 AD
subjects.
All participants were administered the Mini Mental State Examination (MMSE) [4].
The MMSE is a widely used general cognitive screening test. In the ADNI database,
the clock drawing test was performed through the ADNI program, which has two
components: a command condition in which the participant draws a clock to verbal
instructions, and a copy condition in which the participant copies a model clock
drawn at the top of response form. This test, especially the copy condition, is
considered correlated with the visuospatial perception [5]. Thus, we included the
results of copy conditions of clock drawing test for further analysis.
III. Network construction and calculations
Previously established network quantification method based on graph theory was used
in this study to calculate the network parameters [6]. Firstly, the association matrices
based on the pair-wise interregional RSFC at 0.021-0.045Hz were created
respectively for “what” (29 × 29) and “where” (25 × 25) vision networks of each
subject. Then the matrices were thresholded into binary matrices. Here, a range of
network densities (0.06 – 0.40, with an interval of 0.01) were used to binarize the
matrics to ensure that the resultant networks of different subject could have the same
number of edges thus to make the networks comparable. Therefore, the binarized
matrix was recognized as a graph, each ROI as a node, and the RSFC between two
nodes as an edge.
Two fundamental quantities, the characteristic path length (L) and clustering
coefficient (C), were used in this study to respectively quantify the capability of
integrated and segregated information processing of the brain network. The L is
defined as the average of the shortest path lengths between nodes [7].
Here N represents the total number of the nodes, i and j stand for two generic nodes,
and Lij is the shortest path length between i and j. Thus a shorter path length
represents greater network integration. The C is the average of the local clustering
coefficients of all nodes which are calculated by dividing the number of existing
connections among the node and its neighbors by all the possible connections.
,
Here G and Gi respectively represent the whole network and the sub-network which is
composed of a node i and its nearest neighbors (e.g., node j and k). Thus the C
quantifies the extent to which neighboring brain regions are connected with each
other, i.e. the network segregation [8].
IV. ADNI dataset
The ADNI was launched in 2003 as a public-private partnership, led by Principal
Investigator Michael W. Weiner, MD. The primary goal of ADNI has been to test
whether serial MRI, positron emission tomography (PET), other biological markers,
and clinical and neuropsychological assessment can be combined to measure the
progression of MCI and AD.
ADNI is funded by the National Institute on Aging, the National Institute of
Biomedical Imaging and Bioengineering, and through generous contributions from
the following: AbbVie, Alzheimer’s Association; Alzheimer’s Drug Discovery
Foundation; Araclon Biotech; BioClinica, Inc.; Biogen; Bristol-Myers Squibb
Company; CereSpir, Inc.; Eisai Inc.; Elan Pharmaceuticals, Inc.; Eli Lilly and
Company; EuroImmun; F. Hoffmann-La Roche Ltd and its affiliated company
Genentech, Inc.; Fujirebio; GE Healthcare; IXICO Ltd.; Janssen Alzheimer
Immunotherapy Research & Development, LLC.; Johnson & Johnson Pharmaceutical
Research & Development LLC.; Lumosity; Lundbeck; Merck & Co., Inc.; Meso
Scale Diagnostics, LLC.; NeuroRx Research; Neurotrack Technologies; Novartis
Pharmaceuticals Corporation; Pfizer Inc.; Piramal Imaging; Servier; Takeda
Pharmaceutical Company; and Transition Therapeutics. The Canadian Institutes of
Health Research is providing funds to support ADNI clinical sites in Canada. Private
sector contributions are facilitated by the Foundation for the National Institutes of
Health (www.fnih.org). The grantee organization is the Northern California Institute
for Research and Education, and the study is coordinated by the Alzheimer's Disease
Cooperative Study at the University of California, San Diego. ADNI data are
disseminated by the Laboratory for Neuro Imaging at the University of Southern
California.
Supplementary Tables
Supplementary Table 1 Summary of studies selected for the ALE analysis of
"what" and "where" vision
Category/study(author) Year No. of
Subjects
Contrast No. of
Foci
“What” vision
Leube et al. [9] 2003 13 Face > baseline 5
Badzakova-Trajkov et al. [10] 2010 155 Face > baseline 18
Love et al. [11] 2011 20 Face > baseline 10
Vuilleumier et al. [12] 2003 13 Face > baseline 11
Henson et al. [13] 2003 21 Face > scrambled image 4
Furl et al. [14] 2007 12 Face > baseline 7
Fruhholz et al. [15] 2011 24 Face > scrambled image 21
Platek et al. [16] 2009 12 Face > baseline 1
Schwartz et al. [17] 2013 8 Face > baseline 3
Joseph et al. [18] 2011 43 Face > baseline 11
Jehna et al. [19] 2011 30 Face > scrambled image 5
Benuzzi et al. [20] 2007 24 Face > noise 36
Rossion et al. [21] (2003) 2003 11 Face > baseline 4
Rossion et al. [22] 2012 40 Face > scrambled 12
Gobbini et al. [23] 2011 12 Face > baseline 30
Hoffman et al. [24] 2000 9 Face > baseline 8
Pagliaccio et al. [25] 2013 52 Face > baseline 1
Ramon et al. [26] 2010 13 Face > scrambled image 13
Blonder et al. [27] 2004 14 Face > baseline 6
Liu et al. [28] 2009 17 Face > baseline 11
Orlov et al. [29] 2010 12 Face > baseline 10
Olulade et al. [30] 2013 15 Word> baseline 8
Joseph et al. [31] 2006 11 Word > baseline 7
Minati et al. [32] 2008 10 Word > baseline 15
Szwed et al. [33] 2011 16 Word > scrambled image 9
Longcamp et al. [34] 2011 10 Word > baseline 1
Pegado et al. [35] 2011 14 Letter > baseline 4
Dehaene et al. [36] 2001 15 Word > noise 6
Dehaene et al. [37] 2004 26 Word > noise 6
Casarotto et al. [38] 2006 8 Word > baseline 4
Peng et al. [39] (2003) 2003 8 Word > baseline 34
Gold et al. [40] 2007 16 Word > baseline 2
Gros et al. [41] 2001 6 Letter > baseline 2
James et al. [42] 2009 19 Letter > baseline 12
Thoma et al. [43] 2011 17 Object > scrambled image 8
Pegado et al. [35] 2011 14 Object > baseline 6
Yamamoto et al. [44] 2008 6 Object > baseline 3
Podzebenko et al. [45] 2005 16 Object > baseline 13
Minati et al. [32] 2008 10 Object > baseline 23
Szwed et al. [33] 2011 16 Object > scrambled image 7
Creem-Regehr et al. [46] 2007 23 Object > scrambled image 10
Durand et al. [47] 2009 17 Object > scrambled image 6
Cardin et al. [48] 2011 14 Object > baseline 9
Valyear et al. [49] 2006 7 Object > baseline 1
Niemeier et al. [50] 2005 10 Object > baseline 2
Casarotto et al. [38] 2006 8 Object > baseline 1
Peelen et al. [51] 2012 26 Object > baseline 2
Bucher et al. [52] 2006 16 Object > baseline 7
Blonder et al. [27] 2004 14 Object > baseline 6
Amemiya et al. [53] 2012 27 Scene > baseline 10
Jehna et al. [19] 2011 30 Scene > baseline 4
Blonder et al. [27] 2004 14 Scene > baseline
Claeys et al. [54] 2004 16 Color > baseline 10
Leh et al. [55] 2009 1 Color > baseline 4
Kaufmann et al. [56] 2008 12 Color > baseline 1
Orlov et al. [29] 2010 12 Body > baseline 18
“Where” vision
Homola et al. [57] 2012 24 Motion > baseline 2
Schraa-Tam et al. [58] 2008 22 Motion > baseline 12
Morito, Tanabe et al. [59] 2009 28 Motion > baseline 17
Caplan et al. [60] 2006 12 Motion > baseline 16
Katsuyama et al. [61] 2011 31 Motion > baseline 3
Podzebenko et al. [45] 2005 16 Motion > baseline 9
Braddick et al. [62] 2001 3 Motion > noise 14
Braddick et al. [63] 2000 4 Motion > baseline 4
Oreja-Guevara et al. [64] 2004 9 Motion > baseline 61
Bucher et al. [52] 2006 16 Motion > baseline 12
Santi et al. [65] 2003 10 Motion > baseline 14
Aso et al. [66] 2007 12 Motion > baseline 5
Iwami et al. [67] 2002 10 Spatial > baseline 11
Grady et al. [68] 2014 14 Spatial > baseline 11
Joseph et al. [69] 2003 10 Spatial > baseline 4
Joseph et al. [69] 2003 11 Spatial > baseline 5
Joseph et al. [69] 2003 10 Spatial > baseline 3
Joseph et al. [69] 2003 11 Spatial > baseline 3
Zeidman et al. [70] 2012 19 Spatial > baseline 2
Yamamoto et al. [44] 2008 13 Spatial > baseline 30
Katsuyama et al. [61] 2011 31 Spatial > baseline 7
Fraedrich et al. [71] 2010 18 Spatial > baseline 6
Georgieva et al. [72] 2008 18 Spatial > scrambled image 12
Creem-Regehr et al. [46] 2007 23 Spatial > scrambled image 6
Kana et al. [73] 2013 14 Spatial > baseline 11
Wu et al. [74] 2012 19 Spatial > baseline 9
Valyear et al. [49] 2006 3 Spatial > baseline 1
Braddick et al. [63] 2000 4 Spatial > baseline 8
Kaufmann et al. [56] 2008 12 Spatial > baseline 3
Sterzer et al. [75] 2005 12 Spatial > baseline 4
Brouwer et al. [76] 2005 7 Spatial > baseline 7
Cant et al. [77] 2007 9 Spatial > baseline 10
Schoth et al. [78] 2007 22 Spatial > baseline 12
Dumoulin et al. [79] 2004 4 Spatial > noise 2
Baecke et al. [80] 2009 26 Spatial > baseline 6
Gros et al. [41] 2001 6 Spatial > baseline 2
Ritzl et al. [81] 2003 11 Spatial > baseline 10
Negawa et al. [82] 2003 13 Spatial > scrambled image 13
Supplementary Table 2 Activation likelihood estimation results of “what” vision
Cluster
No.
x y z Anatomic region (AAL) Broadmann area
1 -40 -52 -20 Fusiform_L* 37
-44 -76 -6 Occipital_Mid_L* 19
-46 -68 -14 Temporal_Inf_L 19
-28 -84 -16 Occipital_Inf_L* 18
-30 -96 6 Occipital_Mid_L* 17
2 40 -48 -22 Fusiform_R* 37
48 -74 -6 Temporal_Mid_R* 19
42 -62 -14 Temporal_Inf_R 37
26 -66 -10 Fusiform_R 19
52 -64 6 Temporal_Mid_R 37
3 44 10 28 Frontal_Inf_Tri_R* 44
4 52 -40 8 Temporal_Sup_R* 21
5 30 -92 2 Occipital_Mid_R* 18
6 20 -6 -16 Hippocampus_R* -
7 -46 4 30 Frontal_Inf_Oper_L* 44
8 -32 24 0 Insula_L* 47
9 -20 -8 -14 Amygdala_L* -
10 12 -92 6 Occipital_Sup_L* 17
11 -56 -46 10 Temporal_Sup_L* 22
12 -10 -96 2 Occipital_Sup_L* 17
13 -26 -70 -10 Fusiform_L* 18
14 -22 -100 -2 Occipital_Mid_L* 17
15 22 -30 0 Thalamus_R* -
16 -38 -46 48 Parietal_Inf_L* 40
17 32 24 -2 Frontal_Inf_Orb_R* 47
18 48 2 52 Precentral_R* 6
19 -42 18 22 Frontal_Inf_Tri_L* 48
20 -26 -62 62 Parietal_Sup_L* 7
21 32 -56 52 Parietal_Inf_R* 7
22 -28 -70 46 Parietal_Sup_L* 7
23 56 -52 14 Temporal_Sup_R* 21
24 0 22 44 Frontal_Sup_Medial_L
*
32
25 -18 -68 -16 Lingual_L* 18
Statistical level: p< 0.01; coordinates are in MNI spaces
* coordinate chosen for defining the nodes of “where” vision.
L: left; R: right; Inf: inferior; Mid: middle; Sup: superior.
Supplementary Table 3 Activation likelihood estimation results of “where” vision
Cluster
No.
x y z Anatomical region (AAL) Broadmann area
1 -44 -72 -12 Occipital_Inf_L* 19
-46 -72 4 Temporal_Mid_L* 37
-42 -84 4 Occipital_Mid_L 19
-38 -60 -12 Occipital_Inf_L 37
2 52 -68 4 Temporal_Mid_R* 37
44 -70 -12 Temporal_Inf_R* 19
42 -74 -4 Occipital_Mid_R 19
42 -84 2 Occipital_Mid_R 19
36 -84 -14 Occipital_Inf_R 19
3 30 -86 8 Occipital_Mid_R* 18
26 -76 34 Occipital_Sup_R* 19
34 -78 28 Occipital_Mid_R 19
36 -88 14 Occipital_Mid_R 19
32 -78 12 Occipital_Mid_R 19
26 -92 24 Occipital_Sup_R 18
4 22 -62 58 Parietal_Sup_R* 7
10 -62 60 Precuneus_R 7
28 -54 54 Parietal_Sup_R 7
5 -26 -56 60 Parietal_Sup_L* 7
6 30 4 60 Frontal_Mid_R* 8
7 40 -34 46 Postcentral_R* 2
8 -20 -96 12 Occipital_Sup_L* 17
9 -44 -48 56 Postcentral_L* 40
10 -44 -14 56 Precentral_L* 6
-36 -16 62 Precenral_L* 6
11 -26 -88 28 Occipital_Mid_L* 18
12 -42 -32 58 Postcentral_L* 3
13 -10 -22 46 Cingulum_Mid_L*
14 -38 -48 -22 Fusiform_L* 37
15 30 -58 -20 Fusiform_R* 37
16 -28 -70 42 Parietal_Sup_L* 7
17 40 -46 -26 Fusiform_R* 37
18 -24 -2 52 Frontal_Mid_L* 6
19 38 -44 60 Hippocampus_R* 2
20 20 -28 0 Thalamus_R*
21 -22 -8 58 Frontal_Sup_L* 6
22 -34 -44 52 Postcentral_L* 40
Statistical level: p< 0.01; coordinates are in MNI spaces
* coordinate chosen for defining the nodes of “where” vision.
L: left; R: right; Inf: inferior; Mid: middle; Sup: superior.
Supplementary figures
Supplementary figure 1. The ALE map of “what” vision (L, left; R, right).
Supplementary figure 2. The ALE map of “where” vision (L, left; R, right).
References:
[1] M.H. Coletti, H.L. Bleich, Medical subject headings used to search the
biomedical literature, J Am Med Inform Assoc 8 (2001) 317-323.
[2] J.L. Lancaster, D. Tordesillas-Gutierrez, M. Martinez, F. Salinas, A. Evans, K.
Zilles, J.C. Mazziotta, P.T. Fox, Bias between MNI and Talairach coordinates
analyzed using the ICBM-152 brain template, Hum Brain Mapp 28 (2007) 1194-
1205.
[3] A.R. Laird, P.M. Fox, C.J. Price, D.C. Glahn, A.M. Uecker, J.L. Lancaster, P.E.
Turkeltaub, P. Kochunov, P.T. Fox, ALE meta-analysis: controlling the false
discovery rate and performing statistical contrasts, Hum Brain Mapp 25 (2005) 155-
164.
[4] M.F. Folstein, S.E. Folstein, P.R. McHugh, “Mini-mental state”: A practical
method for grading the cognitive state of patients for the clinician, J Psychiat Res 12
(1975) 189-198.
[5] R.W. Parks, S.N. Thiyagesh, T.F. Farrow, L. Ingram, K. Wilkinson, M.D. Hunter,
I.D. Wilkinson, C. Young, P.W. Woodruff, Performance on the clock drawing task
correlates with FMRI response to a visuospatial task in Alzheimer's disease, Int J
Neurosci 120 (2010) 335-343.
[6] D.J. Watts, S.H. Strogatz, Collective dynamics of 'small-world' networks, Nature
393 (1998) 440-442.
[7] E.T. Bullmore, D.S. Bassett, Brain graphs: graphical models of the human brain
connectome, Annu Rev Clin Psychol 7 (2011) 113-140.
[8] M.R. Brier, J.B. Thomas, A.M. Fagan, J. Hassenstab, D.M. Holtzman, T.L.
Benzinger, J.C. Morris, B.M. Ances, Functional connectivity and graph theory in
preclinical Alzheimer's disease, Neurobiol Aging 35 (2014) 757-768.
[9] D.T. Leube, H.W. Yoon, A. Rapp, M. Erb, W. Grodd, M. Bartels, T.T. Kircher,
Brain regions sensitive to the face inversion effect: a functional magnetic resonance
imaging study in humans, Neurosci Lett 342 (2003) 143-146.
[10]G. Badzakova-Trajkov, I.S. Haberling, R.P. Roberts, M.C. Corballis, Cerebral
asymmetries: complementary and independent processes, PLoS One 5 (2010) e9682.
[11]S.A. Love, F.E. Pollick, M. Latinus, Cerebral correlates and statistical criteria of
cross-modal face and voice integration, Seeing Perceiving 24 (2011) 351-367.
[12]P. Vuilleumier, J.L. Armony, J. Driver, R.J. Dolan, Distinct spatial frequency
sensitivities for processing faces and emotional expressions, Nat Neurosci 6 (2003)
624-631.
[13]R.N. Henson, Y. Goshen-Gottstein, T. Ganel, L.J. Otten, A. Quayle, M.D. Rugg,
Electrophysiological and haemodynamic correlates of face perception, recognition
and priming, Cereb Cortex 13 (2003) 793-805.
[14]N. Furl, N.J. van Rijsbergen, A. Treves, R.J. Dolan, Face adaptation aftereffects
reveal anterior medial temporal cortex role in high level category representation,
Neuroimage 37 (2007) 300-310.
[15]S. Fruhholz, B. Godde, P. Lewicki, C. Herzmann, M. Herrmann, Face recognition
under ambiguous visual stimulation: fMRI correlates of "encoding styles", Hum Brain
Mapp 32 (2011) 1750-1761.
[16]S.M. Platek, S.M. Kemp, Is family special to the brain? An event-related fMRI
study of familiar, familial, and self-face recognition, Neuropsychologia 47 (2009)
849-858.
[17]B.L. Schwartz, C.J. Vaidya, D. Shook, S.I. Deutsch, Neural basis of implicit
memory for socio-emotional information in schizophrenia, Psychiatry Res 206 (2013)
173-180.
[18]J.E. Joseph, A.D. Gathers, R.S. Bhatt, Progressive and regressive developmental
changes in neural substrates for face processing: testing specific predictions of the
Interactive Specialization account, Dev Sci 14 (2011) 227-241.
[19]M. Jehna, C. Neuper, A. Ischebeck, M. Loitfelder, S. Ropele, C. Langkammer, F.
Ebner, S. Fuchs, R. Schmidt, F. Fazekas, C. Enzinger, The functional correlates of
face perception and recognition of emotional facial expressions as evidenced by
fMRI, Brain Res 1393 (2011) 73-83.
[20]F. Benuzzi, M. Pugnaghi, S. Meletti, F. Lui, M. Serafini, P. Baraldi, P. Nichelli,
Processing the socially relevant parts of faces, Brain Res Bull 74 (2007) 344-356.
[21]B. Rossion, R. Caldara, M. Seghier, A.M. Schuller, F. Lazeyras, E. Mayer, A
network of occipito-temporal face-sensitive areas besides the right middle fusiform
gyrus is necessary for normal face processing, Brain 126 (2003) 2381-2395.
[22]B. Rossion, B. Hanseeuw, L. Dricot, Defining face perception areas in the human
brain: a large-scale factorial fMRI face localizer analysis, Brain Cogn 79 (2012) 138-
157.
[23]M.I. Gobbini, C. Gentili, E. Ricciardi, C. Bellucci, P. Salvini, C. Laschi, M.
Guazzelli, P. Pietrini, Distinct neural systems involved in agency and animacy
detection, J Cogn Neurosci 23 (2011) 1911-1920.
[24]E.A. Hoffman, J.V. Haxby, Distinct representations of eye gaze and identity in
the distributed human neural system for face perception, Nat Neurosci 3 (2000) 80-84.
[25]D. Pagliaccio, J.L. Luby, M.S. Gaffrey, A.C. Belden, K.N. Botteron, M.P. Harms,
D.M. Barch, Functional brain activation to emotional and nonemotional faces in
healthy children: evidence for developmentally undifferentiated amygdala function
during the school-age period, Cogn Affect Behav Neurosci 13 (2013) 771-789.
[26]M. Ramon, L. Dricot, B. Rossion, Personally familiar faces are perceived
categorically in face-selective regions other than the fusiform face area, Eur J
Neurosci 32 (2010) 1587-1598.
[27]L.X. Blonder, C.D. Smith, C.E. Davis, M.L. Kesler-West, T.F. Garrity, M.J.
Avison, A.H. Andersen, Regional brain response to faces of humans and dogs, Brain
Res Cogn Brain Res 20 (2004) 384-394.
[28]J. Liu, J. Tian, J. Li, Q. Gong, K. Lee, Similarities in neural activations of face
and Chinese character discrimination, Neuroreport 20 (2009) 273-277.
[29]T. Orlov, T.R. Makin, E. Zohary, Topographic representation of the human body
in the occipitotemporal cortex, Neuron 68 (2010) 586-600.
[30]O.A. Olulade, D.L. Flowers, E.M. Napoliello, G.F. Eden, Developmental
differences for word processing in the ventral stream, Brain Lang 125 (2013) 134-
145.
[31]J.E. Joseph, M.A. Cerullo, A.B. Farley, N.A. Steinmetz, C.R. Mier, fMRI
correlates of cortical specialization and generalization for letter processing,
Neuroimage 32 (2006) 806-820.
[32]L. Minati, C. Rosazza, I. Zucca, L. D'Incerti, V. Scaioli, M.G. Bruzzone, Spatial
correspondence between functional MRI (fMRI) activations and cortical current
density maps of event-related potentials (ERP): a study with four tasks, Brain Topogr
21 (2008) 112-127.
[33]M. Szwed, S. Dehaene, A. Kleinschmidt, E. Eger, R. Valabregue, A. Amadon, L.
Cohen, Specialization for written words over objects in the visual cortex, Neuroimage
56 (2011) 330-344.
[34]M. Longcamp, Y. Hlushchuk, R. Hari, What differs in visual recognition of
handwritten vs. printed letters? An fMRI study, Hum Brain Mapp 32 (2011) 1250-
1259.
[35]F. Pegado, K. Nakamura, L. Cohen, S. Dehaene, Breaking the symmetry: mirror
discrimination for single letters but not for pictures in the Visual Word Form Area,
Neuroimage 55 (2011) 742-749.
[36]S. Dehaene, L. Naccache, L. Cohen, D.L. Bihan, J.F. Mangin, J.B. Poline, D.
Riviere, Cerebral mechanisms of word masking and unconscious repetition priming,
Nat Neurosci 4 (2001) 752-758.
[37]S. Dehaene, A. Jobert, L. Naccache, P. Ciuciu, J.B. Poline, D. Le Bihan, L.
Cohen, Letter binding and invariant recognition of masked words: behavioral and
neuroimaging evidence, Psychol Sci 15 (2004) 307-313.
[38]S. Casarotto, A.M. Bianchi, S. Cerutti, N. Vanello, E. Ricciardi, C. Gentili, L.
Sani, D. Bonino, M. Guazzelli, P. Pietrini, L. Landini, G.A. Chiarenza, Combination
of event-related potentials and functional magnetic resonance imaging during single-
letter reading, Conf Proc IEEE Eng Med Biol Soc 1 (2006) 984-987.
[39]D.L. Peng, D. Xu, Z. Jin, Q. Luo, G.S. Ding, C. Perry, L. Zhang, Y. Liu, Neural
basis of the non-attentional processing of briefly presented words, Hum Brain Mapp
18 (2003) 215-221.
[40]B.T. Gold, K. Rastle, Neural correlates of morphological decomposition during
visual word recognition, J Cogn Neurosci 19 (2007) 1983-1993.
[41]H. Gros, K. Boulanouar, G. Viallard, E. Cassol, P. Celsis, Event-related
functional magnetic resonance imaging study of the extrastriate cortex response to a
categorically ambiguous stimulus primed by letters and familiar geometric figures, J
Cereb Blood Flow Metab 21 (2001) 1330-1341.
[42]K.H. James, I. Gauthier, Letter processing automatically recruits a sensory-motor
brain network, Neuropsychologia 44 (2006) 2937-2949.
[43]V. Thoma, R.N. Henson, Object representations in ventral and dorsal visual
streams: fMRI repetition effects depend on attention and part-whole configuration,
Neuroimage 57 (2011) 513-525.
[44]T. Yamamoto, S. Takahashi, T. Hanakawa, S. Urayama, T. Aso, H. Fukuyama,
Y. Ejima, Neural correlates of the stereokinetic effect revealed by functional magnetic
resonance imaging, J Vis 8 (2008) 11-14.
[45]K. Podzebenko, G.F. Egan, J.D. Watson, Real and imaginary rotary motion
processing: functional parcellation of the human parietal lobe revealed by fMRI, J
Cogn Neurosci 17 (2005) 24-36.
[46]S.H. Creem-Regehr, V. Dilda, A.E. Vicchrilli, F. Federer, J.N. Lee, The influence
of complex action knowledge on representations of novel graspable objects: evidence
from functional magnetic resonance imaging, J Int Neuropsychol Soc 13 (2007) 1009-
1020.
[47]J.B. Durand, R. Peeters, J.F. Norman, J.T. Todd, G.A. Orban, Parietal regions
processing visual 3D shape extracted from disparity, Neuroimage 46 (2009) 1114-
1126.
[48]V. Cardin, K.J. Friston, S. Zeki, Top-down modulations in the visual form
pathway revealed with dynamic causal modeling, Cereb Cortex 21 (2011) 550-562.
[49]K.F. Valyear, J.C. Culham, N. Sharif, D. Westwood, M.A. Goodale, A double
dissociation between sensitivity to changes in object identity and object orientation in
the ventral and dorsal visual streams: a human fMRI study, Neuropsychologia 44
(2006) 218-228.
[50]M. Niemeier, H.C. Goltz, A. Kuchinad, D.B. Tweed, T. Vilis, A contralateral
preference in the lateral occipital area: sensory and attentional mechanisms, Cereb
Cortex 15 (2005) 325-331.
[51]M.V. Peelen, A. Caramazza, Conceptual object representations in human anterior
temporal cortex, J Neurosci 32 (2012) 15728-15736.
[52]K. Bucher, T. Dietrich, V.L. Marcar, S. Brem, P. Halder, S. Boujraf, P. Summers,
D. Brandeis, E. Martin, T. Loenneker, Maturation of luminance- and motion-defined
form perception beyond adolescence: a combined ERP and fMRI study, Neuroimage
31 (2006) 1625-1636.
[53]S. Amemiya, K. Ohtomo, Effect of the observed pupil size on the amygdala of
the beholders, Soc Cogn Affect Neurosci 7 (2012) 332-341.
[54]K.G. Claeys, P. Dupont, L. Cornette, S. Sunaert, P. Van Hecke, E. De Schutter,
G.A. Orban, Color discrimination involves ventral and dorsal stream visual areas,
Cereb Cortex 14 (2004) 803-822.
[55]S.E. Leh, A. Ptito, M. Schonwiesner, M.M. Chakravarty, K.T. Mullen, Blindsight
mediated by an S-cone-independent collicular pathway: an fMRI study in
hemispherectomized subjects, J Cogn Neurosci 22 (2010) 670-682.
[56]L. Kaufmann, S.E. Vogel, G. Wood, C. Kremser, M. Schocke, L.B.
Zimmerhackl, J.W. Koten, A developmental fMRI study of nonsymbolic numerical
and spatial processing, Cortex 44 (2008) 376-385.
[57]G.A. Homola, S. Jbabdi, C.F. Beckmann, A.J. Bartsch, A brain network
processing the age of faces, PLoS One 7 (2012) e49451.
[58]C.K. Schraa-Tam, A. van der Lugt, M.A. Frens, M. Smits, P.C. van Broekhoven,
J.N. van der Geest, An fMRI study on smooth pursuit and fixation suppression of the
optokinetic reflex using similar visual stimulation, Exp Brain Res 185 (2008) 535-
544.
[59]Y. Morito, H.C. Tanabe, T. Kochiyama, N. Sadato, Neural representation of
animacy in the early visual areas: a functional MRI study, Brain Res Bull 79 (2009)
271-280.
[60]J.B. Caplan, T.L. Luks, G.V. Simpson, M. Glaholt, A.R. McIntosh, Parallel
networks operating across attentional deployment and motion processing: a multi-
seed partial least squares fMRI study, Neuroimage 29 (2006) 1192-1202.
[61]N. Katsuyama, N. Usui, I. Nose, M. Taira, Perception of object motion in three-
dimensional space induced by cast shadows, Neuroimage 54 (2011) 485-494.
[62]O.J. Braddick, J.M. O'Brien, J. Wattam-Bell, J. Atkinson, T. Hartley, R. Turner,
Brain areas sensitive to coherent visual motion, Perception 30 (2001) 61-72.
[63]O.J. Braddick, J.M. O'Brien, J. Wattam-Bell, J. Atkinson, R. Turner, Form and
motion coherence activate independent, but not dorsal/ventral segregated, networks in
the human brain, Curr Biol 10 (2000) 731-734.
[64]C. Oreja-Guevara, R. Kleiser, W. Paulus, W. Kruse, R.J. Seitz, K.P. Hoffmann,
The role of V5 (hMT+) in visually guided hand movements: an fMRI study, Eur J
Neurosci 19 (2004) 3113-3120.
[65]A. Santi, P. Servos, E. Vatikiotis-Bateson, T. Kuratate, K. Munhall, Perceiving
biological motion: dissociating visible speech from walking, J Cogn Neurosci 15
(2003) 800-809.
[66]T. Aso, T. Hanakawa, K. Matsuo, K. Toma, H. Shibasaki, H. Fukuyama, T.
Nakai, Subregions of human parietal cortex selectively encoding object orientation,
Neurosci Lett 415 (2007) 225-230.
[67]T. Iwami, Y. Nishida, O. Hayashi, M. Kimura, M. Sakai, K. Kani, R. Ito, A.
Shiino, M. Suzuki, Common neural processing regions for dynamic and static
stereopsis in human parieto-occipital cortices, Neurosci Lett 327 (2002) 29-32.
[68]C.L. Grady, C.J. Mondloch, T.L. Lewis, D. Maurer, Early visual deprivation from
congenital cataracts disrupts activity and functional connectivity in the face network,
Neuropsychologia 57 (2014) 122-139.
[69]J.E. Joseph, A.D. Gathers, Effects of structural similarity on neural substrates for
object recognition, Cogn Affect Behav Neurosci 3 (2003) 1-16.
[70]P. Zeidman, S.L. Mullally, D.S. Schwarzkopf, E.A. Maguire, Exploring the
parahippocampal cortex response to high and low spatial frequency spaces,
Neuroreport 23 (2012) 503-507.
[71]E.M. Fraedrich, S. Glasauer, V.L. Flanagin, Spatiotemporal phase-scrambling
increases visual cortex activity, Neuroreport 21 (2010) 596-600.
[72]S.S. Georgieva, J.T. Todd, R. Peeters, G.A. Orban, The extraction of 3D shape
from texture and shading in the human brain, Cereb Cortex 18 (2008) 2416-2438.
[73]R.K. Kana, Y. Liu, D.L. Williams, T.A. Keller, S.E. Schipul, N.J. Minshew, M.A.
Just, The local, global, and neural aspects of visuospatial processing in autism
spectrum disorders, Neuropsychologia 51 (2013) 2995-3003.
[74]X. Wu, W. Li, M. Zhang, J. Qiu, The neural basis of impossible figures: evidence
from an fMRI study of the two-pronged trident, Neurosci Lett 508 (2012) 17-21.
[75]P. Sterzer, A. Kleinschmidt, A neural signature of colour and luminance
correspondence in bistable apparent motion, Eur J Neurosci 21 (2005) 3097-3106.
[76]G.J. Brouwer, R. van Ee, J. Schwarzbach, Activation in visual cortex correlates
with the awareness of stereoscopic depth, J Neurosci 25 (2005) 10403-10413.
[77]J.S. Cant, M.A. Goodale, Attention to form or surface properties modulates
different regions of human occipitotemporal cortex, Cereb Cortex 17 (2007) 713-731.
[78]F. Schoth, T.D. Waberski, T. Krings, R. Gobbele, H. Buchner, Cerebral
processing of spontaneous reversals of the rotating Necker cube, Neuroreport 18
(2007) 1335-1338.
[79]S.O. Dumoulin, R.F. Hess, Cortical specialization for concentric shape
processing, Vision Res 47 (2007) 1608-1613.
[80]S. Baecke, R. Lutzkendorf, C. Tempelmann, C. Muller, D. Adolf, M. Scholz, J.
Bernarding, Event-related functional magnetic resonance imaging (efMRI) of depth-
by-disparity perception: additional evidence for right-hemispheric lateralization, Exp
Brain Res 196 (2009) 453-458.
[81]A. Ritzl, J.C. Marshall, P.H. Weiss, O. Zafiris, N.J. Shah, K. Zilles, G.R. Fink,
Functional anatomy and differential time courses of neural processing for explicit,
inferred, and illusory contours. An event-related fMRI study, Neuroimage 19 (2003)
1567-1577.
[82]T. Negawa, S. Mizuno, T. Hahashi, H. Kuwata, M. Tomida, H. Hoshi, S. Era, K.
Kuwata, M pathway and areas 44 and 45 are involved in stereoscopic recognition
based on binocular disparity, Jpn J Physiol 52 (2002) 191-198.