Skip to main content
Advertisement
  • Loading metrics

Visual perception of highly memorable images is mediated by a distributed network of ventral visual regions that enable a late memorability response

  • Benjamin Lahner ,

    Roles Conceptualization, Formal analysis, Funding acquisition, Investigation, Validation, Visualization, Writing – original draft, Writing – review & editing

    blahner@mit.edu

    ‡ These authors share first authorship on this work.

    Affiliation Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, Massachusetts, United States of America

  • Yalda Mohsenzadeh ,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Writing – original draft, Writing – review & editing

    ‡ These authors share first authorship on this work.

    Affiliations The Brain and Mind Institute, The University of Western Ontario, London, Canada, Department of Computer Science, The University of Western Ontario, London, Canada, Vector Institute for Artificial Intelligence, Toronto, Ontario, Canada

  • Caitlin Mullin ,

    Roles Conceptualization, Data curation, Investigation, Writing – original draft, Writing – review & editing

    ‡ These authors share first authorship on this work.

    Affiliation Vision: Science to Application (VISTA), York University, Toronto, Ontario, Canada

  • Aude Oliva

    Roles Conceptualization, Funding acquisition, Writing – original draft, Writing – review & editing

    Affiliation Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, Massachusetts, United States of America

Abstract

Behavioral and neuroscience studies in humans and primates have shown that memorability is an intrinsic property of an image that predicts its strength of encoding into and retrieval from memory. While previous work has independently probed when or where this memorability effect may occur in the human brain, a description of its spatiotemporal dynamics is missing. Here, we used representational similarity analysis (RSA) to combine functional magnetic resonance imaging (fMRI) with source-estimated magnetoencephalography (MEG) to simultaneously measure when and where the human cortex is sensitive to differences in image memorability. Results reveal that visual perception of High Memorable images, compared to Low Memorable images, recruits a set of regions of interest (ROIs) distributed throughout the ventral visual cortex: a late memorability response (from around 300 ms) in early visual cortex (EVC), inferior temporal cortex, lateral occipital cortex, fusiform gyrus, and banks of the superior temporal sulcus. Image memorability magnitude results are represented after high-level feature processing in visual regions and reflected in classical memory regions in the medial temporal lobe (MTL). Our results present, to our knowledge, the first unified spatiotemporal account of visual memorability effect across the human cortex, further supporting the levels-of-processing theory of perception and memory.

1. Introduction

Although visual memorability feels at first a subjective and hard to quantify property of an image, many behavioral, neuroscience, and computational works have shown that it is not an inexplicable phenomenon—we can calculate an image’s memorability score as its probability of being correctly remembered at a later viewing. Visual memorability has been shown to be intrinsic to the stimulus and reproducible across individuals. This means that despite varied experiences, individuals tend to remember and forget the same images [17]. This suggests images with a high memorability score are better encoded into and retrieved from memory not due to individual preferences, but due to perceptual properties of the stimuli [2,810].

Measuring these perceptual differences between images with high and low memorability scores can support or discourage theories relating perception and memory processes. Differences arising in the first few hundred milliseconds of perception within perceptual regions may better align with the levels-of-processing theory of memory, which posits that memory of a stimulus is a result of deeply processing the stimulus during perception [11,12]. Differences arising in non-perceptual regions temporally distinct from perception may better align with the multi-store models of memory, which states that stimulus processing and memory storage are separate [13]. A spatiotemporal account of the memorability effect would allow for controlled examination of this crucial link between perception and memory.

The current spatiotemporal account of memorability sensitivity is coarse and may be incomplete, relying on results pieced together from independent unimodal experiments with different subjects, stimulus sets, and tasks. Previous work spatially localizes memorability sensitivity to ventral and medial temporal lobe (MTL) regions [810] and temporally localizes it to a middle stage of visual perception, at around 150 ms poststimulus onset [14]. When specific regions of interest (ROIs) become sensitive to memorability relative to both stimulus onset and other perceptual processes remains unknown.

Here, we describe the dynamics of image memorability in humans in high spatiotemporal resolution by combining 2 neuroimaging techniques, magnetoencephalography (MEG) and functional magnetic resonance imaging (fMRI). This method is known as MEG-fMRI fusion [1521]. We isolate the effect of image memorability using semantically matched High Memorable and Low Memorable stimulus sets controlled for low-level features (see S1 Table). In doing so, we provide a unified account of the spatiotemporal dynamics of the memorability effect throughout the whole brain and identify which brain regions are sensitive to memorability and when.

To anticipate our results, we find that images with high memorability scores recruit a distributed network of ventral visual regions for temporally extended processing late in time, beginning at approximately 250 ms poststimulus onset. We identify 5 ROIs, inferior temporal cortex, fusiform gyrus, lateral occipital cortex, banks superior temporal sulcus (banks of the STS), and, notably, early visual cortex (EVC), as showing significant sensitivity to image memorability. We provide evidence that image memorability overlaps with both visual perceptual and memory regions by measuring how well a representation derived from human-based memorability score magnitudes matches neural representations in classical visual and memory regions. We find the memorability magnitude representation arises after high-level image feature processing but still during the perceptual process and memory regions in the MTL (hippocampus, amygdala, and parahippocampal gyrus) significantly represent memorability judgments. Together, our results provide a detailed spatiotemporal account of the memorability effect at the voxel and region-level and support the levels-of-processing theory linking perception with memory.

2. Results

2.1. MEG-fMRI fusion maps of the time course of the memorability effect in the brain

The effect of image memorability has typically been studied in either time or space, but not both. Here, we document when and where the effect of image memorability propagates throughout the whole-brain with MEG-fMRI fusion [15,16,19,22], a technique that enables comparison of MEG and fMRI data by mapping them into a common representational space through the calculation of representational dissimilarity matrices (RDMs) [23] (Fig 1B and 1E). By correlating a voxel’s fMRI RDM with each time point’s MEG RDM, we estimate that voxel’s contribution to the MEG signal over time. Repeated for every voxel in the brain, we achieve a matrix of correlations describing visual processing over the whole-brain and time (Fig 1C and 1F).

thumbnail
Fig 1. Whole-brain MEG-fMRI fusion methodology and results.

(A, D) Stimuli examples: Illustrative examples of experimental stimuli and their High and Low Memorable pairs of the same semantic concept. For example, the bottom images in panels (A) and (D) show a High Memorable image of “a bike rack outside of a building” and a Low Memorable image of “a bike rack outside of a building,” respectively. The real experimental stimuli are not shown due to copyright protection. The analysis steps (BC) and (EF) were separately performed on the 78 High Memorable conditions and the 78 Low Memorable conditions, respectively, and are otherwise identical. (B, E) Computing pairwise dissimilarities: For the MEG experiment, at each time point t, an RDM was computed where the (i,j) indices correspond to a dissimilarity measure (SVM classification accuracy) between condition i and condition j. For the fMRI experiment, at each voxel v, an RDM was computed where the (i,j) indices correspond to a dissimilarity measure (1-Pearson’s R) between the activation patterns within a spherical searchlight (radius of 4 voxels). (C, F) MEG-fMRI fusion correlation map: The partial correlation (Spearman’s R) between the MEG RDMs at each time point t and the fMRI RDMs at each voxel v while regressing out a GIST image feature RDM results in a time series of correlations across the whole-brain while controlling for possible confounding image features. (G) High Memorable and Low Memorable correlation map difference: The Low Memorable correlation maps were subtracted from the High Memorable correlation maps to create a new High > Low Difference correlation map over the whole-brain and across time. (H) High > Low Memorable fusion results: Axial slices of the brain at 56 mm and 26 mm between 250 ms and 600 ms poststimulus onset depict significant differences in correlation between the High Memorable and Low Memorable correlation maps (n = 15, cluster-definition threshold P < 0.001, significance threshold (alpha) P < 0.01, permutations = 1,000). fMRI, functional magnetic resonance imaging; MEG, magnetoencephalography; RDM, representational dissimilarity matrix; SVM, support vector machine.

https://doi.org/10.1371/journal.pbio.3002564.g001

We compute an MEG-fMRI fusion correlation series on the 78 High Memorable conditions (Fig 1A–1C) and 78 Low Memorable conditions (Fig 1D–1F) separately. Both MEG-fMRI fusion correlation series show significant clusters of correlations onset in the EVC around 80 ms and propagate throughout the visual cortex, as expected from previous works [15,21] (see S1 Movie).

We then subtract the Low Memorable fusion correlation series from the High Memorable fusion correlation series to obtain a spatiotemporal view of sensitivity to image memorability (Fig 1G). Since the High and Low Memorable stimuli sets are controlled for high- and low-level perceptual features (see S1 Table), any effect that remains after the High > Low Memorable difference contrast is attributable to the difference in memorability scores. Furthermore, this difference contrast is not biased by hand-selected memorability features in order to provide a general description of memorability sensitivity throughout the whole brain.

Fig 1H shows snapshots of the whole-brain High Memorable > Low Memorable difference fusion (cluster statistics) over the time points 250 to 600 ms in 2 select axial slices highlighting significant differences in correlations distributed throughout the visual cortex. Significant differences in correlations first arise in the posterior temporal lobe at around 275 ms followed by responses in the posterior occipital lobe and lateral occipital lobe at 300 ms. Significant differences in correlations cease by 600 ms (see S1 Movie for a link to the High Memorable, Low Memorable, and High > Low Memorable difference whole-brain MEG-fMRI fusion movies side-by-side).

Overall, results from this spatially and temporally unconstrained fusion analysis provide evidence that sensitivity to image memorability is distributed throughout visual cortex late in time. This whole-brain MEG-fMRI fusion analysis complements previous MEG-only work showing sensitivities to image memorability late in time [14] and fMRI-only work showing sensitivity to image memorability near high-level ventral visual stream and posterior temporal lobe (but not in extra-visual regions associated with memory) [8,9]. In contrast to these previous works, we identify voxels sensitive to image memorability in the posterior occipital lobe, a region typically associated with low-level image feature processing. The late and distributed responses warrant region-specific analyses to further quantify onset times and build a clearer picture of the effect of image memorability on visual perception.

2.2. ROI-based MEG-fMRI fusion shows late onset of image memorability sensitivity

We extend the MEG-fMRI fusion analysis to individual ROIs defined for both the fMRI and MEG data to identify which regions are sensitive to image memorability and when they show this sensitivity. We identify 6 non-overlapping ROIs from the Desikan–Killiany atlas [24] to spatially constrain both the MEG and fMRI data. Specifically, we use the ROIs EVC (combination of lingual gyrus, cuneus cortex, and pericalcarine cortex regions), lateral occipital cortex, inferior temporal cortex, fusiform gyrus, middle temporal gyrus, and banks of the STS for their non-overlapping but abutting coverage in or near the ventral visual stream that we beforehand might expect to be sensitive to image memorability [810,14]. These ROIs span typical visual regions (EVC, lateral occipital cortex, inferior temporal cortex, fusiform gyrus) [2531] and nearby multimodal integration regions (middle temporal gyrus and banks of the STS) [3236] for broad functional coverage of visual perception.

The ROI-based MEG-fMRI fusion method (see section 4.8.2) correlates a source-estimated MEG time series of RDMs (see section 4.6.3) with an fMRI RDM (see section 4.7.3) for a given ROI (Fig 2A, example green ROI is lateral occipital cortex) to obtain a ROI-specific correlation time series for each subject. This procedure is done separately for the High Memorable (red) and Low Memorable conditions (blue) (Fig 2A). Subtracting the Low Memorable correlation time series from the High Memorable correlation time series results in the High > Low Memorable Difference correlation time series (orange) (Fig 2A). We estimate onset latencies for the High, Low, and High > Low Difference conditions within each ROI using a jackknife-based baseline technique [3740]. In detail, the jackknife-based baseline technique computes leave-one-out subsamples and, for each subsample, considers a time point as significant if: (1) the value at that time point exceeds 2 times the baseline standard deviation; and (2) the average value over 10 consecutive 50 ms windows also exceeds 2 times the baseline standard deviation. The first time point that satisfies both criteria is marked as the onset for that subsample (see Methods section 4.8.2 for more details). Fig 2B shows the grand-average correlation time series curves, the average onset for each condition (the average of all subsample onsets), and the significant time points for the High, Low, and High > Low difference condition for each ROI (time points where all 15 jackknife subsamples met the 2 criteria).

thumbnail
Fig 2. Source-estimated MEG-fMRI ROI fusion correlation time series.

(A) ROI MEG-fMRI fusion methods: For each of the 6 ROIs (in this example, the green ROI is lateral occipital), an fMRI RDM and a source-estimated MEG RDM time series are partially correlated (regressing out the GIST image features) for the High Memorable (red) and Low Memorable (blue) images, separately. The resulting Low Memorable fusion correlation curves are subtracted from the High Memorable fusion correlation curves to compute the High > Low Memorable Difference curve (orange). (B) Source-estimated MEG-fMRI ROI fusion results: The 15-subject grand average correlation curves for the High Memorable, Low Memorable, and High > Low Memorable Difference conditions are plotted for 6 ROIs in or abutting the ventral visual stream. The corresponding color-coded horizontal bars above the curves represent the significant time points at which all 15 jackknifed subsamples passed the jackknife-based baseline criteria. The black dots depict the average onset within each High, Low, and High > Low condition, and the color-coded vertical bar depicts the onset’s 95% confidence interval (one-sample t test, right-sided, variance corrected for the subsamples). The dotted black vertical lines connect each conditions’ average onset to the x-axis for easier viewing of the onset times. fMRI, functional magnetic resonance imaging; MEG, magnetoencephalography; RDM, representational dissimilarity matrix; ROI, region of interest.

https://doi.org/10.1371/journal.pbio.3002564.g002

Fig 2B shows the presence of significant time points in the High > Low Difference correlations for EVC, lateral occipital cortex, inferior temporal cortex, fusiform gyrus, and banks of the STS with an average onset of 303, 415, 278, 368, and 269 ms poststimulus onset, respectively (see S2 Table for results for all conditions). This result shows that perception of High Memorable images differs from perception of Low Memorable images in a spatially distributed set of regions late in time.

2.3. Memorability score magnitude is represented after high-level perception

In order to pinpoint when image memorability is represented with respect to the perceptual time course, we use representational similarity analysis (RSA) [23,41] to correlate low-level, high-level, and memorability behavior model RDMs with the source-estimated MEG RDM time series of classical visual processing regions [15,4245]. To approximate the low-level and high-level image representations of the human visual system, RDMs are computed by extracting the features after the first convolutional layer (here referred to as Conv1) and penultimate fully connected layer (here referred to as FC7) from an AlexNet deep convolutional neural network (DNN) trained on image classification [46] (Fig 3A). To capture the human brain’s behavioral readout of image memorability, a model RDM is computed using the magnitude of difference in memorability scores between each pair of stimuli (Fig 3B).

thumbnail
Fig 3. Model and source-estimated MEG RSA.

(A) Creation of AlexNet layer RDMs: Features for each of the 156 stimuli were extracted after the first convolution (Conv1, pink) and after the penultimate fully connected layer (FC7, green) in an AlexNet model pretrained on ImageNet. Dissimilarity between each pair of images was measured as 1 –Pearson correlational similarity between the flattened image features and entered into the RDM. (B) Creation of memorability behavior RDM: Each value in the (i,j) index of the memorability behavior RDM (orange) is the absolute value of the image i’s memorability score minus image j’s memorability score. (C) Correlation of model RDMs with MEG RMDs: The three 156 × 156 model RDMs (AlexNet Conv1, AlexNet FC7, and memorability behavior) were partially correlated (spearman; Conv1 partialled out FC7, FC7 partialled out Conv1, and memorability behavior partialled out GIST) with the source-estimated MEG time series for each of the 4 typical visual regions. The color-coded horizontal bars above the curves represent the significant time points at which all 15 jackknifed subsamples passed the jackknife-based baseline criteria. The average onset time and 95% confidence interval for each curve are denoted by black dots with a dotted black line extending down to the x-axis (for easier visualization) and a short color-coded vertical bar, respectively. The onset times were computed using a jackknife approach. The latencies between condition onsets are denoted by dual color-coded bars below the curves. The latencies were computed using a two-sided t test and corrected for multiple comparisons using the Bonferroni–Holm procedure (q < 0.05, n = 12 comparisons). An asterisk to the right of the latency bar denotes significance, and N.S. denotes “Not Significant.” MEG, magnetoencephalography; RDM, representational dissimilarity matrix; RSA, representational similarity analysis.

https://doi.org/10.1371/journal.pbio.3002564.g003

Correlating a model RDM with a region’s source-estimated MEG RDM time series measures how strongly the ROI shares the model’s representation over time, indicative of the ROI’s underlying neural processes [23,41,42]. Crucially, the model and source-estimated MEG RDMs use all 156 image conditions to capture how images across High and Low Memorable conditions are represented. Comparing differences in onset latencies between each of these 3 model representations then allows us to establish a temporal ordering of these neural processes. We restrict this analysis to classical visual regions (EVC, lateral occipital cortex, fusiform gyrus, and inferior temporal cortex) for fair comparison with a vision-based model. We compute the conditions’ onsets (as opposed to peaks) to measure the earliest time the conditions are represented in the brain.

We find that the onset of image memorability processing (orange bar) is later than low-level (pink bar) and/or high-level (green bar) visual processing in all 4 visual regions by an average of 47 milliseconds and 42 milliseconds, respectively (Fig 3C). EVC, lateral occipital cortex, and inferior temporal cortex show the processing of memorability magnitudes occurs significantly later than both low-level and high-level processing, suggesting that the processing of image memorability is temporally distinct, and later than, both low-level and high-level perceptual processing (two-sided t test, Bonferroni–Holm correction for 12 comparisons, q < 0.05). In fusiform gyrus, the processing of memorability magnitudes occurs later than the high-level processing. Latency calculations involving the low-level features, especially latency differences between low-level and high-level features, may not have become significant due to larger variability in the low-level features (see S3 Table for all onset and latency values).

Together, this analysis shows that the image memorability representation that reflects behavioral recognition performance arises shortly after high-level visual processing (approximately 42 ms after) but still well within the time period of perceptual processing (approximately 107 ms poststimulus onset). Thus, the behavioral readout of image memorability is a late-stage representation but is still reflected during perception.

2.4. Memorability score magnitude is reflected in classical memory regions

To determine the extent that classical memory regions also represent image memorability, we create RDMs for the hippocampus, parahippocampal gyrus, amygdala, and left prefrontal cortex (LPFC) and correlate them with the memorability score magnitude RDM. The hippocampus, parahippocampal gyrus, and amygdala compose part of the MTL and are often implicated in encoding, retrieving, and/or mediating memory [4752]. LPFC has been consistently observed to positively respond in subsequent memory literature [53] and used in prior memorability work to dissociate image memorability from individual memory [8].

The memorability magnitude RDM (Fig 4A), also used in the analysis in section 2.3, summarizes the relationship between each pair of images by their difference in ability to be correctly remembered by human observers.

thumbnail
Fig 4. Correlation of classical memory fMRI RDMs with memorability magnitude RDM.

(A) Correlation analysis: The 156 × 156 memorability magnitude RDM is partially correlated (spearman, partialling out GIST features) with each of the 15 subject’s fMRI ROI RDM (in this example, the hippocampus ROI). (B) Memorability representation in classical memory regions: For each of the 4 memory regions, the partial correlation (Spearman, GIST features partialled out) between the 15 subjects’ fMRI RDM and the memorability magnitude RDM was computed (black dots). Statistical significance was computed using a right sided t test against a null correlation of 0 (q < 0.05, Bonferroni–Holm correction, n = 4 comparisons). The boxplot shows the median correlation (red line), 25th and 75th percentile (blue box), and the most extreme non-outlier points (black whiskers). An asterisk above the boxplot denotes significance, and N.S. denotes “Not Significant.” fMRI, functional magnetic resonance imaging; LPFC, left prefrontal cortex; RDM, representational dissimilarity matrix; ROI, region of interest.

https://doi.org/10.1371/journal.pbio.3002564.g004

Similar to [8], we find that hippocampus (p = 2.1e-5), parahippocampal gyrus (p = 2.4e-6), and amygdala (p = 0.0427) significantly correlate with the memorability magnitude RDM, while LPFC (p = 0.1004) does not (right-sided t test, Bonferroni–Holm correction for n = 4 comparisons, q < 0.05; see S4 Table for equivalent nonparametric tests). This result shows the behavioral representation of image memorability is represented in not only visual regions (see section 2.3) but also memory regions.

3. Discussion

3.1. Summary of results

This study combines high spatial resolution fMRI with high temporal resolution MEG to provide a unified account of the when and where of the memorability effect throughout the whole brain. Spatiotemporal analyses reveal that sensitivity to image memorability is more distributed than previously thought. We find sensitivity in not only the high-level visual regions of inferior temporal cortex, fusiform gyrus, and lateral occipital cortex, but also the banks of the STS and EVC. This sensitivity occurs late in visual processing, at approximately 303, 415, 278, 368, and 269 ms for EVC, lateral occipital cortex, inferior temporal cortex, fusiform gyrus, and banks of the STS, respectively. We then analyze the extent classic visual regions and classic memory regions represent a specific model of image memorability based on humans’ image recognition performance. The visual brain represents memorability score magnitudes significantly after high-level features with its first representations emerging approximately 107 ms poststimulus onset, suggesting that the memorability effect arises late in, but during, perception. Additionally, classical memory regions of hippocampus, parahippocampal gyrus, and amygdala reflect representations of memorability score magnitudes, further showing how visual perception and memory are intertwined.

3.2. EVC shows sensitivity to image memorability

Both our whole-brain MEG-fMRI fusion (section 2.1) and ROI-based MEG-fMRI fusion (section 2.2) analyses exhibit strong agreement that EVC is sensitive to differences in image memorability late in time (around 303 ms poststimulus onset). This result is in contrast with previous work that identifies image memorability as a high-level process different from low-level vision and did not find significant involvement of EVC [8,9,14]. Why does this set of MEG-fMRI fusion analyses find significant sensitivity in EVC?

We first note that this study’s fMRI-only univariate and multivariate analyses (see S2 Fig and S3 Fig), closely resembling the analyses of Bainbridge and colleagues (2017) and Bainbridge and Rissman (2018), also do not appear to show sensitivity in EVC. However, these multivariate analyses show a lack of statistically significant correlation between an EVC voxel’s representational geometry and a specific memorability-related representational geometry. Such an analysis is not sufficient to conclude that EVC is not sensitive to any aspects of image memorability.

The MEG-fMRI fusion analysis, on the other hand, can be viewed as testing 1,201 representational geometries (one for each MEG time point) that do not necessarily reflect high-level theories of memorability, but rather representations defined by the MEG signal at a time point. Thus, it is likely that EVC is sensitive to image memorability, but its representational geometry corresponds to a more intermediate level of processing best brought out by the MEG time points’ representational geometries, not the memorability theories previously tested.

A second (not mutually exclusive) possible explanation as to why we see significant late correlational differences in EVC may be due to the temporal limitations of fMRI. The temporally coarse and sluggish BOLD response captured at each voxel in EVC, and subsequently its RDM, is likely dominated by a strong representational geometry corresponding to its initial low-level visual processing early in time. If the signal of EVC’s sensitivity to differences in image memorability is weaker than the signal of EVC’s initial low-level visual processing, the memorability effect’s representational contribution to the overall EVC RDM would be present but relatively weak. MEG’s temporally resolved representational geometries, as utilized in MEG-fMRI fusion analyses, may be able to highlight EVC’s memorability representation.

3.3. Spatiotemporal results support the levels-of-processing theory of memory

Our results support the levels-of-processing theory of memory, which posits that a stimulus’s better encoding into and retrieval from memory is a consequence of deeper levels of perceptual processing [11,12]. Results show the memorability effect is distributed across regions around the occipital cortex often implicated in perception (sections 2.1, 2.2, and 2.4). The memorability score magnitude RDM is first reflected in visual ROIs after high-level feature processing but still during perception (approximately 107 ms after stimulus onset) (section 2.3), and significant High Memorable and Low Memorable differences are observed at even later time points (approximately 300 ms depending on the ROI) (section 2.2). Finally, we show the memorability score magnitude RDM is significantly represented in both classical visual regions (section 2.3) and classical memory regions in the MTL (but not LPFC) (section 2.4). Together, we show that the memorability effect is observed in both visual and memory regions and characterized by spatially distributed and temporally extended perceptual processing strongly suggestive of increased processing depth [11,54].

Marrying our results with previous literature, we propose that both High and Low Memorable images are similarly processed in an initial feedforward pass through the visual perceptual hierarchy [55,56] to extract their low-level (e.g., spatial frequency, color, contrast) and high-level (e.g., shapes, concepts) properties. The high-level visual regions at the end stages of the initial visual perception, specifically the inferior temporal cortex (Fig 2B, middle left panel) and banks of the STS (Fig 2B, lower right panel) seen here at around 278 ms and 269 ms, respectively, may preferentially process High Memorable images due to their relatively more conceptually [27,57] and socially [33,34] useful information content [58].

Banks of the STS, with its strong feedback connections to EVC [59,60], may then project the signal back to EVC at approximately 303 ms (Fig 2B, top left). EVC may serve to extract additional fine details of High Memorable images for extended processing and eventual encoding to or retrieval from memory, in line with EVC’s known involvement in high-order perception [30,31,59,61,62], retention of fine details in memory [6365], and our results showing temporally distinct onsets between low-level, high-level, and memorability behavior representations (see section 2.3). Thus, we hypothesize that High Memorable images are subject to deeper levels of processing in order to extract additional relevant details in line with levels-of-processing theory.

3.4. Limitations and future work

The correlation between the MEG and fMRI RDMs in the MEG-fMRI fusion method amplifies the noise present in both MEG and fMRI signals, thus requiring a higher signal-to-noise ratio for MEG-fMRI fusion to attain statistical significance. Moreover, MEG-fMRI fusion only captures neural activity present in both MEG and fMRI separately. Thus, analyses along only the temporal dimension using MEG data or only the spatial dimension using fMRI data will likely yield stronger signal. Conversely, these limitations have the benefit of making the MEG-fMRI fusion technique conservative, meaning that any statistically significant results are more likely to be true effects.

All spatial information stems from fMRI and all temporal information stems from MEG, which theoretically can result in spatially and temporally ambiguous representational correlations [66]. Incorporating temporal information in the fMRI data or spatial information in the MEG data is the best way to resolve any potential ambiguity. While our slow (2 second TR) fMRI acquisition did not allow for a useful increase in temporal resolution in the fMRI data, we did increase the spatial resolution of the MEG data using source-estimation in order to minimize any potential ambiguity. Furthermore, even without incorporating temporal resolution in fMRI and spatial resolution in MEG, MEG-fMRI fusion has been shown to produce reliable spatiotemporal signals both across studies and subjects [1522]. MEG-fMRI fusion results also demonstrate strong agreement with independently collected spatiotemporal results [15,16,6770].

The goal of our ROI analysis was to broadly test the ventral visual cortex, abutting regions, and classical memory regions that may be implicated in visual perception. A study selecting ROIs to test specific hypotheses of an ROI’s contribution to the memorability effect can yield impactful results. Achieving source-estimated temporal information in the memory regions of hippocampus, parahippocampal gyrus, and amygdala would have allowed us to further elucidate the interplay between perception and memory, but due to the medial anatomical location of these ROIs, such MEG source-estimated information was not possible to obtain. Promising future directions include defining a temporal hierarchy for the memorability effect in light of the strong evidence of feedback processes (see section 2.2 and 2.3; [22]). Additionally, there exist many categories that might ecologically benefit from a stronger encoding into memory. We wonder if regions when sensitive to image memorability show categorical preferences, as in object recognition [71], or are agnostic to categorical information. Identifying the memorability effect’s relationship to object recognition and memory encoding and retrieval would be a dramatic step forward in understanding human visual cognition.

3.5. Conclusions

In this work, we elucidate the spatiotemporal dynamics of image memorability sensitivity in the brain by combing the high temporal resolution of MEG with the high spatial resolution of fMRI. We find sensitivity to image memorability in EVC, inferior temporal cortex, fusiform gyrus, lateral occipital cortex, and the banks of the STS, and we quantify onset times of sensitivity to image memorability in relation to stimulus onset. We find that High Memorable images are preferentially processed in high-level visual regions [810,14] and EVC late in time (but not in extra-perceptual regions). The behavioral readout of image memorability is represented after high-level feature processing in visual regions and significantly represented in classical memory regions in the MTL. In support of the levels-of-processing theory of memory, we hypothesize that High Memorable images undergo additional perceptual processing to extract detailed conceptually and socially relevant information for better encoding and retrieval from memory.

4. Materials and methods

4.1. Participants

Fifteen healthy volunteers (9 female, 27.87 ± 5.17 years old) participated in this experiment. All participants were right-handed with normal or corrected-to-normal vision and provided written consent. The studies were conducted in accordance with the Declaration of Helsinki and approved by the local Institutional Review Board ethics committee (Committee on the Use of Humans as Experimental Subjects (COUHES) of the Massachusetts Institute of Technology, approval code: 1510287948).

4.2. Stimuli

We selected 78 pairs of images (for a total of 156 images) from the LaMem Memorability dataset [72] and the 10K USA Adult Faces [6] to form a categorically diverse stimulus set. Each pair contained 1 High Memorable image and 1 Low Memorable image matched for high-level semantic description.

The 78 High Memorable images and 78 Low Memorable images showed no significant differences in the low-level features of color and spatial frequency (see S1 Table) [7375]. An image’s memorability score ranges from 0 to 1 and is calculated as the decimal percentage of correct image recognition trials during a behavioral memory experiment, averaged over all participants [3]. In this study, the memorability scores are 0.854 +/− 0.089 for images in the High Memorable set and 0.519 +/− 0.105 for those in the Low Memorable set.

A memorability score magnitude RDM was computed from the stimuli’s memorability scores to reflect how each pair of images differed from one another based on human results from the memory experiment. For each pair of images i and j, the difference in magnitudes of their memorability score was entered into index (i,j) of the memorability behavior RDM. For example, 2 images i and j that were recognized equally well (i.e., the same memorability score) were represented with 0 dissimilarity at index (i,j) in the RDM. This memorability magnitude RDM was used to measure the extent that this specific behavioral readout of memorability was represented in a cortical region.

4.3. GIST and AlexNet stimuli features

We computed summary image statistics for each stimuli using the GIST descriptor [76,77]. These GIST features provided an additional control against image-feature confounds between the High Memorable and Low Memorable image sets not accounted for by our matching of semantics, color, and spatial frequency (see S1 Table). Specifically, the GIST features were computed by applying a series of Gabor filters across 8 orientations and 4 spatial frequencies to each image. Each image was then windowed using a 4 × 4 grid to result in a 512 × 1 vector.

We computed a High Memorable and Low Memorable GIST RDM, each of size 78 × 78, by calculating the pairwise distance (1 –Pearson correlation) between the GIST features of the 78 High Memorable and 78 Low Memorable images, respectively. We used the High and Low Memorable GIST RDMs for partial correlation analysis (for sections 2.1, 2.2, 2.3, and 2.4) as the potentially confounding variable to control for the possibility that GIST image features, not differences in memorability scores, drove the correlations seen in the results below. See S1 Fig for a visualization of the High and Low Memorable GIST RDMs.

We extracted stimuli features from the deep convolutional neural network AlexNet [46] provided by PyTorch pretrained to classify images from ImageNet [78]. Specifically, stimuli features were extracted after the first convolutional layer and after the penultimate fully connected layer, respectively, termed “features.0” and “classifiers.4” in the model definition. Separately for each layer, the features were flattened and the pairwise dissimilarity (1 –Pearson correlation) was computed between each stimulus’s feature vector and entered into an RDM. In this way, we created a 78 × 78 high memorable RDM, 78 × 78 low memorable RDM, and a 156 × 156 all conditions RDM for the AlexNet layers. The features from the first convolutional layer in AlexNet (referred to as Conv1 here) reflect low-level image features, and the features from the penultimate fully connected layer (referred to as FC7 here) reflect high-level image features useful to probe interpretable representations in the primate visual cortex [4244].

4.4. Region of interest selection

We defined a total of 10 non-overlapping ROIs, 9 from the Desikan–Killiany Atlas [24] and 1 spherical ROI following Bainbridge and colleagues (2017) [8]. The Desikan–Killiany Atlas used to define the fMRI ROIs was downloaded from the AtlasReader [79] GitHub page (https://github.com/miykael/atlasreader/tree/master/atlasreader/data) in the filepath “atlasreader/atlasreader/data/atlases/atlas_desikan_killiany.nii.gz.” The atlas was created by converting FreeSurfer’s (version 6.0) [80] “aparc+aseg.mgz” file output from the FreeSurfer folder “subjects/cvs_avg35_inMNI152/mri” into a standard NifTi file using FreeSurfer’s mri_convert function. The labels follow FreeSurfer’s “FreeSurferColorLUT.txt” file. Six of the 9 Desikan–Killiany Atlas [24] ROIs spanned the ventral visual stream where we a priori might have expected to see sensitivities to image memorability [8,9,14]. Namely, we defined EVC (which was made by concatenating estimated cortical signals in Lingual, Cueneus, and Pericalcarine), lateral occipital, inferior temporal, middle temporal, fusiform, and banks of the STS [24]. We used these same ROIs for MEG source-estimation and fMRI ROI analyses. The remaining 3 ROIs from the Desikan–Killiany Atlas [24] were the classical memory regions of hippocampus, parahippocampal gyrus, and amygdala. Finally, we defined the memory region of LPFC as a spherical ROI with radius 4 voxels centered around MNI coordinate (−50, 9, 31), as in Bainbridge and colleagues (2017) [8].

4.5. Experimental procedure

fMRI and MEG data were acquired in separate sessions (session orders were counterbalanced) and followed presentation parameters optimized for each imaging modality. Participants passively viewed the stimuli during 1 MEG session and 2 fMRI sessions while engaged in an orthogonal vigilance task. In all sessions, images were presented at the center of the screen at 6.0° visual angle with 500 ms duration, overlaid with a black fixation cross. Subjects were not informed that images had different memorability scores.

For fMRI, participants completed 2 sessions of 5 to 8 runs of 610 s each (totaling 11 to 15 runs over both sessions). Each image was presented once per run and image order was randomized. Null trials occurred 25% of the time, where the fixation cross against a gray background was white for 100 ms before returning to its black color for the remainder of the trial. Participants were instructed to respond to the change in fixation color with a button press. Image and null trials were both 3 s (500 ms stimulus display 2.5 s fixation only).

For MEG, participants completed 1 session of 25 runs, with trials of 1 to 1.2 s each. Each image was presented once per run in a randomized order. Null trials occurred randomly every 3 to 5 trials and consisted of an image of an eye. Subjects were instructed to only blink on these null trials and respond with a button press.

4.6. MEG acquisition and analysis

4.6.1. MEG acquisition and preprocessing.

MEG signals were acquired from 306 channels (204 planar gradiometers, 102 magnetometers, Elekta Neuromag TRIUX, Elekta, Stockholm, Sweden) using a sampling rate of 1 kHz, filtered between 0.03 and 330 Hz. We applied temporal source space separation (maxfilter software, Elekta, Stockholm) [81,82] before analyzing data with Brainstorm [83]. For each trial, we extracted peri-stimulus data from −200 ms to +1,000 ms, removed baseline mean, and smoothed data with a 30 Hz low pass filter. We obtained 25 trials for each condition, session, and participant.

4.6.2. MEG multivariate pattern analysis.

We conducted MEG multivariate pattern analysis on each subject data independently (Fig 1C and 1G). First, we extracted preprocessed MEG signals from 200 ms before to 1,000 ms after stimulus onset and arranged the 306 sensor measurements in 306-dimensional vector patterns at each time point. This yields N pattern vectors per time point and stimulus condition. We then trained linear support vector machine (SVM) classifiers (LibSVM implementation, www.csie.ntu.edu.tw/~cjlin/libsvm) to discriminate patterns between any 2 conditions. Following previous works [14,22,84,85], we binned MEG trials randomly into groups of 3 for each condition. To reduce noise and computational complexity, we averaged the trials assigned to each bin, resulting in M pseudo trials [66,8587]. Then, at each time point and for each pair of conditions, we assigned M-1 pseudo trials for training the SVM classifier. The trained SVM was then tested on the Mth left out vector from each of the 2 conditions. The random assignment to bins, sub-averaging, training and testing procedure were repeated for 100 times. The classifier performance (decoding accuracy), shown to be a reliable and high-SNR dissimilarity index [87], over the 100 iterations was averaged and used to populate a 156 × 156 matrix, each row and conditions indexed by the stimulus conditions. This procedure yielded one 156 × 156 symmetric matrix of decoding accuracies for every time point, referred to as MEG representational dissimilarity matrix (MEG RDM). Extracting the pairwise decoding accuracies of just the High Memorable conditions and just the Low Memorable conditions yields two 78 × 78 MEG RDMs per time point.

4.6.3. MEG source-estimation.

We mapped MEG signals to cortical sources using default anatomy [88] and based on Freesurfer automatic segmentation [89]. Performing multivariate pattern analysis on MEG signals in the 6 cortical regions of interest, we localized MEG representational information on that region. For source localization, using an overlapping spheres model [90], we first computed the forward model and then mapped MEG signals on the cortex using a dynamic statistical parametric mapping (dSPM) approach [91]. Pattern vectors in each of the 6 cortical regions were then extracted and employed for a multivariate pattern analysis in that region. Source-estimation was computed separately for the left and right hemispheres of each region. Once the RDM time series was computed for each hemisphere and region, the left and right hemispheres were combined.

4.7. fMRI acquisition and analysis

4.7.1. fMRI acquisition and preprocessing.

Both sessions were conducted at the Athinoula A. Martinos Imaging Center at the MIT McGovern Institute, using a 3T Siemens Trio scanner (Siemens, Erlangen, Germany) with a 32-channel phased-array head coil. In each session, we acquired structural images using a standard T1-weighted sequence (176 sagittal slices, FOV = 256 mm2, TR = 2530 ms, TE = 2.34 ms, flip angle = 9°). Across the 2 sessions, we acquired 11 to 15 runs of 305 volumes for each participant (gradient-echo EPI sequence: TR = 2,000 ms, TE = 29 ms, flip angle = 90°, FOV read = 200 mm, FOV phase = 100%, bandwidth 2,368 Hz/Px, gap = 20%, resolution = 3.1 mm isotropic, slices = 33, ascending acquisition).

The fMRI data were preprocessed using SPM software (Wellcome Trust Center for Neuroimaging, University College London). For each participant, fMRI data were slice-time corrected, realigned and co-registered to the T1 structural scan of the first session, and finally normalized to the standard MNI space. We fit a general linear model (GLM) to estimate the fMRI responses to the 156 images. Stimulus onsets and durations, as well as motion and run regressors were included in the GLM. All the regressors were convolved with a hemodynamic response function (canonical HRF) with time resolution of TR/N and onset of the first slice (TR = 2,000 msec, N = 33 slices).

4.7.2. fMRI voxelwise searchlight analysis.

We obtained 156 condition-specific t-maps by computing the contrast of each condition’s GLM estimate against an implicit baseline. We then computed a searchlight analysis [92,93] to calculate an RDM at every voxel for each subject (Fig 1B and 1E). Specifically, we defined a spherical searchlight of radius 4 voxels and centered it on voxel v. Following previous work [16,94,95], the condition-specific t-values were extracted from all voxels within the searchlight, and the dissimilarities between each condition were calculated (1-Pearson’s R) to obtain an RDM of shape (156 × 156) for voxel v. Each row i and column j of the RDM corresponded to a specific condition, and the (i,j) index into the RDM was the dissimilarity measure between condition i and condition j. The searchlight sphere was then moved to the next voxel, and the procedure was repeated until an RDM was calculated at each voxel. This searchlight analysis was performed on each subject separately.

4.7.3. fMRI ROI searchlight analysis.

To obtain 1 RDM for each of the 6 ROIs, we averaged the individual voxel RDMs (from the whole-brain searchlight procedure detailed in Methods section 4.7.2) contained within the corresponding ROI mask. We repeated this procedure for each subject separately.

4.8. MEG-fMRI fusion.

M/EEG-fMRI fusion uses RSA [23,41] to relate temporally resolved MEG or EEG signal with spatially resolved fMRI signal [15,17,18,96,97]. The method assumes that if 2 images are similarly represented in MEG patterns, they should also be similarly represented in fMRI patterns. Translating this neural data into representational space allows us to connect a location in the brain to a particular time point. This creates a spatiotemporally resolved view of the emergence of a specific visual representation in the brain.

4.8.1. Whole-brain MEG-fMRI fusion.

We used a partial correlation analysis to correlate the fMRI RDMs at each voxel (as described in section 4.7.2) with the whole-brain MEG RDMs (as described in section 4.6.2) while regressing out the GIST image feature RDMs (as described in section 4.3) (Fig 1C and 1F) in order to control for low and mid-level image features not explicitly accounted for in the stimulus set. In line with previous work [15,16], analysis was conducted on the subject-averaged MEG RDMs with the subject-specific fMRI RDMs. For each time point, we computed the similarity (Spearman’s R, thus avoiding scaling issues and parametric assumptions) between the MEG RDM and the fMRI RDM of each voxel while controlling for the potentially confounding GIST image features. This yielded a 3D correlation map of representational similarities, indicating locations in the brain at which neuronal processing emerged at a particular time point. Repeating this analysis for each time point creates a spatiotemporally resolved view of neural activity in the brain during perception free of any spatial priors.

We computed MEG-fMRI fusion separately for High Memorable and Low Memorable conditions by using their respective MEG, fMRI, and GIST High and Low Memorable RDMs. The spatiotemporal dynamics of the effect of image memorability was the difference between the High and Low Memorable MEG-fMRI fusion time series (Fig 1G). Nonparametric permutation-based cluster correction statistics (n = 15, cluster-definition threshold P < 0.001, significance threshold (alpha) P < 0.01, permutations = 1,000, right-tailed) were used to identify significant spatial and temporal clusters for the 4D (voxel X, voxel Y, voxel Z, time point t) correlational fusion series [98,99]. Specifically, we first computed a mean 4D correlation map by averaging each subject’s 4D correlation map. Next, we used each voxel’s values in the baseline period (−200 to 0 ms) in the mean correlation map to empirically construct a distribution of the null hypothesis. We estimated a maximal cluster size threshold by randomly shuffling the signs of each subject’s data 1,000 times, averaging over the subjects, and defining clusters by the values’ spatial and temporal proximity at the cluster-definition threshold (defined to be right-sided, P = 0.001). An empirical distribution of the maximal cluster size was computed via permutation sampling and a cluster size threshold of P = 0.01. A cluster is significant if it exceeds the cluster size threshold.

4.8.2. Source-estimated MEG-fMRI ROI fusion.

ROI-based MEG-fMRI fusion, as used in this study, first required the calculation of source-estimated MEG-fMRI fusion. We first computed a given ROI’s source-estimated MEG RDM time series (as described in section 4.6.3). We then computed a given ROI’s fMRI RDM by averaging each voxel’s RDM contained within the ROI mask (as described in section 4.7.3). Finally, we performed a partial correlation between the subject-averaged MEG ROI RDMs at each time point and the subject-specific fMRI ROI RDM with the GIST RDM serving as the potentially confounding variable. This procedure resulted in a correlation time series for each ROI of size 15 × 1,201 (n_subjects × n_timepoints).

When determining the onset latencies of the ROI fusion correlation time series, we faced 2 main challenges: low signal-to-noise ratio (as common with most neuroimaging data) and no prior knowledge of the signal’s canonical shape. We addressed these 2 challenges using a jackknife-based baseline criteria technique. Starting with an ROI correlation time series matrix of size 15 × 1,201 (n_subjects × n_timepoints), we followed previous work to identify significant latencies [3740].

First, we computed the standard deviation of the baseline values across time (−200 to 0 ms) from the 15-subject grand-averaged correlation time series to establish the baseline criteria. Next, we iterated over each subject to compute 15 subsamples of leave-one-out mini grand averages, where each of the remaining 14 subjects are averaged at each iteration (jackknife) [38]. For each of the jackknifed subsamples, a time point t was considered significant if it satisfied 2 criteria [37,39]: (1) the value at time point t was greater than or equal to 2 times the baseline standard deviation; and (2) the average value of the next 10 windows of t + 50 ms was also greater than or equal to 2 times the baseline standard deviation. The first significant time point for each of the 15 jackknifed subsamples was the onset for that subsample. This procedure resulted in 15 samples of onset latencies per condition.

The strict baseline criteria ensured that significant time points satisfy an amplitude criterion (2 times the baseline standard deviation) and a temporal grouping criterion (10 moving windows of 50 ms) that noise was unlikely to satisfy, and the jackknifing procedure increased the signal-to-noise ratio to better estimate signals with non-canonical shapes. For an ROI’s High Memorable, Low Memorable, or High > Low Memorable curves to be considered significant, at least 1 time point must exist where all 15 subsamples passed the 2 criteria. Otherwise, the correlation curve was considered not significant.

Since the subsamples were no longer independent, we adjusted the variance by multiplying by a factor of (n—1) [2] (where n = 15) when computing subsequent statistics (for detailed proof and discussion of this adjustment, see [37,38,100]). The mean onset latency of each condition and ROI was computed by averaging the onset latencies of the 15 subsamples. The 95% confidence interval around the onset latency for each condition and ROI was computed by multiplying the standard error of the onset latencies (with the adjusted variance) by a critical t-statistic of 2.145 to represent a 95% confidence interval with 14 degrees of freedom. Onset latencies were computed with a one-sided 1 sample t test (against an onset of 0 ms, stimulus onset), and onset latency differences between conditions (as in section 2.3) were computed with a two-sided paired t test.

Statistical tests for latency differences between conditions (section 2.3) and memorability magnitude to memory RDM correlations (section 2.4) were corrected for multiple comparisons using Bonferroni–Holm correction (n = 12 tests and n = 4 tests, respectively) at q = 0.05 [101].

Supporting information

S1 Fig. Visualization of RDMs.

We show the RDMs for the 9 fMRI ROIs, memorability score magnitudes, and GIST features for the 78 High Memorable image conditions, 78 Low Memorable image conditions, all 156 image conditions. The fMRI and GIST RDMs are computed by calculating the pairwise dissimilarity (1—Pearson correlation) between each stimulus’s feature vector. The memorability magnitude RDM is computed by calculating the magnitude of difference in memorability scores between each pair of images. For the GIST RDMs, each stimulus feature is a 512 × 1 vector computed from a GIST model. For the fMRI RDMs, each stimulus feature is a n_voxel × 1 vector of t-values, where n_voxels is the number of voxels contained within a spherical searchlight (radius of 4 voxels). All RDMs contained within the ROI’s mask are averaged together, and the average RDM over 15 subjects is shown here. The dissimilarity is ranked and divided by the maximum dissimilarity, resulting in the displayed RDMs with percent values.

https://doi.org/10.1371/journal.pbio.3002564.s001

(TIF)

S2 Fig. fMRI-only univariate memorability contrast.

For each subject, a GLM was used to fit the observed fMRI data to the experimental conditions. The beta weight map from all 78 high memorable conditions was contrasted against the beta weight map from all 78 low memorable conditions to create a high memorable > low memorable contrast image (one-sample T test, p < 0.001, voxel extent threshold (k) = 10). Each high memorable > low memorable contrast image was input into a group level analysis to compute the group effect (one-sample T test, p < 0.001, voxel extent threshold (k) = 10), shown here. The colorbar scale has units of t-statistic.

https://doi.org/10.1371/journal.pbio.3002564.s002

(TIF)

S3 Fig. fMRI-only multivariate memorability contrast.

For each subject, a GLM was used to fit the observed fMRI data to the experimental conditions. The beta weight map from each of the 156 conditions was contrasted against the beta weight map from response (null) trials to create a t-statistic significance map (referred to as t-map) for each condition (one-sample T test, p < 0.001, voxel extent threshold (k) = 10). A searchlight analysis was then performed on these t-maps to create a 156 × 156 RDM at each voxel in the brain, where each (i, j) index into the RDM is a measure of the dissimilarity between condition i and condition j. We then created a single hypothesized 156 × 156 RDM where each (i, j) index was the Euclidean distance (absolute value of the difference) between the memorability score of condition i and the memorability score of condition j. The memorability score associated with each condition is a high-level behavioral measure of image memorability, and therefore, the hypothesized RDM is a high-level behavioral measure of image memorability in representation space. We then correlated (Spearman’s R) the hypothesized RDM with each RDM at each voxel in the brain and perform statistical analysis on the correlation map shown here (cluster statistics, n = 15, cluster-definition threshold P < 0.01, cluster threshold P < 0.01, permutations = 1,000).

https://doi.org/10.1371/journal.pbio.3002564.s003

(TIF)

S1 Table. High and Low Memorable stimuli controlled for low-level image features.

We use the Natural Image Statistical Toolbox [73,74,101] to compare low-level statistics (color, brightness, contrast, and spatial frequency at various energy levels) across the set of 78 High Memorable images and the set of 78 Low Memorable images. The spectral energy levels of 10%, 30%, 50%, 70%, and 90% refer to the percentage of total energy of the spectrum in the image sets (larger percentages encompass the energy levels of lesser percentages). In RGB color space, the “R,” “G,” and “B” denote significance testing within the “Red,” “Green,” and “Blue” color channels, respectively. In Lab color space, “L,” “a,” and “b” denote significance testing within “Luminance,” “a color channel,” and “b color channel,” respectively. At a p-value of p < 0.05, no statistically significant differences were found.

https://doi.org/10.1371/journal.pbio.3002564.s004

(XLSX)

S2 Table. Source-estimated ROI-based MEG-fMRI fusion onset latency.

The mean latency and 95% one-sided confidence intervals (milliseconds) for the source-estimated MEG-fMRI ROI fusion correlation time series in the High Memorable, Low Memorable, High > Low Difference, and High > Low Difference minus High Memorable conditions were computed using a jackknife-based relative baseline criteria on the searchlight ROI-based MEG-fMRI fusion correlation time series. For each condition, the standard deviation of the 15 subject grand-averaged baseline values was first computed (from −200–0 ms before stimulus onset). Then, for each of the jackknife subsamples, any time point that met the (1) amplitude criteria of a value greater than or equal to the baseline value and (2) the temporal criteria of an averaged value in 10 consecutive 50 ms windows greater than or equal to the baseline value was deemed significant. The first significant time point in each of the jackknife subsamples (n = 15) was used to compute the average latency and 95% confidence interval. The variance in the standard error calculation was corrected by a factor of (n– 1) [2] to account for the dependence between the jackknifed subsamples. A time series that contained at least 1 subsample with zero significant time points (that did not meet the 2 criteria stated above) was deemed “Not Significant,” denoted in the table by “N.S.”

https://doi.org/10.1371/journal.pbio.3002564.s005

(XLSX)

S3 Table. Model MEG RSA latency.

The mean latency and 95% one-sided confidence intervals (milleseconds) for the source-estimated MEG-model RDM correlation time series for AlexNet Conv1, AlexNet FC7, and memorability magnitude model RDMs were computed using a jackknife-based relative baseline criteria. For each correlation time series, the standard deviation of the 15 subject grand-averaged baseline values was first computed (from −200–0 ms before stimulus onset). Then, for each of the jackknife subsamples, any time point that met the (1) amplitude criteria of a value greater than or equal to the baseline value and (2) the temporal criteria of an averaged value in 10 consecutive 50 ms windows greater than or equal to the baseline value was deemed significant. The first significant time point in each of the jackknife subsamples (n = 15) was used to compute the average latency and 95% confidence interval (one-sided t test). For the latency difference calculations between MEG-model correlation time series, a two-sided paired t test was used. The variance in the standard error calculation was corrected by a factor of (n– 1) [2] to account for the dependence between the jackknifed subsamples. In the latency difference calculations, multiple comparisons were corrected (n = 12 comparisons) using Bonferroni–Holm correction [100] at q = 0.05. The p-values shown here are adjusted for Bonferroni–Holm multiple comparison correction, but the 95% confidence intervals are not adjusted to retain transparency of the original computation and avoid misinterpretations from the step-wise corrections employed by the Bonferroni–Holm method.

https://doi.org/10.1371/journal.pbio.3002564.s006

(XLSX)

S4 Table. Parametric and nonparametric statistics for memory ROI correlations with memorability magnitude RDM.

For each of the 4 memory regions studied here, the correlation between the fMRI ROI 156 × 156 RDM and the 156 × 156 memorability magnitude RDM was computed (Spearman, partialled out GIST features). Significance was computed using a one-sided t test (parametric) and one-sided Wilcoxon sign rank test (nonparametric). Both tests were corrected for multiple comparisons (n = 4 comparisons, Bonferroni–Holm correction) at q = 0.05 [100]. The adjusted p-value is shown.

https://doi.org/10.1371/journal.pbio.3002564.s007

(XLSX)

S1 Movie. Whole-brain MEG-fMRI fusion movies.

Youtube link: https://youtu.be/skYb5l5xWBQ. The video depicts axial slices of the whole-brain MEG-fMRI fusion movies for the High Memorable images, Low Memorable images, and the difference between the High and Low Memorable fusion correlation time series (High > Low Memorable difference). The 3 movies were computed independently with identical statistics (permutation-based cluster statistics; cluster-definition threshold = 0.001, significance threshold (alpha) = 0.01, n_permutations = 1,000) and stitched together in the one movie linked here for easy side-by-side comparison.

https://doi.org/10.1371/journal.pbio.3002564.s008

(MP4)

Acknowledgments

We thank Dimitrios Pantazis for guidance for the MEG experiments and data analyses. We thank the National Institute of Health for their shared instrument grant 1S10OD021569 and Siemens PrismaFit 3T scanner (Erlangen, Germany) (https://www.nih.gov/). The study was conducted at the Athinoula A. Martinos Imaging Center at the McGovern Institute for Brain Research, Massachusetts Institute of Technology.

References

  1. 1. Bainbridge WA. Chapter One—Memorability: How what we see influences what we remember. In: Federmeier KD, Beck DM, editors. Psychology of Learning and Motivation. Academic Press; 2019. p 1–27.
  2. 2. Bainbridge WA. The resiliency of image memorability: A predictor of memory separate from attention and priming. Neuropsychologia. 2020;141:107408.
  3. 3. Isola P, Xiao J, Parikh D, Torralba A, Oliva A. What Makes a Photograph Memorable? IEEE Trans Pattern Anal Mach Intell. 2014;36:1469–1482. pmid:26353315
  4. 4. Goetschalckx L, Moors P, Wagemans J. Image memorability across longer time intervals. Memory. 2018;26:581–588. pmid:28974153
  5. 5. Isola P, Parikh D, Torralba A, Oliva A. Understanding the Intrinsic Memorability of Images. In: Advances in Neural Information Processing Systems. Curran Associates, Inc.; 2011. Available from: https://proceedings.neurips.cc/paper/2011/hash/286674e3082feb7e5afb92777e48821f-Abstract.html (accessed 2022 May 24).
  6. 6. Bainbridge WA, Isola P, Oliva A. The intrinsic memorability of face photographs. J Exp Psychol Gen. 2013;142:1323–1334. pmid:24246059
  7. 7. Bylinskii Z, Isola P, Bainbridge C, Torralba A, Oliva A. Intrinsic and extrinsic effects on image memorability. Vision Res. 2015;116:165–178. pmid:25796976
  8. 8. Bainbridge WA, Dilks DD, Oliva A. Memorability: A stimulus-driven perceptual neural signature distinctive from memory. Neuroimage. 2017;149:141–152. pmid:28132932
  9. 9. Bainbridge WA, Rissman J. Dissociating neural markers of stimulus memorability and subjective recognition during episodic retrieval. Sci Rep. 2018;8:1–11.
  10. 10. Jaegle A, Mehrpour V, Mohsenzadeh Y, Meyer T, Oliva A, Rust N. Population response magnitude variation in inferotemporal cortex predicts image memorability. Elife. 2019;8:e47596. pmid:31464687
  11. 11. Craik FIM. Levels of processing: Past, present… and future? Memory. 2002;10:305–318.
  12. 12. Craik FIM, Lockhart RS. Levels of processing: A framework for memory research. J Verbal Learn Verbal Behav. 1972;11:671–684.
  13. 13. Atkinson RC, Shiffrin RM. Human Memory: A Proposed System and its Control Processes. In: Spence KW, Spence JT, editors. Psychology of Learning and Motivation. Academic Press; 1968. p 89–195.
  14. 14. Mohsenzadeh Y, Mullin C, Oliva A, Pantazis D. The perceptual neural trace of memorable unseen scenes. Sci Rep. 2019;9:6033. pmid:30988333
  15. 15. Cichy RM, Pantazis D, Oliva A. Resolving human object recognition in space and time. Nat Neurosci. 2014;17:455–462. pmid:24464044
  16. 16. Cichy RM, Pantazis D, Oliva A. Similarity-Based Fusion of MEG and fMRI Reveals Spatio-Temporal Dynamics in Human Cortex During Visual Object Recognition. Cereb Cortex. 2016;26:3563–3579. pmid:27235099
  17. 17. Hebart MN, Bankson BB, Harel A, Baker CI, Cichy RM. The representational dynamics of task and object processing in humans. Elife. 2018;7:e32816. pmid:29384473
  18. 18. Henriksson L, Mur M, Kriegeskorte N. Rapid Invariant Encoding of Scene Layout in Human OPA. Neuron. 2019;103:161–171.e3. pmid:31097360
  19. 19. Khaligh-Razavi S-M, Cichy RM, Pantazis D, Oliva A. Tracking the Spatiotemporal Neural Dynamics of Real-world Object Size and Animacy in the Human Brain. J Cogn Neurosci. 2018;30:1559–1576. pmid:29877767
  20. 20. Lowe MX, Mohsenzadeh Y, Lahner B, Charest I, Oliva A, Teng S. Cochlea to categories: The spatiotemporal dynamics of semantic auditory representations. Cogn Neuropsychol. 2021;38:468–489. pmid:35729704
  21. 21. Mohsenzadeh Y, Mullin C, Lahner B, Cichy RM, Oliva A. Reliability and Generalizability of Similarity-Based Fusion of MEG and fMRI Data in Human Ventral and Dorsal Visual Streams. Vision. 2019;3:8. pmid:31735809
  22. 22. Mohsenzadeh Y, Qin S, Cichy RM, Pantazis D. Ultra-Rapid serial visual presentation reveals dynamics of feedforward and feedback processes in the ventral visual pathway. Elife. 2018;7:e36329. pmid:29927384
  23. 23. Kriegeskorte N, Mur M, Bandettini P. Representational similarity analysis—connecting the branches of systems neuroscience. Front Syst Neurosci. 2008;2. Available from: https://www.frontiersin.org/article/10.3389/neuro.06.004.2008 (accessed 2022 May 24). pmid:19104670
  24. 24. Desikan RS, Ségonne F, Fischl B, Quinn BT, Dickerson BC, Blacker D, et al. An automated labeling system for subdividing the human cerebral cortex on MRI scans into gyral based regions of interest. Neuroimage. 2006;31:968–980. pmid:16530430
  25. 25. Weiner KS, Zilles K. The anatomical and functional specialization of the fusiform gyrus. Neuropsychologia. 2016;83:48–62. pmid:26119921
  26. 26. Miyashita Y. Inferior temporal cortex: where visual perception meets memory. Annu Rev Neurosci. 1993;16:245–263. pmid:8460893
  27. 27. Conway BR. The Organization and Operation of Inferior Temporal Cortex. Annu Rev Vis Sci. 2018;4:381–402. pmid:30059648
  28. 28. Grill-Spector K, Kourtzi Z, Kanwisher N. The lateral occipital complex and its role in object recognition. Vision Res. 2001;41:1409–1422. pmid:11322983
  29. 29. Palejwala AH, Dadario NB, Young IM, O’Connor K, Briggs RG, Conner AK, et al. Anatomy and White Matter Connections of the Lingual Gyrus and Cuneus. World Neurosurg. 2021;151:e426–e437. pmid:33894399
  30. 30. Gilbert C, Ito M, Kapadia M, Westheimer G. Interactions between attention, context and learning in primary visual cortex. Vision Res. 2000;40:1217–1226. pmid:10788637
  31. 31. Lee TS, Mumford D, Romero R, Lamme VAF. The role of the primary visual cortex in higher level vision. Vision Res. 1998;38:2429–2454. pmid:9798008
  32. 32. Beauchamp MS, Lee KE, Argall BD, Martin A. Integration of Auditory and Visual Information about Objects in Superior Temporal Sulcus. Neuron. 2004;41:809–823. pmid:15003179
  33. 33. Deen B, Koldewyn K, Kanwisher N, Saxe R. Functional Organization of Social Perception and Cognition in the Superior Temporal Sulcus. Cereb Cortex. 2015;25:4596–4609. pmid:26048954
  34. 34. Isik L, Koldewyn K, Beeler D, Kanwisher N. Perceiving social interactions in the posterior superior temporal sulcus. Proc Natl Acad Sci U S A. 2017;114:E9145–E9152. pmid:29073111
  35. 35. Cabeza R, Nyberg L. Imaging Cognition II: An Empirical Review of 275 PET and fMRI Studies. J Cogn Neurosci. 2000;12:1–47.
  36. 36. Chao LL, Haxby JV, Martin A. Attribute-based neural substrates in temporal cortex for perceiving and knowing about objects. Nat Neurosci. 1999;2:913–919. pmid:10491613
  37. 37. Kiesel A, Miller J, Jolicœur P, Brisson B. Measurement of ERP latency differences: A comparison of single-participant and jackknife-based scoring methods. Psychophysiology. 2008;45:250–274. pmid:17995913
  38. 38. Miller J, Patterson T, Ulrich R. Jackknife-based method for measuring LRP onset latency differences. Psychophysiology. 1998;35:99–115. pmid:9499711
  39. 39. Osman A, Bashore TR, Coles MGH, Donchin E, Meyer DE. On the transmission of partial information: Inferences from movement-related brain potentials. J Exp Psychol Hum Percept Perform. 1992;18:217–232. pmid:1532189
  40. 40. Sassenhagen J, Draschkow D. Cluster-based permutation tests of MEG/EEG data do not establish significance of effect latency or location. Psychophysiology. 2019;56:e13335. pmid:30657176
  41. 41. Kriegeskorte N, Kievit RA. Representational geometry: integrating cognition, computation, and the brain. Trends Cogn Sci. 2013;17:401–412. pmid:23876494
  42. 42. Cichy RM, Khosla A, Pantazis D, Torralba A, Oliva A. Comparison of deep neural networks to spatio-temporal cortical dynamics of human visual object recognition reveals hierarchical correspondence. Sci Rep. 2016;6:27755. pmid:27282108
  43. 43. Yamins DL, Hong H, Cadieu C, DiCarlo JJ. Hierarchical Modular Optimization of Convolutional Networks Achieves Representations Similar to Macaque IT and Human Ventral Stream. In: Advances in Neural Information Processing Systems. Curran Associates, Inc.; 2013. Available from: https://proceedings.neurips.cc/paper/2013/hash/9a1756fd0c741126d7bbd4b692ccbd91-Abstract.html (accessed 2023 Feb 13).
  44. 44. Yamins DLK, Hong H, Cadieu CF, Solomon EA, Seibert D, DiCarlo JJ. Performance-optimized hierarchical models predict neural responses in higher visual cortex. Proc Natl Acad Sci U S A. 2014;111:8619–8624. pmid:24812127
  45. 45. Rajalingham R, Issa EB, Bashivan P, Kar K, Schmidt K, DiCarlo JJ. Large-Scale, High-Resolution Comparison of the Core Visual Object Recognition Behavior of Humans, Monkeys, and State-of-the-Art Deep Artificial Neural Networks. J Neurosci. 2018;38:7255–7269. pmid:30006365
  46. 46. Krizhevsky A. One weird trick for parallelizing convolutional neural networks. 2014.
  47. 47. Dickerson BC, Eichenbaum H. The Episodic Memory System: Neurocircuitry and Disorders. Neuropsychopharmacology. 2010;35:86–104. pmid:19776728
  48. 48. Ward AM, Schultz AP, Huijbers W, Van Dijk KRA, Hedden T, Sperling RA. The parahippocampal gyrus links the default-mode cortical network with the medial temporal lobe memory system. Hum Brain Mapp. 2014;35:1061–1073. pmid:23404748
  49. 49. McGaugh JL, Cahill L, Roozendaal B. Involvement of the amygdala in memory storage: Interaction with other brain systems. Proc Natl Acad Sci U S A. 1996;93:13508–13514. pmid:8942964
  50. 50. Phelps EA. Human emotion and memory: interactions of the amygdala and hippocampal complex. Curr Opin Neurobiol. 2004;14:198–202. pmid:15082325
  51. 51. Bird CM, Burgess N. The hippocampus and memory: insights from spatial processing. Nat Rev Neurosci. 2008;9:182–194. pmid:18270514
  52. 52. Squire LR, Stark CEL, Clark RE. The Medial Temporal Lobe. Annu Rev Neurosci. 2004;27:279–306. pmid:15217334
  53. 53. Kim H. Neural activity that predicts subsequent memory and forgetting: A meta-analysis of 74 fMRI studies. Neuroimage. 2011;54:2446–2461. pmid:20869446
  54. 54. Treisman A. Levels of processing in human memory. 2014. Available from: https://books.google.com/books/about/Levels_of_Processing_in_Human_Memory_PLE.html?id=w2GLAwAAQBAJ (accessed 2023 Apr 26).
  55. 55. Felleman DJ, Van Essen DC. Distributed hierarchical processing in the primate cerebral cortex. Cereb Cortex N Y N. 1991;1:1–47. pmid:1822724
  56. 56. Logothetis NK, Sheinberg DL. Visual object recognition. Annu Rev Neurosci. 1996;19:577–621. pmid:8833455
  57. 57. Kanwisher N. Functional specificity in the human brain: A window into the functional architecture of the mind. Proc Natl Acad Sci U S A. 2010;107:11163–11170. pmid:20484679
  58. 58. Bylinskii Z, Goetschalckx L, Newman A, Oliva A. Memorability: An Image-Computable Measure of Information Utility. In: Ionescu B, Bainbridge WA, Murray N, editors. Human Perception of Visual Information. Cham: Springer International Publishing; 2022. p 207–239.
  59. 59. Noesselt T, Rieger JW, Schoenfeld MA, Kanowski M, Hinrichs H, Heinze H-J, et al. Audiovisual Temporal Correspondence Modulates Human Multisensory Superior Temporal Sulcus Plus Primary Sensory Cortices. J Neurosci. 2007;27:11431–11441. pmid:17942738
  60. 60. Falchier A, Clavagnier S, Barone P, Kennedy H. Anatomical Evidence of Multimodal Integration in Primate Striate Cortex. J Neurosci. 2002;22:5749–5759. pmid:12097528
  61. 61. Supèr H, Spekreijse H, Lamme VAF. Two distinct modes of sensory processing observed in monkey primary visual cortex (V1). Nat Neurosci. 2001;4:304–310. pmid:11224548
  62. 62. Wokke ME, Vandenbroucke ARE, Scholte HS, Lamme VAF. Confuse Your Illusion: Feedback to Early Visual Cortex Contributes to Perceptual Completion. Psychol Sci. 2013;24:63–71. pmid:23228938
  63. 63. Bainbridge WA, Hall EH, Baker CI. Drawings of real-world scenes during free recall reveal detailed object and spatial information in memory. Nat Commun. 2019;10:5. pmid:30602785
  64. 64. Brady TF, Konkle T, Alvarez GA, Oliva A. Visual long-term memory has a massive storage capacity for object details. Proc Natl Acad Sci U S A. 2008;105:14325–14329. pmid:18787113
  65. 65. Konkle T, Brady TF, Alvarez GA, Oliva A. Conceptual distinctiveness supports detailed visual long-term memory for real-world objects. J Exp Psychol Gen. 2010;139:558–578. pmid:20677899
  66. 66. Cichy RM, Oliva A. A M/EEG-fMRI Fusion Primer: Resolving Human Brain Responses in Space and Time. Neuron. 2020;107:772–781. pmid:32721379
  67. 67. Bullier J. Feedback connections and conscious vision. Trends Cogn Sci. 2001;5:369–370. pmid:11520692
  68. 68. Robinson DL, Rugg MD. Latencies of visually responsive neurons in various regions of the rhesus monkey brain and their relation to human visual responses. Biol Psychol. 1988;26:111–116. pmid:3061476
  69. 69. Schmolesky MT, Wang Y, Hanes DP, Thompson KG, Leutgeb S, Schall JD, et al. Signal Timing Across the Macaque Visual System. J Neurophysiol. 1998;79:3272–3278. pmid:9636126
  70. 70. VanRullen R, Thorpe SJ. The Time Course of Visual Processing: From Early Perception to Decision-Making. J Cogn Neurosci. 2001;13:454–461.
  71. 71. Ratan Murty NA, Bashivan P, Abate A, DiCarlo JJ, Kanwisher N. Computational models of category-selective brain regions enable high-throughput tests of selectivity. Nat Commun. 2021;12:5540. pmid:34545079
  72. 72. Khosla A, Raju AS, Torralba A, Oliva A. Understanding and Predicting Image Memorability at a Large Scale. 2015. p 2390–2398.
  73. 73. Bainbridge WA, Oliva A. A toolbox and sample object perception data for equalization of natural images. Data Brief. 2015;5:846–851. pmid:26693521
  74. 74. Torralba A, Oliva A. Statistics of natural image categories. Netw Comput Neural Syst. 2003;14:391–412. pmid:12938764
  75. 75. Park S, Konkle T, Oliva A. Parametric Coding of the Size and Clutter of Natural Scenes in the Human Brain. Cereb Cortex. 2015;25:1792–1805. pmid:24436318
  76. 76. Oliva A, Torralba A. Modeling the Shape of the Scene: A Holistic Representation of the Spatial Envelope. 2001. p. 31.
  77. 77. Oliva A, Torralba A. Chapter 2 Building the gist of a scene: the role of global image features in recognition. In: Progress in Brain Research. Elsevier; 2006. p. 23–36.
  78. 78. Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L. ImageNet: A large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition. 2009. p. 248–255.
  79. 79. Notter M, Gale D, Herholz P, Markello R, Notter-Bielser M-L, Whitaker K. AtlasReader: A Python package to generate coordinate tables, region labels, and informative figures from statistical MRI images. J Open Source Softw. 2019;4:1257.
  80. 80. Fischl B. FreeSurfer. Neuroimage. 2012;62:774–781. pmid:22248573
  81. 81. Taulu S, Kajola M, Simola J. Suppression of Interference and Artifacts by the Signal Space Separation Method. Brain Topogr. 2004;16:269–275. pmid:15379226
  82. 82. Taulu S, Simola J. Spatiotemporal signal space separation method for rejecting nearby interference in MEG measurements. Phys Med Biol. 2006;51:1759–1768. pmid:16552102
  83. 83. Tadel F, Baillet S, Mosher JC, Pantazis D, Leahy RM. Brainstorm: A User-Friendly Application for MEG/EEG Analysis. Comput Intell Neurosci. 2011;2011:e879716. pmid:21584256
  84. 84. Cichy RM, Khosla A, Pantazis D, Oliva A. Dynamics of scene representations in the human brain revealed by magnetoencephalography and deep neural networks. Neuroimage. 2017;153:346–358. pmid:27039703
  85. 85. Pantazis D, Fang M, Qin S, Mohsenzadeh Y, Li Q, Cichy RM. Decoding the orientation of contrast edges from MEG evoked and induced responses. Neuroimage. 2018;180:267–279. pmid:28712993
  86. 86. Cichy RM, Pantazis D. Multivariate pattern analysis of MEG and EEG: A comparison of representational structure in time and space. Neuroimage. 2017;158:441–454. pmid:28716718
  87. 87. Guggenmos M, Sterzer P, Cichy RM. Multivariate pattern analysis for MEG: A comparison of dissimilarity measures. Neuroimage. 2018;173:434–447. pmid:29499313
  88. 88. Holmes CJ, Hoge R, Collins L, Woods R, Toga AW, Evans AC. Enhancement of MR Images Using Registration for Signal Averaging. J Comput Assist Tomogr. 1998;22:324–333. pmid:9530404
  89. 89. Fischl B, Salat DH, van der Kouwe AJW, Makris N, Ségonne F, Quinn BT, et al. Sequence-independent segmentation of magnetic resonance images. Neuroimage. 2004;23:S69–S84. pmid:15501102
  90. 90. Huang MX, Mosher JC, Leahy RM. A sensor-weighted overlapping-sphere head model and exhaustive head model comparison for MEG. Phys Med Biol. 1999;44:423–440. pmid:10070792
  91. 91. Dale AM, Liu AK, Fischl BR, Buckner RL, Belliveau JW, Lewine JD, et al. Dynamic Statistical Parametric Mapping: Combining fMRI and MEG for High-Resolution Imaging of Cortical Activity. Neuron. 2000;26:55–67. pmid:10798392
  92. 92. Haynes J-D, Sakai K, Rees G, Gilbert S, Frith C, Passingham RE. Reading Hidden Intentions in the Human Brain. Curr Biol. 2007;17:323–328. pmid:17291759
  93. 93. Kriegeskorte N, Goebel R, Bandettini P. Information-based functional brain mapping. Proc Natl Acad Sci U S A. 2006;103:3863–3868. pmid:16537458
  94. 94. Carlin JD, Calder AJ, Kriegeskorte N, Nili H, Rowe JB. A Head View-Invariant Representation of Gaze Direction in Anterior Superior Temporal Sulcus. Curr Biol. 2011;21:1817–1821. pmid:22036180
  95. 95. Haynes J-D. A Primer on Pattern-Based Approaches to fMRI: Principles, Pitfalls, and Perspectives. Neuron. 2015;87:257–270. pmid:26182413
  96. 96. Muukkonen I, Ölander K, Numminen J, Salmela VR. Spatio-temporal dynamics of face perception. Neuroimage. 2020;209:116531. pmid:31931156
  97. 97. Salmela V, Salo E, Salmi J, Alho K. Spatiotemporal Dynamics of Attention Networks Revealed by Representational Similarity Analysis of EEG and fMRI. Cereb Cortex. 2018;28:549–560. pmid:27999122
  98. 98. Maris E, Oostenveld R. Nonparametric statistical testing of EEG- and MEG-data. J Neurosci Methods. 2007;164:177–190. pmid:17517438
  99. 99. Pantazis D, Nichols TE, Baillet S, Leahy RM. A comparison of random field theory and permutation methods for the statistical analysis of MEG data. Neuroimage. 2005;25:383–394. pmid:15784416
  100. 100. Ulrich R, Miller J. Using the jackknife-based scoring method for measuring LRP onset effects in factorial designs. Psychophysiology. 2001;38:816–827. pmid:11577905
  101. 101. Holm S. A Simple Sequentially Rejective Multiple Test Procedure. Scand J Stat. 1979;6:65–70.