Abstract
The Individual Brain Charting (IBC) is a multi-task functional Magnetic Resonance Imaging dataset acquired at high spatial-resolution and dedicated to the cognitive mapping of the human brain. It consists in the deep phenotyping of twelve individuals, covering a broad range of psychological domains suitable for functional-atlasing applications. Here, we present the inclusion of task data from both naturalistic stimuli and trial-based designs, to uncover structures of brain activation. We rely on the Fast Shared Response Model (FastSRM) to provide a data-driven solution for modelling naturalistic stimuli, typically containing many features. We show that data from left-out runs can be reconstructed using FastSRM, enabling the extraction of networks from the visual, auditory and language systems. We also present the topographic organization of the visual system through retinotopy. In total, six new tasks were added to IBC, wherein four trial-based retinotopic tasks contributed with a mapping of the visual field to the cortex. IBC is open access: source plus derivatives imaging data and meta-data are available in public repositories.
Similar content being viewed by others
Background & Summary
Mapping cognition across the whole human brain requires the multi-dimensional analysis of the correlates of behavior corresponding to a wide range of psychological domains. This phenotyping of behavioral responses relies on brain activation maps obtained from functional Magnetic Resonance Imaging (fMRI), that quantify the underlying neural correlates modulated by mental functions across tasks1,2,3,4,5.
In order to achieve generalisation of cognitive processes, behavioural phenotyping with fMRI can be carried out through data pooling analyses, in which data from different sources, and therefore different tasks, from different participants in different studies, are aggregated. In this context, it becomes difficult to dissociate effects elicited by: (1) common functional signatures of cognitive processes (even if different tasks performed by different participants share some psychological domains); (2) within-subject variability in task performance; (3) individual functional differences in both between- and within-task performance; and (4) different feature distributions across multiple data-acquisition sites.
Inter-subject variability arising from individual functional differences has been widely recognized in neuroimaging and affects all kinds of group-level analyses. It has been shown to undermine not only the estimation of statistical significance6, but also the exact demarcation of functional regions according to their contribution in elementary cognitive processes4. In the past decade, many studies have thus started to adopt individual analysis, in order to examine both functional and anatomical individual differences7,8,9,10,11,12,13.
Within-subject variability has been recently addressed in a few studies. They reveal the importance of collecting comprehensive individual data to minimize this type of variability, as it improves—across individuals—task-fMRI replicability measures14 and prediction of cognitive phenotypes4,15. Additionally, the benefits of increasing the amount of individual data to consequently decrease within-subject variability becomes particularly important in small sample-size regimes14,15; this is typically the case of fMRI studies.
Inter-site variability can also affect precision in functional imaging. Multi-site data analyses have become common practice, especially in consortium projects, and technical factors pertaining to scanner platform, imaging sequences and acquisition procedures–including setup of task protocols–have thus been identified as sources of site effects16.
On the other hand, cognitive neuroscience has traditionally relied upon event- or block-organized controlled designs using unnatural stimuli17. Single-task fMRI experiments are usually conceived in this way, wherein experimental designs tightly control the variables and isolate targeted cognitive constructs as means to link them to the function of a subset of brain regions. Because this approach has been the mainstay in cognitive neuroscience, many of the publicly available task datasets, as well as the ensuing data-pooling studies are mostly based on this type of stimuli3,17,18,19,20,21. In contrast, the study of real-life, dynamic and multimodal sensory stimuli–a.k.a. naturalistic stimuli–was introduced in parallel in the field of cognitive neuroimaging by two seminal papers22,23, and led to initiatives such as the studyforrest dataset24,25,26,27. They are thought to reduce biases inherent to the choice of artificial behavioral settings, and importantly, a priori designed contrasts17. The main challenges in using naturalistic stimuli include: (1) experimental implementation, as naturalistic paradigms are lengthy in duration because their high-dimensional feature space requires the collection of a larger amount of data; (2) statistical modelling, as the standard General Linear Model (GLM) applied to naturalistic stimuli leads to high-dimensional controlled-design models due to the larger amount of features extracted from the paradigms; and (3) unsupervised data-driven approaches are preferred because of (2), but high-dimensional imaging data (many voxels) require decomposition methods that scale.
To obtain a large sample of behavioral features and simultaneously achieve a whole-brain coverage free from inter-site and between-task inter-subject variability, extensive functional mapping at high-spatial resolution in the same set of individual brains exposed to a comprehensive collection of task paradigms is necessary. We thus present herein the third release of the Individual Brain Charting (IBC) dataset: a multi-task fMRI-data collection obtained from a permanent cohort of twelve participants acquired with a spatial resolution of 1.5 mm. Its task-wise organization–pertaining to the deep phenotyping of the cohort’s behavioral responses–combined with a higher spatial resolution allows for better estimates of a broad variety of mental functions both at the individual4,28 and group level4,5.
The first and second releases of the IBC dataset encompass a broad range of cognitive domains, with an emphasis of the second release on higher-order cognition29,30. In contrast, the third release is now focused in sensory modalities. It includes two naturalistic tasks on visualization of naturalistic scenes (the Clips31 task; see Section Clips task for more details) and movie watching (the Raiders8 task; see Section Raiders task for more details), probing mainly the visual system (i.e. areas comprising the visual cortex and visual association cortex). Movie watching also covers the auditory system as well as the language system, through inclusion of higher-order components pertaining to speech and narrative comprehension. To improve the coverage of the visual system, we have also included the classic retinotopic tasks32 dedicated to map the polar angle and eccentricity in the visual cortex. Table 1 summarizes all tasks from these three releases. Considering our individual functional-atlasing approach based on deep-behavioral phenotyping for cognitive mapping of the human brain, as described in Pinho et al.4, Thirion et al.5, and Thirion et al.28, these tasks are intended to complement those from the first and second releases, through the inclusion of not only existing components but also new ones. Indeed, components concerning visual attention as well as visual object, face and body recognition, already found in ARCHI Spatial, ARCHI Emotional, HCP Emotion, HCP Relational, HCP Working Memory, Preference battery, Theory-of-Mind (TOM) battery and Visual short-term memory (VSTM) task, are now present in both naturalistic tasks. On the other hand, retinotopic tasks bring new modules on visual representation (i.e. internal representation of visual information; see the Cognitive Atlas33), namely functional localization of the meridians and quadrants of the visual field, as well as its foveal, mid-peripheral and far-peripheral areas. Data from the three releases are organized in 31 tasks (see Table 1), of which 29 (the trial-based tasks) amount for 216 contrasts described on the basis of 127 cognitive atoms from the Cognitive Atlas.
In this paper, we provide a thorough description of the tasks taking part for this extension. We also present their technical validation, using Quality-Assurance (QA) metrics that show functional activation to be a direct response to behavior. Additionally and given the aforementioned challenges posed by the analysis of fMRI data relative to naturalistic paradigms, we complement the technical validation of the Clips and Raiders tasks with an application of the Fast Shared Response Model (FastSRM), as described in Richard&Thirion34. To this end, we showcase how to extract networks linked to mental functions elicited by these tasks.
IBC is an open-access dataset consisting of high-resolution, functional maps from a dozen of subjects. It aims at providing quantitative insights about individual differences of elementary processes in cognition, by leveraging a deep behavioral-phenotyping approach. Many sessions pertaining to different tasks are thus undertaken per participant. Unlike longitudinal studies, data collection of each task only takes place once during the lifetime of the project. IBC is thus intended to serve as a source of functional correlates of various cognitive conditions, in order to support research in human neuroscience.
Methods
To avoid ambiguity with MRI-related terms, definitions follow the Brain-Imaging-Data-Structure (BIDS) Specification version 1.8.035.
Participants
The present release of the IBC dataset consists of brain fMRI data from twelve individuals (two female) acquired between April 2016 and February 2019. The experiments were carried out with the understanding and formal consent of the participants, in accordance with the ethical principles of the Helsinki declaration, as well as the French public health regulation (https://ansm.sante.fr/), which approved the study protocol.
Detailed description of the age, sex and handedness of the group is provided in Table 2. Age varied between 26 and 40 years old (median = 30 years) upon recruitment and handedness was determined with the Edinburgh Handedness Inventory36. For more information on the cohort’s recruitment, consult Pinho et al.29 and Pinho et al.30.
Materials
Stimuli
This release covers data from three main behavioral protocols: Clips task, Retinotopy battery of tasks (i.e. Wedge and Ring tasks), and Raiders task (see Experimental Paradigms Section for details about tasks' paradigms). The stimuli of the tasks were delivered through custom-made scripts that ensured a fully automated environment and computer-controlled collection of the behavioral data. All protocols were set under Python 2.7. The protocol of the Clips task was adapted from the original study31 using standard Python libraries; the ones pertaining to the Retinotopy and Raiders tasks were respectively designed with PsychoPy v1.90.337 and Expyriment v0.9.038. Visual stimuli of Clips consisted of the same color natural movies as described in Nishimoto et al.31, whereas the audiovisual stimuli presented in the Raiders task corresponded to the 2009 DVD edition of Raiders of the Lost Ark dubbed to French.
These materials are available in a public GitHub repository dedicated to the behavioral protocols of the tasks featuring the IBC dataset: https://github.com/individual-brain-charting/public_protocols (consult Section Code Availability for further details about the repository).
Eye-tracker
The video-based, eye-tracker system EyeLink 1000 Plus was used for the behavioral, training sessions of the Clips plus Retinotopy tasks (for more information about the training sessions, consult Section Experimental Procedure).
MRI equipment
The fMRI data were acquired using an MRI scanner Siemens 3 T Magnetom Prismafit along with a Siemens Head/Neck 64-channel coil. Behavioral responses for the Retinotopy tasks were obtained with a MR-compatible, five-button ergonomic pad (Current Designs, Package 932 with Pyka HHSC-1 × 5-N4) and the MRI-environment audio system for the Raiders task was set with the MR-Confon package.
All sessions were conducted at the NeuroSpin platform of the CEA Research Institute, Saclay, France.
Experimental procedure
Upon arrival to the research institute, participants were instructed about the execution and timing of the tasks referring to the upcoming session.
Particularly, behavioral training sessions prior to the MRI sessions were conducted for the Clips and Retinotopy tasks. We stress that the center of the visual field must be approximately constant during data acquisition of these tasks, as means to obtain a consistent map of the visual system within and between individuals (consult Experimental Paradigms Section for more details). To this end, participants were prepared during the training sessions to gain perception of their eyes' movements. They were instructed to move them as little as possible while fixating toward a flickering point placed on the center of a screen, which was also displaying video scenes at the same time. These video scenes were excerpts of those presented for the Clips task. An eye-tracker was also coupled to the experimental setup of the training session. Participants were thus provided with a real-time feedback of their eyes' movements in the form of a green moving point displayed on the screen, too. The main goal of this training exercise was to keep the green moving point as close as possible to the flickering one, which was fixed. By this way, participants could then practice how to keep looking continuously toward the center of the screen—i.e. the perimetric origin—for as long as possible.
All MRI sessions were composed of several runs dedicated to one or two tasks as described in Section Experimental Paradigms. The structure of the sessions according to the MRI modality employed at every run is detailed in Table 3. Information on the imaging parameters of the referred modalities can be found in Imaging Data Section.
Experimental paradigms
Tasks were aggregated in different sessions according to Table 3. The following sections intend to provide a description of the paradigms employed for each task. Materials used for stimulus presentation (see Section Stimuli) have been made publicly available, together with video-demo presentations of the corresponding protocols, on https://github.com/individual-brain-charting/public_protocols (see Section Code Availability for a detailed description of the repository).
A comprehensive explanation of the experimental paradigms as well as conditions and contrasts of the Retinotopy tasks can be also found on the online documentation of the IBC dataset: https://individual-brain-charting.github.io/docs.
Clips task
The Clips task stands for a reproduction of the study reported in Nishimoto et al.31, in which participants were to visualize naturalistic scenes edited as video clips of ten and a half minutes each.
Data collection was carried out in uninterrupted runs, each of them dedicated to the presentation of a specific video clip. As in the original study, runs were grouped in two categories: training and test. Scenes from the training category were shown only once in the corresponding video clip. Contrariwise, scenes from the test category were composed of approximately one-minute-long excerpts selected from the clips presented during training. Excerpts were concatenated to construct the sequence of every test run; each sequence was predetermined by randomly permuting many excerpts that were repeated ten times each across all runs. The same randomized sequences, employed across test runs, were used to collect data from all participants.
There were twelve and nine runs dedicated to the collection of training data and test data, respectively. Data from nine runs of each category were interspersedly acquired over three full sessions. The three remaining runs for training-data collection were acquired in half of one last session, prior to the Retinotopy tasks (see Section Retinotopy tasks for complete description of these tasks).
To assure the same topographic reference of the visual field for all participants, a colored fixation point was always presented at the center of the images. Such point was changing three times per second to ensure that it was visible regardless the color of the movie. To account for stabilization of the BOLD signal, ten extra seconds of acquisition were added at the beginning and end of every run. The total duration of each run was thus ten minutes and fifty seconds.
Retinotopy tasks
The Retinotopy tasks refer to the classic retinotopic paradigms—i.e. the Wedge and the Ring tasks—consisting of four kinds of visual stimuli: (1-2) a slowly rotating clockwise or counterclockwise, semicircular checkerboard stimulus, as part of the Wedge task; and (3-4) a thick, dilating or contracting ring, as part of the Ring task. The phase of the periodic response at the fundamental frequency of rotation or dilation/contraction measured at each voxel is related to the measurement of the perimetric parameters of polar angle and eccentricity, respectively32.
In the present study, six runs were devoted to this task. Each of them were five-and-a-half minutes long. They were programmed for the same session following the last three “training-data” runs of the Clips task (see Section Clips task for complete description of this task). Four runs were dedicated to the presentation of the rotating checkerboard stimulus (two runs for each direction) and the remaining two were dedicated to the dilating or contracting ring, one at a time.
Similarly to the Clips task, a point was displayed at the center of the visual stimulus in order to keep constant the perimetric origin in all participants. Participants were thus to fixate continuously this point whose color flickered throughout the entire run. To keep the participants engaged in the task, they were instructed that, at the end of every run (i.e. after MRI acquisition was finished), they would be asked which color had most often been presented. They had to select one of the four possible options–i.e. red, green, blue or yellow–by pressing on the corresponding button in the response box.
Raiders task
The Raiders task was adapted from Haxby et al.8, in which the full-length action movie Raiders of the Lost Ark was presented to the participants. The main goal of the original study was the estimation of the hyperalignment parameters that transform voxel space of functional data into feature space of brain responses, linked to the visual characteristics of the movie displayed.
Similarly to the Clips task, the movie was shown herein to the IBC participants in contiguous runs determined according to the chapters of the movie defined in the DVD (see Section Stimuli for details about the DVD edition). Yet, one shall note that while the raiders task involved free-viewing, the Clips task required fixation.
This task was completed in two sessions. In order to use the acquired fMRI data in train-test split and cross-validation experiments, we performed three extra-runs at the end of the second session in which the three first chapters of the movie were repeated.
To account for stabilization of the BOLD signal, ten seconds of acquisition were added at the end of the run. The structure of both sessions and duration of their runs are detailed in Table 3.
Data acquisition
Data across participants were acquired throughout four MRI sessions, whose structure is described in Table 3. Deviations from this structure was only registered for participant 8 (sub-08), whose data pertaining to Run #13 of Raiders task was acquired in the session dedicated to the Theory-of-Mind and Pain Matrices task battery from the second release of the IBC dataset (for more details about this release, consult Pinho et al.30).
Behavioral data
Behavioral data from the training sessions for the Clips and Retinotopy tasks were recorded throughout four sessions.
Scores were obtained according to the response accuracy (in percentage) of the participant for a given trial. They indicate how close participant’s sight was from the center of the screen during the trial. Each session was composed of four different trials. Therefore, we collected a maximum of sixteen scores per participant across all sessions. They are presented in the Supplementary Material. Each participant completed at least one training session. However, the completion of the four sessions was not mandatory, as there was a trade-off between extensive training and quality of performance. Since participants could experience some fatigue while fixating their eyes for a continuous period of time, we thus recommended carrying out at least one training session (for more information about how training sessions were conducted, consult Section Experimental Procedure). We thus denote that the recorded scores were merely intended to facilitate participants' training and no further analyses were conducted with these data.
Additionally, participants were also asked to provide a button-press response at the end of every run concerned with the Retinotopy tasks (for more details about these tasks, consult Section Retinotopy tasks). The registry of these behavioral data was held in log files generated by the stimulus-delivery software and the response accuracy across runs for each participant is provided in Section Behavioral Results.
Imaging data
FMRI data were collected using a Gradient-Echo (GE) pulse, whole-brain Multi-Band (MB) accelerated39,40 Echo-Planar Imaging (EPI) T2*-weighted sequence with Blood-Oxygenation-Level-Dependent (BOLD) contrasts, using the following parameters: the repetition time (TR) is 2000 ms; the echo time (TE) is 27 ms; the flip angle is 74°; the field-of-view (FOV) is 192 × 192 × 140 mm3; voxel size is 1.5 × 1.5 × 1.5 mm3; the slice orientation is axial; slices are acquired in interleaved fashion; in-plane acquisitions were accelerated by a factor (GRAPPA) of 2; and across slices, a multi-band factor of 3 was used. Two different acquisitions for the same task were always performed using two opposite phase-encoding directions: one from Posterior to Anterior (PA) and the other from Anterior to Posterior (AP). The main purpose was to mitigate geometrical distortions while assuring built-in, within-subject replication of the same tasks.
Spin-Echo (SE) EPI-2D image volumes were acquired in order to correct for spatial distortions, using the following parameters: a TR of 7680 ms; a TE of 46 ms; a FOV of 192 × 192 × 140 mm3; a voxel size of 1.5 × 1.5 × 1.5 mm3; axial slice orientation; and acceleration factor (GRAPPA) = 2. Similarly to the GE-EPI sequences, two different acquisitions were also performed using PA and AP phase-encoding directions.
In addition, a 3D magnetization-prepared rapid gradient-echo (MP-RAGE) T1-weighted anatomical-image volume, covering the whole brain, was acquired with the following parameters: voxel size of 1 × 1 × 1 mm3; sagittal slice orientation; flip angle of 9°; and FOV of 256 × 256 × 160 mm.
A detailed description of the imaging parameters set for each MRI modality is available in Pinho et al.29 and in the IBC-dataset documentation: https://individual-brain-charting.github.io/docs.
Imaging-Data analysis
Prior to any neuroimaging-data analysis, the MRI DICOM images were converted to NIfTI format using the DCM2NII tool, which is available on https://www.nitrc.org/projects/dcm2nii. Conversion to NIfTI format also included a full anonymization of the data, i.e. pseudonyms were removed and images were defaced using the Freesurfer-6.0.0 library41.
Preprocessing
All GE-EPI volumes (aka the functional volumes) were collected twice with reversed phase-encoding directions, resulting in pairs of images with distortions going in opposite directions. Susceptibility-induced off-resonance field was estimated from the two SE-EPI volumes, which were collected twice and also using reversed phase-encoding directions. The GE-EPI volumes were then corrected based on the corresponding deformation model, which was computed using the FSL implementation as described in Andersson et al.42 and Smith et al.43.
Source data were then preprocessed using the same pipeline as described in Pinho et al.29 and Pinho et al.30, which relies on the PyPreprocess library: https://github.com/neurospin/pypreprocess. Pypreprocess stands for a collection of Python scripts–built upon the Nipype interface44–which are oriented toward a common workflow of fMRI-data preprocessing analysis; it uses precompiled modules of both SPM12 software package (Wellcome Department of Imaging Neuroscience, London, UK) and FSL library (Analysis Group, FMRIB, Oxford, UK) v6.0.0.
GE-EPI volumes of each participant were aligned among them45; the mean EPI volume was co-registered onto the corresponding T1-weighted MPRAGE volume (aka the anatomical volume); the individual anatomical volumes were segmented for the estimation of the deformation field to be applied on the normalization of both anatomical and functional volumes46.
The T1-weighted MPRAGE volume and the aligned GE-EPI volumes were then given as input to FreeSurfer v6.0.0, in order to extract the surface meshes of the tissue interfaces and a sampling of the functional activation on these meshes, as described in vanEssen et al.47. These functional activations were then resampled onto the fsaverage7 template of FreeSurfer48.
For more information about the IBC preprocessing pipeline, consult ‘Preprocessing' section of Pinho et al.29 and/or the online documentation of the IBC dataset: https://individual-brain-charting.github.io/docs.
Naturalistic-Data analysis
Naturalistic stimuli typically imply a profuse amount of data descriptors, which leads to high-dimensional design matrices49,50. Therefore, a data-driven approach was instead employed herein as means to capture the underlying structure of the fMRI data and the ensuing effects-of-interest elicited by this type of behavioral paradigms.
Because brains exposed to the same stimuli exhibit synchronous activity23, a shared response can be obtained across different individuals. To this end, we used both a model-free and a model-based approaches.
Inter-Subject Correlations (ISC)23 of individual preprocessed data (see Section Preprocessing) can naturally capture this common response without explicitly model the data. For each run, we thus calculated the Pearson correlation across all voxels for every pair of subjects. We then calculated the mean of the pairwise correlation of each run and subsequently across runs (for results regarding this analysis, refer to Fig. 3).
Nevertheless, one disadvantage that arises from ISC is the fact it misses correspondences that are not accurate at the voxel level. We used the FastSRM implementation for fMRI data–as described in Richard&Thirion34–to analyze the preprocessed imaging data of Clips and Raiders and extract a common subjects' response to these tasks together with their individual spatial components. These individual spatial components thus refer to the subject-specific effects-of-interest elicited by the tasks. To be able to statistically validate these findings, we set an analysis based on Cross Validation (CV), in which we assessed whether reconstructed individual data of one run predicted the original individual data for that run. As referred in Table 3, data per participant and for one task were collected in different runs during one or more sessions and, consequently, data from one run pertain only to one task. Moreover, Raiders and Clips data were analyzed separately. By this way, this prediction framework is task specific.
Concretely, it consists in a double K-fold cross validation based on co-smoothing51, wherein K = 3 for N = 12 subjects and K = 2 for R runs. The total number of runs per task is different. For Raiders, R = 13 and, therefore, the size of the two folds are six and seven. For Clips, data were split in terms of number of runs for training and test, which were predefined according to the acquisition design of the task (see Clips task for more details); the size of the training and test sets are thus twelve and nine, respectively. For the two tasks, each fold comprised runs with both phase-encoding directions. This CV procedure served therefore as a scoring algorithm. It computes the correlation coefficient between predicted brain-activity maps of test runs and subjects–which were obtained from the shared response and k = 20 spatial components learnt on the training runs–and the corresponding original data. The resulting vertex-wise correlation coefficients represent a similarity measure of the shared content between subjects. This procedure is depicted in further detail on Fig. 4. To compute the group-level significance of these individual estimates, we obtained p-values through a mass-univariate group analysis and corrected for multiple comparisons, using the False Discovery Rate (FDR); subsequent vertex-wise q–values ≤ 0.05 served as a mask to display the group-level correlation coefficients (for results regarding this analysis, refer to Fig. 5b). To evaluate the upper-bound model performance of FastSRM, we determined the noise ceiling using data from the three first runs of session 1 of the Raiders task. Because these runs were repeated at the end of session 2, the noise ceiling was estimated, for every subject, as the vertex-wise correlation between the same run and its repetition. The correlation between original data and reconstructed data of the same run was thus expressed as a fraction of this noise ceiling. We then took the vertex-wise median of these ratios across subjects, normalized them to ensure they lie between 0 and 1 and, finally, took the median across runs (for results regarding this analysis, refer to Fig. 5a).
We also assessed what regions are significantly different in the performance of Raiders versus Clips by computing a two-sided paired t-test in every vertex, between the individual correlation coefficients of the two tasks (for results regarding this analysis, refer to Fig. 5c).
In order to obtain a precise labeling of the functional regions covered by the previous results, we computed the proportion of significant vertices present in the areas and regions derived from the cortical parcellation of Glasser et al.52. This estimation was performed using the projection of the HCP-MMP1.0 parcellation onto the fsaverage7 template, which is available on Mills (2016)53 (for results regarding this analysis, refer to Supplementary Table 1).
The FastSRM encoding analysis was implemented using the IdentifiableFastSRM module of the FastSRM package that can be found on: https://github.com/hugorichard/FastSRM. Statistical analyses and plotting were performed using SciPy v1.6.1 (https://www.scipy.org) and Nilearn v0.9.254 (https://nilearn.github.io).
Retinotopic-data analysis
The retinotopic mapping data (rotating wedge and expanding/shrinking rings) were analyzed using the standard frequency-domain analysis described in Sereno et al.32 and Warnking et al.55: sine/cosine regressors were specified at the frequency of the periodic stimulus position change (\(\frac{1}{32}Hz\)). The magnitude of the reponse of the voxels was estimated across six sessions, and tested for significance using an F– test, thresholded at p < 0.001 and uncorrected for multiple comparisons. The phase of the reponse in each voxel with a significant effect was then determined for each session, by comparing the relative magnitude of the sine and cosine regressors. The phase information was combined for wedge- and ring-stimuli separately, in order to cancel the hemodynamic-induced phase delay. This is possible because the stimuli were presented in opposite motions (clockwise versus anti-clockwise for the wedge and expanding versus contracting for the ring). In addition, results were averaged across replications (wedge experiments). The phase estimate therefore defines in polar coordinates (eccentricity and polar angle) the visual field position that elicits a maximal amount of activation in each voxel.
All these steps were computed on the data sampled on the fsaverage7 template using the Freesurfer software48. For visualization, the surface-based polar angle and eccentricity maps were displayed using the flat representation available through the Pycortex tool56.
Data Records
Source and derivatives data of the IBC dataset are accessible online to the public. Derivatives account for both preprocessed imaging data (i.e., normalized functional and anatomical data; for detailed information about preprocessing, consult Section Preprocessing) and the subsequent postprocessed imaging data. Postprocessed derivatives refer specifically to the individual and unthresholded z-maps, that are obtained from the contrast maps of the experimental conditions pertaining the trial-based tasks. Therefore, for the present release, these derivatives are only available for the Retinotopy tasks (see Section Retinotopy Study).
The online access of the data is assured by the Human Brain Project (HBP) EBRAINS platform57 under the project Individual Brain Charting (IBC) v3.058. It is also possible to use an integrated data-fetcher API tool to download wholly or partially the IBC dataset, which includes not only the third release but also other releases not herein reported. For more information on how to install and use the data fetcher, consult the IBC documentation available online: https://individual-brain-charting.github.io/docs. For more information about the API tool, consult Section Code Availability.
Source data is also available in the OpenNeuro public repository59, under the accession number ds00268560, and postprocessed derivatives can be found in the NeuroVault repository61, under the collection with the id = 661862.
The NIfTI files as well as paradigm descriptors and imaging parameters are organized per run for each subject and session, according to BIDS Specification, in both EBRAINS and OpenNeuro. For more details, consult ‘Data Records' sections of Pinho et al.29 and Pinho et al.30: the data descriptors of the IBC first and second releases, respectively.
Technical Validation
Behavioral results were obtained through an elementary assessment of curated behavioral data. They are reported in Section Behavioral Results.
All imaging results were obtained following the methodological procedures, as presented in Sections Naturalistic-Data Analysis and Retinotopic-Data Analysis, applied to task-fMRI data previously preprocessed on the surface, as described in Section Preprocessing. They are respectively reported in Sections FastSRM-Encoding Study and Retinotopy Study.
Behavioral results
In the Retinotopy tasks, participants were to give a button-press response at the end of every run corresponding to the color of the flickering fixation point they saw most times. There was only one correct answer out of the four possible options provided for each task run, which was fixed across participants. Table 4 displays the individual response accuracy achieved for all task runs. These scores are presented as percentages of the individual correct responses with respect to the total number of (correct) responses; because there were six runs dedicated to the Retinotopy tasks (see Table 3), six responses were obtained. The average ± standard deviation of the response accuracy across participants (excluding participant 5) are 56 ± 20%, i.e. higher than chance level (25%). These results show that overall participants' fixation was good enough to perform a discrimination task during the course of the run, thus suggesting that fixation was held properly.
Imaging results
Data quality
In order to provide an approximate estimate of data quality, standard QA measures are presented in Figs. 1 and 2 according to:
-
The temporal Signal-to-Noise ratio (tSNR), defined as the mean of each voxels' time course divided by their standard deviation, on normalized and unsmoothed data averaged across all acquisitions (from all subjects). Its values are ≥ 50 in most cortical regions. Given the high resolution of the data (1.5 mm isotropic), such values are indicative of a good image quality63;
-
The histogram of the six rigid body-motion estimates of the brain per scan, in mm/degree, together with their 99% coverage interval. One can notice that this interval ranges approximately within [−1,1] mm/degree, showing that excursions beyond 1 mm/degree motion are rare. No acquisition was discarded due to excessive motion (>2 mm/degree);
-
The framewise displacement, computed as described in Power et al.64. This scalar quantity, estimated from the six rigid body-motion estimates across runs, displays similar distributions for every subject and task. Median of individual displacements is usually lower than 0.1 mm and their upper quartile never goes beyond 0.25 mm.
Since brain activation measured with fMRI during naturalistic stimuli can be expected to consist largely of stimulus-driven, between-subject effects shared by the majority of participants, ISC of individual fMRI time series thus provide a summary statistic of subject synchrony over the entire imaging session. Figure 3 shows where positive inter-subject correlations of BOLD signal are mapped for each task across the brain and highlights the striking similarity of these representations with those obtained from the individual, model-based approach (see Fig. 5b). One can thus observe that spatial locations wherein brain activation is more similar across subjects are also those that likely reflect a significant response to behavior.
Figure 4 describes the procedure implemented to validate naturalistic-stimuli data using FastSRM. It estimates the jointly activated brain areas, in which inference is performed in a two-nested CV loop. Two core assumptions underlie this approach: (1) “Step a” of Fig. 4 assumes the same shared response S for all subjects relative to the same task; and (2) “Step b” assumes that features between similar tasks share the same within-subject representation denoted by Wi. We also investigated the between-subject noise ceiling of FastSRM using data pertaining to six runs of the Raiders task. Figure 5a shows that FastSRM could account for half of the noise ceiling in terms of correlation, in occipital, dorsal and temporal regions.
FastSRM-encoding study
Figure 5b shows the measures of performance—across subjects—of the FastSRM algorithm in terms of Pearson-correlation coefficients (see Section Naturalistic-Data Analysis and Fig. 4 for details), which were obtained between predicted and original data. These results yield functional regions that are synchronously activated across subjects for each task, because predicted data is herein estimated from a shared response to the same stimuli—learnt from the model—which is assumed to be observed in all individuals.
Higher significance was obtained throughout occipital areas for both Raiders and Clips tasks. Overall, these activations suggest a prominent recruitment of areas pertaining to the visual system52,65 during performance of both Raiders and Clips tasks. These areas comprise both early visual areas and middle temporal regions, including the MT + Complex, as well as associative visual areas from the dorsal and ventral streams66.
Similar activations were also found in the Inferior Parietal Lobe and posterior areas of the Superior Temporal Sulcus only for the Raiders task, indicating the additional recruitment of the auditory and language-comprehension systems during performance of this task. They also reflect the main behavioral differences underlying Clips and Raiders tasks, i.e. they highlight the fact that while both tasks refer to naturalistic visual stimuli, only Raiders refers to naturalistic audio stimuli.
To obtain a clear distinction of the regions exhibiting greater contributions for Raiders than Clips and vice-versa, we further inspected the z– maps from Fig. 5c depicting clusters whose difference of activations between the two tasks is significant. The results show a large number of regions where the activation level for Raiders is significantly higher than for Clips, while the extent of regions showing results in the opposite direction is extremely small. The identification of the functional territories covered by the clusters was determined in agreement with the parcellation of Glasser et al.52, which comprises 180 neocortical areas that are subsequently grouped in 22 main regions. Supplementary Table 1 presents, by descending order, the list of areas that contain a proportion of significant vertices larger than 5% in both hemispheres. The correspondence between area and main region was established according to the primary section in which the given area is described by the cortical parcellation (for further details, consult Table 1 of the Neuroanatomical Supplementary Results of Glasser et al.52). We identified 94 areas belonging to 21 main regions displaying higher activation for Raiders than Clips; no areas displaying higher activation for Clips than Raiders were observed above the same threshold. These results confirm the recruitment of additional brain networks necessary in the performance of cognitive tasks—namely auditory, speech and narrative comprehension—that are present in Raiders but not in Clips (see Section Experimental Paradigms for more details). Contributions to these networks come largely from the Auditory Association Cortex, Early Auditory Cortex, Temporo-Parieto-Occipital Junction, Posterior Cingulate Cortex, Parietal Cortex and Lateral Temporal Cortex (see Supplementary Table 1).
Retinotopy study
Figure 6 shows the retinotopic organization of the visual field in the human brain elicited by the Retinotopy tasks (see Sections Materials, Experimental Procedure, Retinotopy tasks for further details about the implementation of these classic retinotopy paradigms). The topographic projection to the V1-4 brain areas of the top-down and left-right reversed representation in the retina of the visual stimuli is mapped for every participant. As in Sereno et al.32, projections from the Wedge task (left column) represent the map of polar angle, whereas those from the Ring task (right column) represent the map of eccentricity. Overall, one can clearly notice a consistent spatial encoding of the visual field through these polar coordinates across all individuals.
In addition, we identified some areas of signal loss, for example in sub-05. After inspection, we concluded that this was probably due to signal correction in the most posterior part of the image during acquisition.
Usage Notes
Our results show that functional-imaging data featuring the third release of the IBC dataset reflect response to behavior during performance of the corresponding tasks. We also show the feasibility of extracting the same type of data derivatives—i.e. contrast maps—from tasks pertaining to different types of experimental designs. Concretely, we demonstrate that cognitive networks of functional data collected from naturalistic paradigms can be extracted using FastSRM—an unsupervised data-driven method—without explicitly modelling features of the stimuli. Due to various shortcuts described in Richard&Thirion (2023)34, the fast and memory-efficient implementation of FastSRM becomes very useful under high-dimensional regimes, such as modelling naturalistic stimuli. In addition, we demonstrate that results obtained for every task–which are adapted from previous studies–are in agreement with the ones originally reported.
Data collection ended in October 2023 and final releases are planned for this year and next year. Tasks featuring these releases will comprise not only other sensory modalities in greater depth but also high-order cognitive modules that will complement those from past releases. For instance, we plan to attain a better coverage of the auditory system with the inclusion of tasks on tonotopy, auditory language comprehension and listening of naturalistic sounds. Other tasks on biological motion, motor inhibition, finger tapping, visual perception (e.g. color, scenes and faces), stimulus salience, working memory, emotional memory, spatial navigation, risk-based decision making, reward processing, language and arithmetic processing will also integrate these future releases.
Although the IBC dataset is dedicated foremost to task-fMRI data, future releases will be also dedicated to resting-state fMRI data as well as to other MRI modalities, concretely high-resolution T1- and T2-weighted, diffusion-weighted and myelin water fraction.
The official website of the IBC dataset (https://individual-brain-charting.github.io/docs) can be consulted anytime for a continuous update about its releases.
Code availability
Metadata concerning the stimuli presented during the BOLD fMRI runs are publicly available at https://github.com/individual-brain-charting/public_protocols. They include: (1) the behavioral protocols of the tasks; (2) the demo presentations of the tasks as video-demo presentations; (3) the instructions to the participants; and (4) the scripts to extract paradigm descriptors from log files for the GLM estimation. Regarding the behavioral protocols, Clips task consists in a reproduction of the protocol featuring the original study, only with minor adjustments and most of them concerned with experimental settings; Retinotopy and Raiders tasks were re-written from scratch in Python with no change of the design referring to the original paradigms. Regarding the demo presentations of the Retinotopy tasks, they can also be found on the YouTube channel of the IBC project: https://www.youtube.com/@individualbraincharting6314.
The scripts used for data analysis are publicly available under the Simplified BSD license: https://github.com/individual-brain-charting/public_analysis_code.
The API code of the IBC data-fetcher from EBRAINS is available on a public GitHub repository at https://github.com/individual-brain-charting/api. This public repository includes also information about installation and a minimal example usage.
References
Genon, S., Reid, A., Langner, R., Amunts, K. & Eickhoff, S. B. How to Characterize the Function of a Brain Region. Trends Cogn Sci 22, 350–364, https://doi.org/10.1016/j.tics.2018.01.010 (2018).
Varoquaux, G. et al. Atlases of cognition with large-scale brain mapping. PLoS computational biology 14. https://doi.org/10.1371/journal.pcbi.1006565 (2018).
King, M., Hernandez-Castillo, C. R., Poldrack, R. A., Ivry, R. B. & Diedrichsen, J. Functional boundaries in the human cerebellum revealed by a multi-domain task battery. Nat Neurosci 22, 1371–1378, https://doi.org/10.1038/s41593-019-0436-x (2019).
Pinho, A. L. et al. Subject-specific segregation of functional territories based on deep phenotyping. Hum Brain Mapp 42, 841–870, https://doi.org/10.1002/hbm.25189 (2021).
Thirion, B., Thual, A. & Pinho, A. L. From deep brain phenotyping to functional atlasing. Curr Opin Behav Sci 40, 201–212, https://doi.org/10.1016/j.cobeha.2021.05.004 (2021). Deep Imaging - Personalized Neuroscience.
Thirion, B. et al. Analysis of a large fMRI cohort: Statistical and methodological issues for group analyses. Neuroimage 35, 105–120, https://doi.org/10.1016/j.neuroimage.2006.11.054 (2007).
Fedorenko, E., Behr, M. K. & Kanwishera, N. Functional specificity for high-level linguistic processing in the human brain. Proc Natl Acad Sci USA 108, 16428–33, https://doi.org/10.1073/pnas.1112937108 (2011).
Haxby, J. V. et al. A common, high-dimensional model of the representational space in human ventral temporal cortex. Neuron 72, 404–416, https://doi.org/10.1016/j.neuron.2011.08.026 (2011).
Nieto-Castañón, A. & Fedorenko, E. Subject-specific functional localizers increase sensitivity and functional resolution of multi-subject analyses. Neuroimage 63, 1646–1669, https://doi.org/10.1016/j.neuroimage.2012.06.065 (2012).
Laumann, T. O. et al. Functional System and Areal Organization of a Highly Sampled Individual Human Brain. Neuron 87, 657–670, https://doi.org/10.1016/j.neuron.2015.06.037 (2015).
Braga, R. M. & Buckner, R. L. Parallel Interdigitated Distributed Networks within the Individual Estimated by Intrinsic Functional Connectivity. Neuron 95, 457–471.e5, https://doi.org/10.1016/j.neuron.2017.06.038 (2017).
Gordon, E. M. et al. Precision functional mapping of individual human brains. Neuron 95, 791–807.e7, https://doi.org/10.1016/j.neuron.2017.07.011 (2017).
Chang, N. et al. BOLD5000, a public fMRI dataset while viewing 5000 visual images. Sci Data 6, 49, https://doi.org/10.1038/s41597-019-0052-3 (2019).
Nee, D. E. fMRI replicability depends upon sufficient individual-level data. Commun Biol 2, 1–4, https://doi.org/10.1038/s42003-018-0073-z (2019).
Ooi, L. Q. R. et al. MRI economics: Balancing sample size and scan duration in brain wide association studies. bioRxiv https://doi.org/10.1101/2024.02.16.580448 (2024).
Bayer, J. M. M. et al. Site effects how-to and when: An overview of retrospective techniques to accommodate site effects in multi-site neuroimaging analyses. Front Neurol 13. https://doi.org/10.3389/fneur.2022.923988 (2022).
Sonkusare, S., Breakspear, M. & Guo, C. Naturalistic stimuli in neuroscience: critically acclaimed. Trends Cogn Sci 23, 699–714, https://doi.org/10.1016/j.tics.2019.05.004 (2019).
Pinel, P. et al. Fast reproducible identification and large-scale databasing of individual functional cognitive networks. BMC neuroscience 8, 91, https://doi.org/10.1186/1471-2202-8-91 (2007).
Barch, D. M. et al. Function in the human connectome: Task-fMRI and individual differences in behavior. Neuroimage 80, 169–89, https://doi.org/10.1016/j.neuroimage.2013.05.033 (2013).
Orfanos, D. P. et al. The Brainomics/Localizer database. Neuroimage 144 (Pt B), 309–314, https://doi.org/10.1016/j.neuroimage.2015.09.052 (2017). Data Sharing Part II.
Pinel, P. et al. The functional database of the ARCHI project: Potential and perspectives. Neuroimage 197, 527–543, https://doi.org/10.1016/j.neuroimage.2019.04.056 (2019).
Bartels, A. & Zeki, S. Functional brain mapping during free viewing of natural scenes. Human Brain Mapping 21, 75–85, https://doi.org/10.1002/hbm.10153 (2004).
Hasson, U., Nir, Y., Levy, I., Fuhrmann, G. & Malach, R. Intersubject synchronization of cortical activity during natural vision. Science 303, 1634–1640, https://doi.org/10.1126/science.1089506 (2004).
Hanke, M. et al. A high-resolution 7-Tesla fMRI dataset from complex natural stimulation with an audio movie. Sci Data 1. https://doi.org/10.1038/sdata.2014.3 (2014).
Hanke, M. et al. High-resolution 7-Tesla fMRI data on the perception of musical genres–an extension to the studyforrest dataset. F1000Res 4, 174, https://doi.org/10.12688/f1000research.6679.1 (2015).
Hanke, M. et al. A studyforrest extension, simultaneous fMRI and eye gaze recordings during prolonged natural stimulation. Sci Data 3. https://doi.org/10.1038/sdata.2016.92 (2016).
Sengupta, A. et al. A studyforrest extension, retinotopic mapping and localization of higher visual areas. Sci Data 3. https://doi.org/10.1038/sdata.2016.93 (2016).
Thirion, B., Aggarwal, H., Ponce, A. F., Pinho, A. L. & Thual, A. Should one go for individual-or group-level brain parcellations? a deep-phenotyping benchmark. Brain Struct Funct 229, 161–181, https://doi.org/10.1007/s00429-023-02723-x (2024).
Pinho, A. L. et al. Individual Brain Charting, a high-resolution fMRI dataset for cognitive mapping. Sci Data 5, 180105, https://doi.org/10.1038/sdata.2018.105 (2018).
Pinho, A. L. et al. Individual Brain Charting dataset extension, second release of high-resolution fMRI data for cognitive mapping. Sci Data 7. https://doi.org/10.1038/s41597-020-00670-4 (2020).
Nishimoto, S. et al. Reconstructing visual experiences from brain activity evoked by natural movies. Curr Biol 21, 1641–6, https://doi.org/10.1016/j.cub.2011.08.031 (2011).
Sereno, M. et al. Borders of multiple visual areas in humans revealed by functional magnetic resonance imaging. Science 268, 889–893, https://doi.org/10.1126/science.7754376 (1995).
Poldrack, R. et al. The cognitive atlas: toward a knowledge foundation for cognitive neuroscience. Front Neuroinform 5, 17, https://doi.org/10.3389/fninf.2011.00017 (2011).
Richard, H. & Thirion, B. Fastsrm: A fast, memory efficient and identifiable implementation of the shared response model. Aperture Neuro 3, 1–13, https://doi.org/10.52294/001c.87954 (2023).
Gorgolewski, K. et al. The Brain Imaging Data Structure: a standard for organizing and describing outputs of neuroimaging experiments. Sci Data 3, 160044, https://doi.org/10.1038/sdata.2016.44 (2016).
Oldfield, R. C. The assessment and analysis of handedness: the Edinburgh inventory. Neuropsychologia 9, 97–113, https://doi.org/10.1016/0028-3932(71)90067-4 (1971).
Peirce, J. et al. PsychoPy2: Experiments in behavior made easy. Behav Res Methods 51, 195–203, https://doi.org/10.3758/s13428-018-01193-y (2019).
Krause, F. & Lindemann, O. Expyriment: A Python library for cognitive and neuroscientific experiments. Behav Res Methods 46, 416–428, https://doi.org/10.3758/s13428-013-0390-6 (2014).
Moeller, S. et al. Multiband multislice GE-EPI at 7 Tesla, with 16-fold acceleration using partial parallel imaging with application to high spatial and temporal whole-brain fMRI. Magn Reson Med 63, 1144–53, https://doi.org/10.1002/mrm.22361 (2010).
Feinberg, D. A. et al. Multiplexed echo planar imaging for sub-second whole brain fmri and fast diffusion imaging. PLoS One 5, 1–11, https://doi.org/10.1371/journal.pone.0015710 (2010).
Bischoff-Grethe, A. et al. A technique for the deidentification of structural brain MR images. Hum Brain Mapp 28, 892–903, https://doi.org/10.1002/hbm.20312 (2007).
Andersson, J. L., Skare, S. & Ashburner, J. How to correct susceptibility distortions in spin-echo echo-planar images: application to diffusion tensor imaging. Neuroimage 20, 870–888, https://doi.org/10.1016/S1053-8119(03)00336-7 (2003).
Smith, S. et al. Advances in functional and structural MR image analysis and implementation as FSL. Neuroimage 23(Suppl 1), S208–S219, https://doi.org/10.1016/j.neuroimage.2004.07.051 (2004).
Gorgolewski, K. et al. Nipype: A Flexible, Lightweight and Extensible Neuroimaging Data Processing Framework in Python. Front Neuroinform 5, 13, https://doi.org/10.3389/fninf.2011.00013 (2011).
Friston, K., Frith, C., Frackowiak, R. & Turner, R. Characterizing dynamic brain responses with fMRI: a multivariate approach. Neuroimage 2, 166–172, https://doi.org/10.1006/nimg.1995.1019 (1995).
Ashburner, J. & Friston, K. J. Unified segmentation. Neuroimage 26, 839–851, https://doi.org/10.1016/j.neuroimage.2005.02.018 (2005).
van Essen, D. C., Glasser, M. F., Dierker, D. L., Harwell, J. & Coalson, T. Parcellations and Hemispheric Asymmetries of Human Cerebral Cortex Analyzed on Surface-Based Atlases. Cereb Cortex 22, 2241, https://doi.org/10.1093/cercor/bhr291 (2012).
Fischl, B., Sereno, M. I., Tootell, R. B. & Dale, A. M. High-resolution intersubject averaging and a coordinate system for the cortical surface. Hum Brain Mapp 8, 272–284, https://doi.org/10.1002/(SICI)1097-0193(1999)8:43.0.CO;2-4 (1999).
Willems, R. M., Nastase, S. A. & Milivojevic, B. Narratives for neuroscience. Trends Cogn Sci 43, 271–273, https://doi.org/10.1016/j.tins.2020.03.003 (2020).
Wheatley, T., Boncz, A., Toni, I. & Stolk, A. Beyond the isolated brain: The promise and challenge of interacting minds. Neuron 103, 186–188, https://doi.org/10.1016/j.neuron.2019.05.009 (2019).
Wu, A., et al. (eds.) Advances in Neural Information Processing Systems (Curran Associates, Inc.), 31. https://proceedings.neurips.cc/paper/2018/file/17b3c7061788dbe82de5abe9f6fe22b3-Paper.pdf (2018).
Glasser, M. F. et al. A multi-modal parcellation of human cerebral cortex. Nature 536, 171–178, https://doi.org/10.1038/nature18933 (2016).
Mills, K. HCP-MMP1.0 projected on fsaverage. Figshare. https://doi.org/10.6084/m9.figshare.3498446.v2 (2016).
Abraham, A. et al. Machine learning for neuroimaging with scikit-learn. Front Neuroinform 8, 14, https://doi.org/10.3389/fninf.2014.00014 (2014).
Warnking, J. et al. fMRI Retinotopic Mapping—Step by Step. Neuroimage 17, 1665–1683, https://doi.org/10.1006/nimg.2002.1304 (2002).
Gao, J. S., Huth, A. G., Lescroart, M. D. & Gallant, J. L. Pycortex: an interactive surface visualizer for fMRI. Front Neuroinform 9, 23, https://doi.org/10.3389/fninf.2015.00023 (2015).
Amunts, K. et al. Linking brain structure, activity, and cognitive function through computation. eNeuro 9. https://doi.org/10.1523/ENEURO.0316-21.2022 (2022).
Pinho, A. L. et al. Individual Brain Charting (IBC). EBRAINS.v3.0. https://doi.org/10.25493/SM37-TS4 (2021).
Poldrack, R. et al. Toward open sharing of task-based fMRI data: the openfMRI project. Front Neuroinform 7, 12, https://doi.org/10.3389/fninf.2013.00012 (2013).
Pinho, A. L. et al. IBC. OpenNeuro. https://doi.org/10.18112/openneuro.ds002685.v1.3.1 (2021).
Gorgolewski, K. et al. NeuroVault.org: a web-based repository for collecting and sharing unthresholded statistical maps of the human brain. Front Neuroinform 9, 8, https://doi.org/10.3389/fninf.2015.00008 (2015).
Pinho, A. L. et al. NeuroVault. id collection = 6618. https://identifiers.org/neurovault.collection:6618 (2020).
Murphy, K., Bodurka, J. & Bandettini, P. A. How long to scan? The relationship between fMRI temporal signal to noise ratio and necessary scan duration. Neuroimage 34, 565–574, https://doi.org/10.1016/j.neuroimage.2006.09.032 (2007).
Power, J. D., Barnes, K. A., Snyder, A. Z., Schlaggar, B. L. & Petersen, S. E. Spurious but systematic correlations in functional connectivity mri networks arise from subject motion. Neuroimage 59, 2142–2154, https://doi.org/10.1016/j.neuroimage.2011.10.018 (2012).
Mantini, D. et al. Interspecies activity correlations reveal functional correspondence between monkey and human brain areas. Nat Methods 9, 277–282, https://doi.org/10.1038/nmeth.1868 (2012).
Contribution of striate inputs to the visuospatial functions of parieto-preoccipital cortex in monkeys. Behav Brain Res 6, 57–77. https://doi.org/10.1016/0166-4328(82)90081-X (1982).
Acknowledgements
We are very thankful to all volunteers who have accepted to be part of this ten-year long study, with many MRI repeated scans over this period of time. We thank to the Gallant Lab for having provided the behavioral protocol and stimuli of naturalistic scenes used to develop the Clips task. We also thank to the Center for Magnetic Resonance Research, University of Minnesota for having kindly provided the Multi-Band Accelerated EPI Pulse Sequence and Reconstruction Algorithms. This project has received funding from the European Union’s Horizon 2020 Framework Program for Research and Innovation under Grant Agreement No 720270 (Human Brain Project SGA1), 785907 (Human Brain Project SGA2) and 945539 (Human Brain Project SGA3). Ana Luísa Pinho is the recipient of a BrainsCAN Postdoctoral Fellowship at Western University, funded by the Canada First Research Excellence Fund (CFREF).
Author information
Authors and Affiliations
Contributions
A.L.P. set the task protocols, performed the MRI acquisitions, performed the analyses of the naturalistic-task fMRI data, post-processed the behavioral data, produced video-demo presentations of the behavioral protocols, contributed to the IBC public analysis pipeline, contributed to Nilearn, contributed to Pypreprocess, wrote the IBC documentation and wrote the paper. H.R. developed FastSRM. A.F.P. estimated the QA metrics of this dataset extension, performed the analyses of the naturalistic-task fMRI data, wrote the documentation and contributed to the IBC public analysis pipeline. M.E. designed the Retinotopy tasks, help setting the Clips task and contributed to Nilearn. A.A. set the MRI sequences. E.D. developed Pypreprocess and Nilearn. I.D. assisted in the implementation of the task protocols, performed some MRI Sessions and contributed to Pypreprocess. J.J.T. wrote the IBC documentation, performed some MRI Sessions, contributed to the IBC public analysis pipeline and contributed to Nilearn. S.S. wrote the IBC documentation, contributed to Pypreprocess, contributed to the IBC public analysis pipeline and assisted on preliminary analyses of naturalistic-task fMRI data. H.A. wrote the documentation, contributed to Pypreprocess and contributed to the IBC public analysis pipeline. A.T. contributed to the IBC public analysis pipeline and contributed to Nilearn. T.C. conceptualized Fig. 4. C.G., S.B.-D., S.R.,Y.L. and V.B. performed the MRI acquisitions plus visual inspection of the neuroimaging data for quality-checking. L.L., V.J.-T. and G.M.-C. recruited the participants and managed routines related to appointment scheduling and ongoing medical assessment. C.D. recruited the participants and managed the scientific communication of the project with them. B.M. managed regulatory issues. G.V. developed Nilearn, contributed to Pypreprocess and advised on the study design plus analysis pipeline. S.D. conceived the general design of the study. L.H.-P. conceived the general design of the study and wrote the ethical protocols. B.T. conceived the general design of the study, managed the project, wrote the ethical protocols, performed some MRI acquisitions, pre-processed fMRI data, performed the analyses of the retinotopy fMRI data, developed FastSRM, developed Nilearn, developed Pypreprocess, developed the IBC public analysis pipeline, uploaded IBC collections on EBRAINS + OpenNeuro + NeuroVault, wrote the IBC documentation and wrote the paper.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Pinho, A.L., Richard, H., Ponce, A.F. et al. Individual Brain Charting dataset extension, third release for movie watching and retinotopy data. Sci Data 11, 590 (2024). https://doi.org/10.1038/s41597-024-03390-1
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41597-024-03390-1
- Springer Nature Limited