= Resting state recordings from the OMEGA database (Original) = ''Authors: Francois Tadel, Guiomar Niso, Elizabeth Bock, Sylvain Baillet '' This tutorial introduces how to download resting state recordings from the OMEGA database and process them into Brainstorm. The goal is to evaluate where in the brain is distributed the power in specific frequency bands at rest. This example includes only 5 healthy subjects, but the same pipeline can be applied over larger populations. Note that the operations used here are not detailed, the goal of this tutorial is not to introduce Brainstorm to new users. For in-depth explanations of the interface and theoretical foundations, please refer to the [[http://neuroimage.usc.edu/brainstorm/Tutorials#Get_started|introduction tutorials]]. <> == OMEGA database == The Open MEG Archive (OMEGA) is the fruit of a collaborative effort by the McConnell Brain Imaging Centre and the Université de Montréal to build a centralized repository in which to regroup MEG data for open dissemination. This continuously expanding repository also contains anatomical MRI volumes, demographic and questionnaire information, and eventually will feature other forms of electrophysiological data (e.g. EEG, field, and cell recordings). OMEGA features the technological framework for multi-centric data aggregation, and is amongst the largest freely available resting-state MEG datasets presently available. Website: https://www.mcgill.ca/bic/resources/omega ''' '''You are free to use all data in OMEGA for research purposes; however, we ask that you '''please cite the following reference''' in your publications if you have used data from OMEGA: Niso G, Rogers C, Moreau JT, Chen LY, Madjar C, Das S, Bock E, Tadel F, Evans A, Jolicoeur P, Baillet S<
>[[http://dx.doi.org/10.1016/j.neuroimage.2015.04.028|OMEGA: The Open MEG Archive]], Neuroimage, 2015. == Presentation of the experiment == ==== Experiment ==== * 5 subjects x 5 minute resting sessions, eyes open ==== MEG acquisition ==== * Acquisition at '''2400Hz''', with a '''CTF 275''' MEG system, * Recorded at the Montreal Neurological Institute in 2015-2016. * Anti-aliasing low-pass filter at 600Hz * Files may be saved with or without the CTF 3rd order gradient compensation applied. * Recorded channels (at least 297): * 26 MEG reference sensors (#2-#27) * 270 MEG axial gradiometers (#28-#297) * 1 ECG bipolar (EEG057/#298) - Not available in the empty room recordings * 1 vertical EOG bipolar (EEG058/#299) - Not available in the empty room recordings * 1 horizontal EOG bipolar (EEG059/#300) - Not available in the empty room recordings * The data in this dataset has been organized according to the BIDS specifications<
>(Brain Imaging Data Structure, http://bids.neuroimaging.io) ==== Head shape and fiducial points ==== 3D digitization using a Polhemus Fastrak device driven by Brainstorm. The .pos files contain: * The center of the CTF coils, * The anatomical references we use in Brainstorm: nasion and ears as illustrated [[http://neuroimage.usc.edu/brainstorm/CoordinateSystems#Pre-auricular_points_.28LPA.2C_RPA.29|here]], * Around 100 head points distributed on the hard parts of the head (no soft tissues). ==== Subject anatomy ==== * Structural T1 image (defaced for anonymization purposes) * Processed with FreeSurfer 5.3 * The anatomical fiducials (NAS, LPA, RPA) have already been marked and saved in the files fiducials.m == Download and installation == * First, make sure you have at least '''40Gb''' of free space on your hard drive (raw files: 12Gb). * The OMEGA database is not in it its final release stage, so we temporarily distribute the 5 subjects used in this tutorial from the Brainstorm website. * Go to the [[http://neuroimage.usc.edu/bst/download.php|Download]] page, and download the file: '''sample_omega.zip ''' * Unzip it in a folder that is not in any of the Brainstorm folders (program folder or database folder). * Start Brainstorm (Matlab scripts or stand-alone version). For help, see the [[http://neuroimage.usc.edu/brainstorm/Installation|Installation]] page. * Select the menu File > Create new protocol. Name it "'''TutorialOmega'''" and select the options: * "'''No, use individual anatomy'''", * "'''No, use one channel file per condition'''". == BIDS specifications == The [[http://bids.neuroimaging.io|Brain Imaging Data Structure (BIDS)]] standard for neuroimaging data organization was first established for MRI and fMRI ([[http://www.nature.com/articles/sdata201644|Gorgolewski, 2016]]). It is based on simple file formats and folder structures that can readily expand to additional data modalities. An extension of the BIDS for MEG datasets is currently under review. The data distributed with this tutorial is organized following this prototype (which may change in the future). One objective is that analysis pipelines designed with major analysis tools (such as Brainstorm, FieldTrip, MNE, SPM and others) can be readily applied without requiring software or pipeline redevelopments. We have already implemented a set of functions that can load automatically into a Brainstorm database any dataset following the BIDS specification. The files that are read by this function are the following: '''sample_omega'''/ * '''derivatives'''/: Everything that cannot be considered as raw data * '''freesurfer'''/: Result of the FreeSurfer segmentation for all the subjects * '''sub-000X_ses-0001'''/: Output of FreeSurfer for subject 000X (session 0001) * '''sub-000X'''/: Raw data for subject 000X * '''ses-0001'''/: Only one session per subject (the subject visited the center only once) * '''anat'''/: Anatomical MRI scans for subject 000X (not used if FreeSurfer output available) * '''sub-000X_T1w.nii.gz''': T1 scan * '''meg'''/: Unprocessed MEG recordings * '''sub-000X_task-rest_run-01_meg.ds''': Resting MEG recordings (subject data) * '''sub-000X_task-noise_run-01_meg.ds''': Empty room recordings (noise measure) == Import the dataset == * Select the menu File > Load protocol > '''Import BIDS dataset''' > Select the folder '''sample_omega'''. * Keep the default values for all the questions that may be asked during the import process (eg. number of vertices in the cortex surfaces). Once done, you should be able to access the data for the 5 subjects in your database explorer: anatomy, and subject and noise recordings. <
><
> {{attachment:load_bids.gif||width="594",height="322"}} * For this example dataset, the fiducials NAS/LPA/RPA were previously marked using the menu '''File > Batch MRI fiducials''', which creates a file '''fiducials.m''' in the FreeSurfer folder. When importing the FreeSurfer folders, these fiducials are used by default. * If you want to use this feature on your own BIDS-formatted dataset, using the menu Import BIDS dataset may ask you to select for each subject the position of the fiducials NAS/LPA/RPA. You can mark them in advance using the menu Batch MRI fiducials as we did. * Another faster solution, if you have access to digitized head shapes for all the subjects in the MEG files: skip the marking of the fiducials, use default positions estimated from MNI coordinates, and use the head shape for refining the registration. To force the BIDS importer to work this way, use the process "'''Import anatomy > Import BIDS dataset'''" instead of the interactive menu, with the option "Align sensors using head points" selected. == Pre-processing == * Switch to the functional view of the protocol (second button above the database explorer). * Drag and drop all the subjects into the Process1 box, and click on [Run]. * Select process: '''Import recordings >''' '''Convert to continuous''': Continuous * Add process: '''Frequency >''' '''Power spectrum density (Welch)''': All file, 4s, 50% overlap, Individual <
><
> {{attachment:psd_process.gif||width="573",height="325"}} * Run, then check the quality of the PSD for all the files, as documented in [[http://neuroimage.usc.edu/brainstorm/Tutorials/ArtifactsFilter#Interpretation_of_the_PSD|this tutorial]]. <
><
> {{attachment:psd_before.gif||width="343",height="152"}} * In Process1, keep the same files selected and click on [Run]. * Select process: '''Pre-process >''' '''Notch filter''': 60 120 180 240 300 Hz, Process the entire file at once * Add process: '''Pre-process >''' '''Band-pass filter''': High-pass filter at 0.3Hz, 60dB, Process entire file. * Add process: '''Frequency >''' '''Power spectrum density (Welch)''': Same options <
><
> {{attachment:filter_process.gif||width="597",height="342"}} * Run, then check the PSD after filtering. <
><
> {{attachment:psd_after.gif||width="389",height="140"}} * Delete all the folders corresponding to the original recordings (_meg) and the notch filtered data (_meg_notch). Only keep the fully processed files (_meg_notch_high). <
><
> {{attachment:delete_original.gif||width="429",height="169"}} {{attachment:delete_after.gif||width="228",height="263"}} == Artifact cleaning == We will now run the automatic procedure for cleaning the heartbeats, as described in the introduction tutorials ([[Tutorials/ArtifactsDetect|detection]], [[Tutorials/ArtifactsSsp|SSP]]). The results we want to illustrate here are robust enough, the recordings do not need to be processed any further. If you want to improve the quality of the data with more manual cleaning (blinks, saccades, movements, bad segments), please refer to the introduction tutorials. * In Process1, select all the subject rest recordings ('''task-rest'''). * Select process: '''Events > Detect heartbeats''': ECG, All file, cardiac * Add process: '''Artifacts > SSP: Heartbeats''': cardiac, MEG, Use existing SSP * Add process: '''File > Snapshot''': Sensor/MRI registration * Add process: '''File > Snapshot''': SSP projectors<
><
> {{attachment:ssp_process.gif||width="548",height="304"}} * Run, then review the registration and the SSP topographies for all the subjects. Make sure that the subjects' heads are well aligned in the MEG, and that the spatial components that are removed are really representative of the cardiac artifact. <
><
> {{attachment:ssp_report.gif||width="489",height="378"}} == Source estimation == * In Process1, select all the noise recordings ('''task-noise'''). * Run process: '''Sources > Compute covariance''': All file, Noise covariance, Copy to other folders. <
><
> {{attachment:noiscov_process.gif||width="584",height="468"}} * In Process1, select all the subject rest recordings ('''task-rest'''). * Run process: '''Sources > Compute head model''': Cortex surface, MEG=Overlapping spheres. <
><
> {{attachment:headmodel_process.gif||width="574",height="423"}} * Run process: '''Sources > Compute sources [2016]''': Kernel only, one per file, dSPM, constrained. <
><
> {{attachment:inverse_process.gif||width="629",height="289"}} == Power maps == We have now access to continuous recordings in source space, we can estimate the average power over this resting period, for various frequency bands. We propose here to compute the PSD from the first 100s instead of the full 600s rest recordings, it is faster and leads to very similar results. For more stable estimates, you can compute the the PSD over all the data available instead. * In Process1, select all the subject rest recordings ('''task-rest'''), click the '''[Process sources]''' button. * Select process: '''Frequency > Power spectrum density (Welch)''': [0,100s], Window=4s, 50% overlap, Group in frequency bands (use the default frequency bands), Save individual PSD values. * Add process: '''Standardize > Spectrum normalization''': Relative power (divide by total power). This normalization is done independently for each source, the output for each frequency band is a percentage of how much this frequency band contributes to the total power of the source signal. * Add process: '''Sources > Project on default anatomy''': Cortex. * Add process: '''Sources > Spatial smoothing''': FWHM=3mm, Overwrite.<
><
> {{attachment:psd_sources_process.gif||width="656",height="359"}} <
><
> {{attachment:psd_sources_normalize.gif||width="634",height="330"}} * Execute this pipeline, then open the results for the different subjects and visualize the different frequency bands. They should look similar across subjects. This measure of relative power is very robust and usually do not vary much between subjects or duration of the input signals. * The values in this figure represent the relative power of each source. At each vertex of the cortex surface, the value between 0 and 1 indicates what is the contribution of the current frequency band to the total power in the signal. To obtain comparable colormaps across multiple figures, you can set a fixed maximum (right-click on a figure > Colormap: Timefreq > Maximum: Custom).<
><
> {{attachment:psd_sources_proj.gif||width="667",height="278"}} * We will now average these files across subjects. In Process1, select all the projected subject averages in the ''Group analysis'' folder, and click on the button '''[Process time-freq]'''. * Run process: '''Average > Average files''': Everything, Arithmetic average, Do not match signals <
><
> {{attachment:psd_sources_avgsubj.gif||width="489",height="459"}} * Double-click on the average to display it. To create the figures below, right-click on the figure > Snapshot > Frequency contact sheet. <
><
> {{attachment:psd_sources_left.gif||width="353",height="152"}} {{attachment:psd_sources_top.gif||width="279",height="179"}} * We can relate these figures with associations commonly reported in the literature: * '''Theta''': Mostly frontal. * '''Alpha''': Parieto-occipital. * '''Beta''': Central regions (motor cortex). * '''Gamma''': Pre-frontal and occipital. However, the occipital sources could be related with muscle activity in the neck. We haven't cleaned the recordings much, we should go back to the recordings and mark as bad all the segments in which we observe a lot of muscle activity. * '''Delta''': Questionable - these source maps showing high values at the poles of the frontal and temporal lobes are typical of eye movements. We should try doing some cleaning of the ocular artifacts (blinks and saccades) and compute these maps again. * You can compute the same results in the head volume. You just have to use a volume head model on a group grid instead of the one we used here. This is explained in the tutorial [[http://neuroimage.usc.edu/brainstorm/Tutorials/CoregisterSubjects#Volume_source_models|Group analysis]]. This seems to confirm that the power in the delta and gamma bands is mostly coming from non-brain regions (eyes and neck muscles). <
><
> {{attachment:psd_volume_delta.gif||width="297",height="293"}} {{attachment:psd_volume_gamma.gif||width="297",height="294"}} * You can produce similar plots at the sensor level: Compute the PSD by frequency band of the rest recordings for each subject, then compute the average across subjects (not as reliable as averaging in source space, as [[http://neuroimage.usc.edu/brainstorm/Tutorials/Workflows#Average-1|explained here]]). <
><
> {{attachment:psd_avgsubj.gif||width="636",height="250"}} == Resting state metrics [TODO] == Discuss what other measures can be used in resting state MEG. Add references to a few recent publications. == Scripting == The following script from the Brainstorm distribution reproduces the analysis presented in this tutorial page: [[https://github.com/brainstorm-tools/brainstorm3/blob/master/toolbox/script/tutorial_omega.m|brainstorm3/toolbox/script/tutorial_omega.m]] <)>><><)>> <>