[![made-with-datalad](https://www.datalad.org/badges/made_with.svg)](https://datalad.org) # **EEG Preprocessing** Steps 1-4 create preprocessing information (e.g., from the ICA, segment labeling), that later is applied to the raw data starting in step 5. All steps use FieldTrip and were executed within MATLAB 2020a. Deviations from the standard pipeline: - no ECG was available, which would normally guide heart IC labeling - all scripts have been adapted to work on a single run Many scripts were run on the high performance computing cluster (HPC) at the Max Planck Institute for Human Development (Berlin, Germany). One of two different deployment approaches was used: either (a) the relevant script was compiled or (b) the script was directly called with multiple MATLAB instances. In the case of (a), the respective code folder will contain a "_prepare" file that was used to compile the code, as well as a '_START' bash script, that deployed the compiled code; in the case of (b) bo compiling was done, and the '_START' script immediately calls the relevant MATLAB code. To re-run scripts outside of the HPC environment, a script needs to be written that loops the script across all subjects. Note that for debugging purposes, code usually checks whether it is run on a mac, and if so, performs all computations on an example subject. Depending on the deployment situation, this section may need to be adapted. --- ### **a1_prepare_preprocessing** - Prepare for ICA - Read into FieldTrip format - Referenced to avg. mastoid (A1, A2) - downsample: 1000Hz to 250 Hz - 4th order Butterworth 1-100 Hz BPF ### **a2_visual_inspection** - Manual labeling of noise periods that will be excluded prior to ICA ### **a3_ica** - prior to ICA, data are segmented into 2 s pseudo-trials - ICA is conducted - use VEOG, HEOG and ECG (not available) to pre-label blink, eye move and heart ICs ### **a4_ica_labeling** - Manual labeling of artefactual ICA components ### **a5_segmentation_raw_data** This step loads the raw data again and segments them to the desired time window, prior filters will be applied to these data in the next step. - load raw data - segment to -1s to +2s relative to stimulus onset - EEG settings: - Referenced to avg. mastoid (A1, A2) - recover implicit reference: POz - 0.2 4th order butterworth HPF - 125 4th order butterworth LPF - demean - downsample: 512 Hz to 500 Hz ### **a6_automatic_artifact_correction** Identify additional artifacts after removing ICA components. This step does NOT yet remove anything. We only calculate the data to be removed in the next step. - get artifact contaminated channels by kurtosis, low & high frequency artifacts - get artifact contaminated channels by FASTER - interpolate artifact contaminated channels - get artifact contaminated epochs & exclude epochs recursively - get channel x epoch artifacts ### **a7_prep_data_for_analysis** - Remove blink, move, heart, ref, art & emg ICA components - Remove artifact-heavy trials - Interpolate artifact-heavy channels - output: preprocessed data: 'sub-XXX_task-xxxx_eeg_art.mat' --- # **DataLad datasets and how to use them** This repository is a [DataLad](https://www.datalad.org/) dataset. It provides fine-grained data access down to the level of individual files, and allows for tracking future updates. In order to use this repository for data retrieval, [DataLad](https://www.datalad.org/) is required. It is a free and open source command line tool, available for all major operating systems, and builds up on Git and [git-annex](https://git-annex.branchable.com/) to allow sharing, synchronizing, and version controlling collections of large files. You can find information on how to install DataLad at [handbook.datalad.org/en/latest/intro/installation.html](http://handbook.datalad.org/en/latest/intro/installation.html). ### Get the dataset A DataLad dataset can be `cloned` by running ``` datalad clone ``` Once a dataset is cloned, it is a light-weight directory on your local machine. At this point, it contains only small metadata and information on the identity of the files in the dataset, but not actual *content* of the (sometimes large) data files. ### Retrieve dataset content After cloning a dataset, you can retrieve file contents by running ``` datalad get ` ``` This command will trigger a download of the files, directories, or subdatasets you have specified. DataLad datasets can contain other datasets, so called *subdatasets*. If you clone the top-level dataset, subdatasets do not yet contain metadata and information on the identity of files, but appear to be empty directories. In order to retrieve file availability metadata in subdatasets, run ``` datalad get -n ``` Afterwards, you can browse the retrieved metadata to find out about subdataset contents, and retrieve individual files with `datalad get`. If you use `datalad get `, all contents of the subdataset will be downloaded at once. ### Stay up-to-date DataLad datasets can be updated. The command `datalad update` will *fetch* updates and store them on a different branch (by default `remotes/origin/master`). Running ``` datalad update --merge ``` will *pull* available updates and integrate them in one go. ### Find out what has been done DataLad datasets contain their history in the ``git log``. By running ``git log`` (or a tool that displays Git history) in the dataset or on specific files, you can find out what has been done to the dataset or to individual files by whom, and when. ### More information More information on DataLad and how to use it can be found in the DataLad Handbook at [handbook.datalad.org](http://handbook.datalad.org/en/latest/index.html). The chapter "DataLad datasets" can help you to familiarize yourself with the concept of a dataset.