Scheduled service maintenance on November 22


On Friday, November 22, 2024, between 06:00 CET and 18:00 CET, GIN services will undergo planned maintenance. Extended service interruptions should be expected. We will try to keep downtimes to a minimum, but recommend that users avoid critical tasks, large data uploads, or DOI requests during this time.

We apologize for any inconvenience.

설명 없음

Jack 10ec6c1499 link to preprint 1 주 전
custom_epitran_files 7a0df466b2 add figs 1 주 전
eeg 7a0df466b2 add figs 1 주 전
env 7a0df466b2 add figs 1 주 전
epo 7a0df466b2 add figs 1 주 전
estimates 7a0df466b2 add figs 1 주 전
fig 7a0df466b2 add figs 1 주 전
fig_code 56c629a02f add fig code 1 주 전
noise_ceiling c934945c24 add noise ceiling 1 주 전
rdm_data f269690761 add poit rdms for all channels 1 주 전
scold 7a0df466b2 add figs 1 주 전
stim_sim 7a0df466b2 add figs 1 주 전
00_get_orth_model_rdms.py 7a0df466b2 add figs 1 주 전
01_get_corpus_model_rdms.py 7a0df466b2 add figs 1 주 전
02_preprocessing.py 3a32e46a35 record total chs interpolated 1 주 전
03_get_neural_rdms.py c5a6112a15 address copy_from_bbox error 1 주 전
04_rsa.R 7a0df466b2 add figs 1 주 전
05_ot_geom.R 7a0df466b2 add figs 1 주 전
06_jacc_geom.R 7a0df466b2 add figs 1 주 전
07_rsa_anns.R 7a0df466b2 add figs 1 주 전
08_controls_rsa.R 7a0df466b2 add figs 1 주 전
09_controls_rsa_all_chs.R 7a0df466b2 add figs 1 주 전
10_sensitivity_analysis.R 7a0df466b2 add figs 1 주 전
AC-64.bvef 7a0df466b2 add figs 1 주 전
Arial-Lgt.ttf 7a0df466b2 add figs 1 주 전
Dockerfile 7a0df466b2 add figs 1 주 전
Helvetica.ttf 7a0df466b2 add figs 1 주 전
LICENSE 975e1e631f Initial commit 1 주 전
README.md 10ec6c1499 link to preprint 1 주 전
SUBTLEX-DE_cleaned_with_Google00_frequencies.csv 7a0df466b2 add figs 1 주 전
arial.ttf 7a0df466b2 add figs 1 주 전
preprocessing_metadata.csv 98909c508a include metadata output 1 주 전
project_tree.md bd904b7800 add proj tree 1 주 전
run.sh 7a0df466b2 add figs 1 주 전
transcribed_phonological_letter_name_vectors.csv 7a0df466b2 add figs 1 주 전

README.md

Alphabetic Decision Task Analysis

This repository contains code for rerunning the analyses reported in our preprint:

Taylor, J.E., Sinn, R., Iaia, C., & Fiebach, C. J. (2024). Beyond Letters: Optimal Transport as a Model for Sub-Letter Orthographic Processing. https://doi.org/10.1101/2024.11.11.622929

Running the Analysis on Docker

The easiest way to re-run these analyses without worrying about package versions is using a Docker container. For this you will need a Docker installation.

1. Build the Docker Image

Set the working directory to the directory containing Dockerfile, then build an image, e.g., called lettersim:

docker image build -t lettersim .

This requires internet access. Building the image will automatically download the raw data in BIDS format from OpenNeuro (https://openneuro.org/datasets/ds005594) into the container.

2. Mount a Volume

Create a volume, e.g., called vol1, from which any outputs can be retrieved:

docker volume create vol1

3. Run a Script in a Container

Run Script Non-Interactively

Create a container from the Docker image to run the script you want to use, mounted to the volume you just created, e.g., here we run 01_get_corpus_model_rdms.py in a container named LS1. The last part, ./run.sh 01_get_corpus_model_rdms.py, tells the container which script to run. Passing it through run.sh will ensure that the correct environment/interpreter is used.

docker container run -t --mount source=vol1,target=/analysis --name=LS1 lettersim ./run.sh 01_get_corpus_model_rdms.py

Note: see Docker documentation for available options. For example, you can use --cpus to limit the number of CPUs available to the container.

Run Script Interactively

Alternatively, you can run scripts interactively, e.g. start an interactive shell:

docker container run -it --mount source=vol1,target=/analysis --name=LSINT lettersim bash

And then run a script through run.sh:

./run.sh 01_get_corpus_model_rdms.py

Accessing Data from the Container

To get any output files, you will need to access the mountpoint of the volume that the container is using. By default, volumes can be accessed at /var/lib/docker/volume.

You can also find the location listed in the volume's configuration:

docker volume inspect vol1
[
    {
        "CreatedAt": "1980-1-1T00:00:01Z",
        "Driver": "local",
        "Labels": null,
        "Mountpoint": "/var/lib/docker/volumes/vol1/_data",
        "Name": "vol1",
        "Options": null,
        "Scope": "local"
    }
]

If using Docker Desktop, you can also browse the files in the Volumes tab.

The actual location used by Docker Desktop varies by Docker and WSL version. On Docker Desktop v4.34.3 on Windows 11, data in a volume called vol1 can be found at

\\wsl.localhost\docker-desktop-data\data\docker\volumes\vol1\_data

Running on HPC

As a docker image, this analysis can be run on high-performance computing (HPC) clusters via Singularity: https://docs.sylabs.io/guides/2.6/user-guide/singularity_and_docker.html

Running the Analysis without Docker

If you prefer not to use Docker, you can run the scripts outside of a container. These instructions assume you have git and a conda distribution installed.

Download the BIDS Dataset

Download the dataset from OpenNeuro, saving it to a directory called eeg/:

git clone https://github.com/OpenNeuroDatasets/ds005594 eeg

Install Environments

There are three conda environments used to run the scripts:

  • mne Python environment for working with EEG data and calculating neural RDMs (conda .yml file in env/environment-mne.yml)

  • rdms Python environment for calculating model RDMs used in the RSA (conda .yml file in env/environment-rdms.yml)

  • r R environment for running all .R scripts (conda .yml file in env/environment-r.yml)

You can install these environments with:

conda env create --file=env/environment-mne.yml --force
conda env create --file=env/environment-rdms.yml --force
conda env create --file=env/environment-r.yml --force

Setup for Reproducing Figures

You may need to run these steps if you want to use scripts in fig_code/ to reproduce figures.

Install LaTeX

Soms scripts in fig_code/ use LaTeX. We use TexLive.

If TexLive is installed, you can use tlmgr to install all the TeX packages that we use:

tlmgr install xcolor tex-gyre underscore etoolbox pgf

Add Fonts to matplotlib

You may need to add fonts to matplotlib. You can do this using:

conda run -n mne fig_code/mpl_setup_fonts.py
conda run -n rdms fig_code/mpl_setup_fonts.py

Running Scripts

If conda is on path, and the environments are set up as in the .yml files (including environment names), you can use the script run.sh to automatically run a script with the correct environment and interpreter, e.g.:

./run.sh 01_get_corpus_model_rdms.py

General Notes on the Analysis

  • All code is intended to be run with the base directory (i.e., the same directory as run.sh) as the working directory.

  • Scripts are numbered based on the order we ran them in. Scripts 00 to 03 should be strictly run in order to reproduce all steps, as they depend on outputs from the previous files. Scripts 04 and up can then be run in any order or skipped.

  • All outputs are included in this repository, so you can run any script in isolation.

  • Scripts for creating figures are in the fig_code directory, with the 99 prefix.

  • The repository structure is listed and explained in project_tree.md