Linescanning - Package for (pre)processing of anatomical and (linescanning) fMRI data

Overview

line scanning repository

plot

This repository contains all of the tools used during the acquisition and postprocessing of line scanning data at the Spinoza Centre for Neuroimaging in Amsterdam. The script master controls the modules prefixed by spinoza_, which in turn call upon various scripts in the utils and bin directory. The scripts in the latter folders are mostly helper scripts to make life a tad easier. The repository contains a mix of languages in bash, python, and matlab.

In active development - do not use unless otherwise instructed by repo owners

Documentation for this package can be found at readthedocs (not up to date)

Policy & To Do

  • install using python setup.py develop
  • Docstrings in numpy format.
  • PEP8 - please set your editor to autopep8 on save!
  • Documentation with Sphinx (WIP)
  • Explore options to streamline code
  • Examples of applications for package (integration of pycortex & pRFpy)

overview of the pipeline

how to set up

Clone the repository: git clone https://github.com/gjheij/linescanning.git.

To setup the bash environment, edit setup file linescanning/shell/spinoza_setup:

  • line 76: add the path to your matlab installation if available (should be, for better anatomicall preprocessing)
  • line 87: add the path to your SPM installation
  • line 92: add your project name
  • line 97: add the path to project name as defined in line 92
  • line 102: add whether you're using (ME)MP(2)RAGE. This is required because the pipeline allows the usage of the average of an MP2RAGE and MP2RAGEME acquisition
  • line 105: add which type of data you're using (generally this will be the same as line 102)

Go to linescanning/shell and hit ./spinoza_setup setup setup. This will print a set of instructions that you need to follow. If all goes well this will make all the script executable, set all the paths, and install the python modules. The repository comes with a conda environment file, which can be activated with: conda create --name myenv --file environment.yml.

How to plan the line

plot

We currently aim to have two separate sessions: in the first session, we acquire high resolution anatomical scans and perform a population receptive field (pRF-) mapping paradigm (Dumoulin and Wandell, 2008) to delineate the visual field. After this session, we create surfaces of the brain and map the pRFs onto that via fMRIprep and pRFpy. We then select a certain vertex based on the parameters extracted from the pRF-mapping: eccentricity, size, and polar angle. Using these parameters, we can find an optimal vertex. We can obtain the vertex position, while by calculating the normal vector, we obtain the orientation that line should have (parellel to the normal vector and through the vertex point). Combining this information, we know how the line should be positioned in the first session anatomy. In the second session, we first acquire a low-resolution MP2RAGE with the volume coil. This is exported and registered to the first session anatomy during the second session to obtain the translations and rotations needed to map the line from the first session anatomy to the currently active second session by inputting the values in the MR-console. This procedure from registration to calculation of MR-console values is governed by spinoza_lineplanning and can be called with master -m 00 -s -h .

Owner
Jurjen Heij
Jurjen Heij
High-resolution networks and Segmentation Transformer for Semantic Segmentation

High-resolution networks and Segmentation Transformer for Semantic Segmentation Branches This is the implementation for HRNet + OCR. The PyTroch 1.1 v

HRNet 2.8k Jan 07, 2023
[ICCV'2021] Image Inpainting via Conditional Texture and Structure Dual Generation

[ICCV'2021] Image Inpainting via Conditional Texture and Structure Dual Generation

Xiefan Guo 122 Dec 11, 2022
RID-Noise: Towards Robust Inverse Design under Noisy Environments

This is code of RID-Noise. Reproduce RID-Noise Results Toy tasks Please refer to the notebook ridnoise.ipynb to view experiments on three toy tasks. B

Thyrix 2 Nov 23, 2022
A unified 3D Transformer Pipeline for visual synthesis

Overview This is the official repo for the paper: NÜWA: Visual Synthesis Pre-training for Neural visUal World creAtion. NÜWA is a unified multimodal p

Microsoft 2.6k Jan 06, 2023
Unofficial Implementation of Oboe (SIGCOMM'18').

Oboe-Reproduce This is the unofficial implementation of the paper "Oboe: Auto-tuning video ABR algorithms to network conditions, Zahaib Akhtar, Yun Se

Tianchi Huang 13 Nov 04, 2022
code for CVPR paper Zero-shot Instance Segmentation

Code for CVPR2021 paper Zero-shot Instance Segmentation Code requirements python: python3.7 nvidia GPU pytorch1.1.0 GCC =5.4 NCCL 2 the other python

zhengye 86 Dec 13, 2022
Deep Ensembling with No Overhead for either Training or Testing: The All-Round Blessings of Dynamic Sparsity

[ICLR 2022] Deep Ensembling with No Overhead for either Training or Testing: The All-Round Blessings of Dynamic Sparsity by Shiwei Liu, Tianlong Chen, Zahra Atashgahi, Xiaohan Chen, Ghada Sokar, Elen

VITA 18 Dec 31, 2022
Python binding for Khiva library.

Khiva-Python Build Documentation Build Linux and Mac OS Build Windows Code Coverage README This is the Khiva Python binding, it allows the usage of Kh

Shapelets 46 Oct 16, 2022
Tello Drone Trajectory Tracking

With this library you can track the trajectory of your tello drone or swarm of drones in real time.

Kamran Asgarov 2 Oct 12, 2022
Diverse Image Captioning with Context-Object Split Latent Spaces (NeurIPS 2020)

Diverse Image Captioning with Context-Object Split Latent Spaces This repository is the PyTorch implementation of the paper: Diverse Image Captioning

Visual Inference Lab @TU Darmstadt 34 Nov 21, 2022
PyTorch implementation of the Pose Residual Network (PRN)

Pose Residual Network This repository contains a PyTorch implementation of the Pose Residual Network (PRN) presented in our ECCV 2018 paper: Muhammed

Salih Karagoz 289 Nov 28, 2022
This is the official released code for our paper, The Emergence of Objectness: Learning Zero-Shot Segmentation from Videos

The-Emergence-of-Objectness This is the official released code for our paper, The Emergence of Objectness: Learning Zero-Shot Segmentation from Videos

44 Oct 08, 2022
An educational resource to help anyone learn deep reinforcement learning.

Status: Maintenance (expect bug fixes and minor updates) Welcome to Spinning Up in Deep RL! This is an educational resource produced by OpenAI that ma

OpenAI 7.6k Jan 09, 2023
Implementation of "StrengthNet: Deep Learning-based Emotion Strength Assessment for Emotional Speech Synthesis"

StrengthNet Implementation of "StrengthNet: Deep Learning-based Emotion Strength Assessment for Emotional Speech Synthesis" https://arxiv.org/abs/2110

RuiLiu 65 Dec 20, 2022
Gluon CV Toolkit

Gluon CV Toolkit | Installation | Documentation | Tutorials | GluonCV provides implementations of the state-of-the-art (SOTA) deep learning models in

Distributed (Deep) Machine Learning Community 5.4k Jan 06, 2023
Deep Learning GPU Training System

DIGITS DIGITS (the Deep Learning GPU Training System) is a webapp for training deep learning models. The currently supported frameworks are: Caffe, To

NVIDIA Corporation 4.1k Jan 03, 2023
Code to use Augmented Shapiro Wilks Stopping, as well as code for the paper "Statistically Signifigant Stopping of Neural Network Training"

This codebase is being actively maintained, please create and issue if you have issues using it Basics All data files are included under losses and ea

J K Terry 32 Nov 09, 2021
Equivariant GNN for the prediction of atomic multipoles up to quadrupoles.

Equivariant Graph Neural Network for Atomic Multipoles Description Repository for the Model used in the publication 'Learning Atomic Multipoles: Predi

16 Nov 22, 2022
PyTorch implementation of our ICCV2021 paper: StructDepth: Leveraging the structural regularities for self-supervised indoor depth estimation

StructDepth PyTorch implementation of our ICCV2021 paper: StructDepth: Leveraging the structural regularities for self-supervised indoor depth estimat

SJTU-ViSYS 112 Nov 28, 2022
Show Me the Whole World: Towards Entire Item Space Exploration for Interactive Personalized Recommendations

HierarchicyBandit Introduction This is the implementation of WSDM 2022 paper : Show Me the Whole World: Towards Entire Item Space Exploration for Inte

yu song 5 Sep 09, 2022