Skip to content

Materials (slides, notebooks, Dockerfile) for a one day workshop on the use of Nipype and Arcana for neuroimaging analysis (https://nbviewer.jupyter.org/github/MonashBI/nipype_arcana_workshop/blob/master/program.ipynb)

License

Notifications You must be signed in to change notification settings

vaibhavi5/nipype_arcana_workshop

 
 

Nipype, Arcana and Banana CVL Workshop - Melbourne 2019

GitHub issues GitHub pull-requests GitHub contributors GitHub Commits GitHub size GitHub HitCount Docker Hub

This repository contains everything for the Automating Neuroimaging Analysis Workflows with Nipype, Arcana and Banana workshop to be held at Swinburne University on the 15 of November 2019 (registration). Many of the materials for this course have been adapted from Michael Notter's excellent Nipype tutorial and a series of workshops he and Peer Herholz have run on "Python in neuroimaging" (Cambridge 2018 and Marburg 2019).

If you get access to the Characterisation Virtual Laboratory (CVL) before the workshop (recommended option) then all the software tools requirements are already installed and ready for you to use. To create an account on CVL@ MASSIVE you need your institutional email. Follow the instructions here: https://www.cvl.org.au/cvl-desktop/cvl-accounts and join project fs97. The CVL aims at giving you (and researchers in Australia) resources such as compute nodes, RAM and GPU to do computational analysis through a virtual desktop. It has all the software required installed and runs using the high performance compute power of MASSIVE. If you have specific questions or issues with the creation of your account, feel free to email [email protected] or check the useful documentation https://docs.massive.org.au/.

Once you have an account please follow the instructions under "Configuring your CVL Account" on the program to get up and running.

If you would like to run the materials from your own system there are (at least) three ways that you can configure it, make sure you do this before the day of the workshop:

1. Docker

Probably the easiest way to go through the workshop material is to use the docker image monashbi/nipype_arcana_workshop. It provides the computational environment to run the notebooks on any system with all necessary dependencies installed. To install Docker on your system, follow one of those links:

Once Docker is installed, open up the docker terminal and test if it works with the command: docker run hello-world

Note: Linux users might need to use sudo to run docker commands or follow the post-installation steps.

Once docker is running on your system you can continue with downloading the docker image for this workshop. For this use the command:

docker pull monashbi/nipype_arcana_workshop

Once the download finished, proceed with the following steps:

  1. Run the following command in a terminal: docker run -it --rm -p 8888:8888 monashbi/nipype_arcana_workshop
  2. Copy paste the link that looks like http://20f109eba8e4:8888/?token=0312c1ef3b61d7a44ff5346d3d150c23249a548850e13868 into your webbrowser.
  3. Replace the hash number 20f109eba8e4 after http:// with localhost or your local IP (probably 192.168.99.100) if you're on windows.
  4. Once Jupyter Notebook is open, click on the program.ipynb notebook, and you're good to go.

And if you want to have access to the output data created within the docker container, add the command -v /path/to/your/output_folder:/output before monashbi/nipype_arcana_workshop, where /path/to/your/output_folder should be a folder on your system, such as /User/your_account/Documents/projectBanana/output.

Some useful Docker Commands

Show running containers
$ docker ps

Show all installed Docker images
$ docker images

Show all (also stopped) containers
$ docker ps -a

Remove a container
$ docker rm $CONTAINER_ID

Remove a docker image
$ docker rmi -f $IMAGE_ID

Start a stopped container & attach to it
$ docker start -ia $CONTAINER_ID

Note: when you stop a container (Ctrl-C), and then re-execute above "docker run" command, you will end up with a second container. If you want to access your previous container (e.g. with downloaded data), you must reconnect to it (see "docker start -ia" command below).

2. Native installation in virtual environment

If you already have the following software dependencies installed on your workstation (or are only doing the non-neuroimaging notebooks):

  • convert3d >=1.0
  • ants >=2.3
  • fsl >=6.0
  • graphviz >=2.30
  • matlab >=r2018a

you can use pip or conda to create the necessary python environment to run the notebooks. While it is possible to install the dependencies globally on your system it is highly recommended to virtualise your installation so it doesn't mess up any other programs that you use.

Option 1: Python 3 + Pip + Venv

Pip and Venv are built in to Python 3 and are pretty easy to use. However, sometimes a Python package might depend on a non-Python library that needs to be installed separately. To install these non-Python dependencies it is strongly recommended to use a "package manager" such as Homebrew (macOS), Chocolately (Windows) or apt/yum (Linux).

  1. Install Python 3 (NB: version 2 won't work) on your system using a package manager
  2. Create a virtual environment somewhere on your system python3 -m venv /path/to/your/venv (replace '/' with '\' on Windows)
  3. Activate the virtual environment source /path/to/your/virtual-env/bin/activate (\path\to\virtual-env\Scripts\activate on Windows)
  4. Install the requirements in the virtualenv pip3 install -r https://raw.githubusercontent.com/MonashBI/nipype_arcana_workshop/master/requirements.txt.

NB: To deactivate the virtualenv afterwards run deactivate, and to reactivate it again just repeat Step 3.

Option 2: Conda

Conda integrates package management for Python and non-Python dependencies into a single cross-platform system. It is therefore perhaps the simplest solution for native installation. However, if you install a lot of (obscure) packages you may find one that isn't supported by Conda and then it gets a bit tricky. This is why I personally prefer the Pip + Venv approach but for 99% of use cases they are interchangeable.

  1. If you haven't yet, get conda on your system: https://conda.io/miniconda.html
  2. Download the environment.yml file from [here].(https://github.com/MonashBI/nipype_arcana_workshop/blob/master/environment.yml)
  3. Open up a conda terminal (or any other terminal), and create a new conda environment with the following command: conda env create --name workshop --file /path/to/file/environment.yml
  4. Activate the conda environment with source activate workshop (for mac and linux) or with activate workshop (for windows)

Opening up the notebook

Once you have your virtual environment installed (either by venv or conda) you should clone this repository, download the test data and start Jupyter notebook with the following commands:

  1. Using Git, clone this repository somewhere sensible on your workstation i.e. (~/git/arcana-workshop).
  2. Download the "A test-retest fMRI dataset for motor, language and spatial attention functions" dataset by Gorgolewski et al. + derivatives using Datalad in the 'notebooks/data' sub-directory of the cloned repository.
    cd notebooks/data
    datalad install -r ///workshops/nih-2017/ds000114
    cd ds000114
    datalad update -r
    datalad get -r .
  3. Symlink the $FSLDIR/data/atlases directory into the notebooks/data sub-directory
  4. Create the notebooks/data/output directory
  5. Go to the root of the cloned repo and run the following command: jupyter notebook program.ipynb. This should open a new window in your browser with the address 'http://localhost:8888'.

3. Jupyter NBViewer

All the notebooks can be looked at via Jupyter nbviewer. Like this you can see everything but cannot really interact with the scripts or run the code.

About

Materials (slides, notebooks, Dockerfile) for a one day workshop on the use of Nipype and Arcana for neuroimaging analysis (https://nbviewer.jupyter.org/github/MonashBI/nipype_arcana_workshop/blob/master/program.ipynb)

Resources

License

Code of conduct

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 99.4%
  • HTML 0.4%
  • Python 0.1%
  • CSS 0.1%
  • Dockerfile 0.0%
  • Shell 0.0%