Last updated: 2018-06-01

workflowr checks: (Click a bullet for more information)
Expand here to see past versions:


Overview

This repository contains code resources to accompany our research paper:

Sarah M. Urbut, Gao Wang and Matthew Stephens (2017). Flexible statistical methods for estimating and testing effects in genomic studies with multiple conditions. bioRxiv doi:10.1101/096552.

We provide three sets of resources:

  1. If you are primarily interested in applying multivariate adaptive shrinkage (mash) methods to your own data, please see the mashr repository.

  2. If you are working with association statistics from the GTEx study, or a similar data set, please see here for our code to convert association statistics in the FastQTL format, or a similar format, to a format that is more convenient for analysis with mashr.

  3. If you would like follow the exact steps we took to produce the results presented in the manuscript, follow the instructions here.

Citing this work

If you find any of the source code in this repository useful for your work, please cite our paper:

Sarah M. Urbut, Gao Wang and Matthew Stephens (2017). Flexible statistical methods for estimating and testing effects in genomic studies with multiple conditions. bioRxiv doi:10.1101/096552.

License

Copyright (c) 2016-2018, Sarah Urbut, Gao Wang, Peter Carbonetto and Matthew Stephens.

All source code and software in this repository are made available under the terms of the MIT license. See the LICENSE file for the full text of the license.

Contributors

This project was developed by Sarah Urbut, Gao Wang, Peter Carbonetto. Matthew Stephens at the University of Chicago, with contributions from

We provide three sets of code resources to accompany Urbut, Wang, Carbonetto & Stephens (2017)

This site contains source code to produce results from Urbut, Wang & Stephens (2017).

If you are primarily interested in applying the statistical methods to your own data, please see the mashr package.

To reproduce the results of Urbut, Wang & Stephens (2017), please follow the instructions here.

Overview

To reproduce the results of Urbut, Wang & Stephens (2017), please follow these instructions.

For more information, please see the README.

The complete analyses require installation of several programs and libraries, and requires several large data sets. To facilitate reproducing our results, we provide pre-processed data for use with the core analysis, and a bioinformatics pipeline with a small toy data-set to demonstrate the pre-processing step. We have also developed a Docker container that includes all software components necessary to run the analyses. Docker can run on most popular operating systems (Mac, Windows and Linux). It also runs on cloud computing services such as Amazon Web Services and Microsoft Azure. If you have not used Docker before, you might want to read this to learn the basic concepts and understand the benefits of Docker.

For details on how the Docker image was configured, see the Dockerfile. The Docker image used for our analyses is based on gaow/lab-base, a Docker image for development with R and Python.

If you prefer to run the analyses without Docker, add a few details about where you can find out more about software and libraries used, and other computing environment setup steps (mention Python 3.x, R, SFA, ExtremeDeconvolution, MOSEK, OpenMP, MKL, GSL, HDF5 tools, pytables rhdf5, and for an improved MASH implementation mashr is also installed).

1. Download and install Docker

Download Docker (note that a free community edition of Docker is available), and install it following the instructions provided on the Docker website. Once you have installed Docker, check that Docker is working correctly by following Part 1 of the Getting Started guide. If you are using Docker for the first time, we recommend reading the entire Getting Started guide. Note that setting up Docker requires that you have administrator access to your computer. (Singularity is an alternative that accepts Docker images and does not require administrator access.)

2. Download and test Docker image

Run this alias command in the shell, which will be used below to run commands inside the Docker container:

alias mash-docker='docker run --security-opt label:disable -t -P -h MASH '\
'-w $PWD -v $HOME:/home/$USER -v /tmp:/tmp -v $PWD:$PWD '\
'-u $UID:${GROUPS[0]} -e HOME=/home/$USER -e USER=$USER gaow/mash-paper'

The -v flags in this command map directories between the standard computing environment and the Docker container. Since the analyses below will write files to these directories, it is important to ensure that:

If any of these statements are not true, please adjust the alias accordingly. The remaining options only affect operation of the container, and so should function the same regardless of your operating system.

Next, run a simple command in the Docker container to check that has loaded successfully:

mash-docker uname -sn

This command will download the Docker image if it has not already been downloaded.

If the container was successfully run, you should see this information about the Docker container outputted to the screen:

Linux MASH

You can also run these commands to show the information about the image downloaded to your computer and the container that has run (and exited):

docker image list
docker container list --all

Note: If you get error “Cannot connect to the Docker daemon. Is the docker daemon running on this host?” in Linux or macOS, see here for Linux or here for Mac for suggestions on how to resolve this issue.

3. Clone or download this repository

Clone or download the gtexresults repository to your computer, then change your working directory in the shell to the root of the repository, e.g.,

cd gtexresults

After doing this, running ls -1 should show the top-level contents of this repository:

LICENSE
README.md
TODO.txt
analysis
data
docs
output
workflows

All commands below will be run from this directory.

4. Fit MASH model and compute posterior statistics

Assuming your working directory is the root of the git repository (you can check by running pwd), run all the steps of the analysis with this command:

mash-docker sos run workflows/gtex6_mash_analysis.ipynb

This command will take several hours to run—see below for more information on the individual steps. All outputs generated by this command will be saved to folder output inside the repository.

Note that you may recognize this file as a Jupyter notebook. Indeed, you may open this notebook in Jupyter. However, you should not step through the code sequentially as you would in a typical Jupyter notebook; this is because the code in this notebook is meant to be run using the Script of Scripts (SoS) framework.

This command will execute the following steps of the analysis:

  • Compute a sparse factorization of the (centered) z-scores using the SFA software, with K = 5 factors, and save the factors in an .rds file. This will be used to construct the mixture-of-multivariate normals prior. This step is labeled sfa, and should only take a few minutes to run.

  • Compute additional “data-driven” prior matrices by computing a singular value decomposition of the (centered) z-scores and low-rank approximations to the empirical covariance matrices. Most of the work in this step involves running the Extreme Deconvolution method. The outcome of running the Extreme Deconvolution method is saved to a new .rds file. This step is labeled mash-paper_1 and may take several hours to run (in one run on a MacBook Pro with a 3.5 GHz Intel Core i7, it took over 6 hours to complete).

  • A final collection of “canonical” and single-rank prior matrices based on SFA and the “BMAlite” models of Flutre et al (2013). These matrices are again written to another .rds file. This step is labeled mash-paper_2, and should take at most a minute to run.

  • The mash-paper_3 step fits the MASH (“multivariate adaptive shrinkage”) model to the GTEx data (the centered z-scores); the model parameters estimated in this fitting step are the weights of the multivariate normal mixture. The outputs from this step are the estimated mixture weights and the conditional likelihood matrix. These two outputs are saved to two separate .rds files. This step is expected to take at most a few hours to complete.

  • The mash-paper_4 step computes posterior statistics using the fitted MASH model from the previous step. These posterior statistics are summarized and visualized in subsequent analyses. The posterior statistics are saved to another .rds file. This step is expected to take a few hours to complete.

Finally, note that all containers that have run and exited will still be retained in the Docker system. Run docker container list --all to list all previous run containers. To clear these previously run containers, run docker container prune. See here for more information.

5. Add Step 5 title here

Install some packages from CRAN:

# Add commands here to install packages.

For convenience, the results needed to generate the figures and tables have been saved in the output folder.

FIXME: update figure plotting instructions

The input data necessary to run this analysis is all available under inputs. This may take some time to run. We have provided the outputs of running mash in Data_vhat.

This repo is organized so that you can run Mash using the gteX data contained in Inputs to produce the parameters and posteriors from mashr.

The directory Plots_for_Paper_vmat contains .Rmd files to plot figures from the paper, using our results which are provided in Results_Data.

Figure 3:Summary of primary patterns identified by mash in GTEx data

Figure 4:Examples illustrating of how mash uses patterns of sharing to inform effect estimates in the GTEx data.

Figure 5:Histogram of Sharing

Figure 6:Pairwise sharing by magnitude of eQTL among tissues

Supplementary Figure 1:Sample sizes and effective sample sizes from mash analysis across tissues

Supplementary Figure 2:There are 4 figures here:

Summary of covariance matrices Uk with largest estimated weight (> 1%) in GTEx data:Uk2

Summary of covariance matrices Uk with largest estimated weight (> 1%) in GTEx data:Uk4

Summary of covariance matrices Uk with largest estimated weight (> 1%) in GTEx data:Uk5

Summary of covariance matrices Uk with largest estimated weight (> 1%) in GTEx data:Uk8

Supplementary Figure 3: Illustration of how Linkage Disequilibrium can impact effect estimate table and figure

Supplementary Figure 4:Pairwise Sharing By Sign

Supplementary Figure 5:Number of “tissue-specific eQTLs” in each tissues.

Supplementary Figure 6:Expression levels in genes with “tissue-specific eQTLs” are similar to those in other genes

Table 1: Heterogeneity Analysis Simulation and Data.

More detailed usage notes

Above we have given the minimal instructions necessary to reproduce the results of Urbut et al (2017). Here are some additional details about the analyses.

TO DO: Things that will go here:

Converting eQTL summary statistics to MASH format

Here we explain how the MatrixEQTLSumStats.Portable.Z.rds data file was generated from the source data downloaded from the GTEx Portal. The source data are the SNP-gene association statistics from release 6 of the GTEx Project (GTEx_Analysis_V6_all-snp-gene-associations.tar).

Under the repo you will find workflows/fastqtl_to_mash.ipynb to convert eQTL summary statistics (default to fastqtl format) to MASH format. Computation is configured to run in parallel for eQTL results from multiple studies. Example data-set can be found at data/eQTLDataDemo. The workflow file is documented in itself, and has a few options to customize the input and output.

To read what’s available, run:

mash-docker sos run workflows/fastqtl_to_mash.ipynb export

and read the HTML files output/fastqtl_to_mash.lite.html and output/fastqtl_to_mash.full.html

To run the conversion:

mash-docker sos run workflows/fastqtl_to_mash.ipynb \
  --data_list data/eQTLDataDemo/FastQTLSumStats.list \
  --gene_list data/eQTLDataDemo/GTEx_genes.txt

In practice for GTEx data the conversion is computationally intensive and is best done on a cluster environment with configurations to run the workflow across different nodes.

Developer notes

Run the following command to update the Docker image:

docker pull gaow/mash-paper

This reproducible R Markdown analysis was created with workflowr 1.0.1.9000