Skip to content

Pipeline for analysis of RRBS data, currently being based of MSSM pipeline

Notifications You must be signed in to change notification settings

MoTrPAC/motrpac-rrbs-pipeline

Repository files navigation

MoTrPAC RRBS Pipeline

DOI

Overview

This repo contains the Reduced Representation Bisulfite Sequencing(RRBS) data processing pipeline implemented in Workflow Description Language (WDL) based on harmonized MOP. This pipeline uses caper, a wrapper python package for the workflow management system Cromwell. All the data was processed on the Google Cloud Platform (GCP). The pipeline uses Bismark for alignment and quantification of methylation levels. The pipeline also generates a qc metrics file, useful for outlier detection and covariate adjustment during differential methylation analysis.

Details

GCP set-up

The WDL/Cromwell framework is optimized to run pipelines in high-performance computing environments. The MoTrPAC Bioinformatics Center runs pipelines on Google Cloud Platform (GCP). We used a number of wrapper tools developed by our colleagues from the ENCODE project to run pipelines on GCP (and other HPC platforms).

A brief summary of the steps to set-up a VM to run the Motrpac pipelines on GCP (for details, please, check the caper repo):

  • Create a GCP account.
  • Enable cloud APIs.
  • Install the Google Cloud SDK (Software Development Kit) on your local machine.
  • Create a service account and download the key file to your local computer (e.g “service-account-191919.json”)
  • Create a bucket for pipeline inputs and outputs (e.g. gs://pipelines/). Note: a GCP bucket is similar to a folder on your computer or a storage unit, but it is stored on Google's servers in the cloud instead of on your local computer.
  • Set up a VM on GCP: create a Virtual Machine (VM) instance from where the pipelines will be run. We recommend the script available in the caper repo. For that, clone the repo on your local machine and run the following command:
$ bash create_instance.sh [INSTANCE_NAME] [PROJECT_ID] [GCP_SERVICE_ACCOUNT_KEY_JSON_FILE] [GCP_OUT_DIR]

# Example for the pipeline:
./create_instance.sh pipeline-instance your-gcp-project-name service-account-191919.json gs://pipelines/results/
  • Finally, clone the repo

git clone https://github.com/MoTrPAC/motrpac-rrbs-pipeline

Software / Dockerfiles

Several tools are required to run the RRBS pipeline. All of them are pre-installed in docker container, which is publicly available in the Artifact Registry. To find out more about the specific versions of tools used to run the pipeline, check the Dockerfile

Configuration Files

An input configuration file (in JSON format) is required to process the data through the pipeline. This configuration file contains several key-value pairs that specify the inputs and outputs of the workflow, the location of the input files, default pipeline paramenters, docker containers, the execution environment, and other parameters needed for execution.

The optimal way to generate the configuration files is to run the make_json_rrbs.py script. Check this help guide to find out more.

Run the pipeline

Connect to the VM and submit the job using the below command

caper submit rrbs_pipeline_scatter.wdl -i input_json/set1_rrbs.json

Check the status of workflows and make sure they have succeeded by typing caper list on the VM instance that's running the job and look for Succeeded

About

Pipeline for analysis of RRBS data, currently being based of MSSM pipeline

Resources

Stars

Watchers

Forks

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •