This bioinformatics workflow will take you through a variant discovery pipeline, from raw reads to actionable insights, all orchestrated with Flyte. The workflow is broken up into sensible steps, each highlighting some of the more advanced Flyte features.
- Clone the repo
- Create a venv, activate it and
pip install -r requirements.txt
- Spin up the demo cluster locally by following these instructions
- Navigate to the minio instance that was started alongside the demo cluster
- log in with
minio
andminiostorage
as username and password - create the folders
my-data/sequences
andmy-data/refs
- upload the sequences and refs from
src/tests/assets
to those locations
- log in with
- Run the workflow with
pyflyte register src/workflows && python src/execute.py
- Navigate to the previous command's provided URL to watch your workflow execute in the Flyte console!
The first section will collect QC metrics, preprocess our samples, and align them to a reference genome using a couple different aligners for comparison sake.
✨Adding custom dependencies alongside Flytekit
You can build out the default Dockerfile that ships with pyflyte init
to add custom dependencies. The Dockerfile here incorporates a number of common bioinformatics tools to create a base image for the project. Any additional python dependencies can also be specified in the requirements.txt file, which gets pulled into and installed in the final image. You'll also be able to use fast registration as you iterate on your code to avoid having to rebuild and push an image every time.
✨Leverage dataclasses to keep things organized
Using dataclasses to define your samples provides a clean and extensible data structure to keep your workflows tidy. Instead of writing to directories and keeping track of things manually on the commandline, these dataclasses will keep relevant metadata about your samples and let you know where to find them in object storage.
✨Run arbitrary shell commands
FastQC is a very common tool for gathering QC metrics about raw reads. It's a java tool and included in our Dockerfile. It doesn't have any python bindings, but luckily Flyte lets us run arbitrary ShellTasks with a clean way of passing in inputs and receiving outputs. Just define a script for what you need to do and ShellTask will handle the rest.
✨Decide wether to continue workflow execution based on QC metrics via conditionals
FastQC generates a summary file with a simple PASS / WARN / FAIL call across a number of different metrics. We can use conditionals in our workflow to check for any FAIL lines in the summary and automatically halt execution. This can surface an early failure without wasting valuable compute or anyone's time doing manual review.
✨Specify resources and parallelize via map task
FastP is another common pre-processing tool for filtering out bad reads, trimming, and adapter removal. It can be a bit more memory hungry than Flyte's defaults are set to; luckily we can use Resources to bump that up and allow it to run efficiently. Additionally, we can make use of a map task in our workflow to parallelize the processing of fastp across all our samples.
✨Pause processing while waiting for human input
As a final check before moving onto the alignment, we can define an explicit approval right in the workflow. Aggregating reports of all processing done up to this point, and visualizing it via Decks (more on that later), a researcher is able to quickly get a high level view of the work done so far and approve the analysis for further processing.
✨Leverage caching to save time on successive runs
Index generation can be a very compute intensive step. Luckily, we can take advantage of Flyte's native caching when building that index for bowtie and hisat. We've also defined a cache_version
in the config that relies on a hash of the reference location in the object store. This means that changing the reference will invalidate the cache and trigger a rebuild, while allowing you to go back to your old reference with impunity.
✨Compare aligners across an arbitrary number of inputs via dynamic workflows
When prototyping a new pipeline, it's usually a good idea to evaluate a few different tools to see how they perform with respect to runtime and resource requirements. This is easy with a dynamic
workflow, which allows us to pass in an arbitrary number of inputs to be used with whatever tasks we want. In the main workflow you'll pass a list of filtered samples to each tool and be able to capture run statistics in the SamFile dataclass as well as visualize their runtimes in the Flyte console.
✨Add a dependency in-line and visualize performance via Decks
Finally, it's easy to tack on a dependency that's not included in your Dockerfile via an ImageSpec definition inline. We do this for MultiQC, an excellent multi-modal visualization tool. After gathering all relative metrics from the workflow, we're able to render that report via Decks, giving us rich run statistics without ever leaving the Flyte console!