Kylee Degatano, Author at Terra https://terra.bio/author/kdegatano/ Science at Scale Wed, 27 Dec 2023 04:55:46 +0000 en-US hourly 1 https://wordpress.org/?v=6.9.1 https://terra.bio/wp-content/uploads/2023/12/Terra-Color-logo-300-150x150.pngKylee Degatano, Author at Terrahttps://terra.bio/author/kdegatano/ 32 32 Scaling variant discovery to a million genomes with the Genomic Variant Storehttps://terra.bio/scaling-variant-discovery-to-a-million-genomes-with-the-genomic-variant-store/ https://terra.bio/scaling-variant-discovery-to-a-million-genomes-with-the-genomic-variant-store/#respond Thu, 27 Oct 2022 18:40:48 +0000 https://terrabioappdev.wpenginepowered.com/scaling-variant-discovery-to-a-million-genomes-with-the-genomic-variant-store/Join the early access program to try out the Broad Institute's new Genomic Variant Store, a highly scalable variant storage and processing solution powered by Google BigQuery on Terra.

The post Scaling variant discovery to a million genomes with the Genomic Variant Store appeared first on Terra.

]]>
Kylee Degatano is a Senior Product Manager in the Data Sciences Platform at the Broad Institute. In this guest blog post, she introduces the Genomic Variant Store, a highly scalable solution for genomic analysis based on Google BigQuery, designed to scale joint variant discovery to a million whole genome samples. Researchers interested in trying out this approach are invited to join the GVS Early Access program


 

Researchers around the world are generating petabytes of genomics data to understand human biology,  identify the causes of diseases and develop new treatments. The analyses involved, such as association studies, linkage analysis, and exploration of Mendelian inheritance all require high confidence in the quality of the genomic variant calls they rely on. 

One of the most successful approaches of the past decade for generating such high quality variant calls involves jointly analyzing the genomes of many different people across a population. This “joint calling” approach increases statistical power for differentiating true variants from artifacts, which makes it possible to identify extremely rare variants with confidence. In the GATK Best Practices for germline short variant discovery, it is implemented as a two-step process: first we identify potential variants individually per sample, then we evaluate the evidence found for each genomic site across all samples to produce “joint calls”, i.e. a multi-sample variant callset. We can then apply additional filtering to further refine the callset for downstream analysis.   

 

Diagram illustrating the overall data generation and analysis process involved in joint variant discovery.

 

As an example of how this approach empowers discovery, my colleague Laura Gauthier recently wrote a blog post commenting on a recent study that used a joint callset of 75,000 exome samples to implicate ten new genes in the development of schizophrenia. Crucially, she explained, the study focused on ultra-rare variants, never seen before and only observed in a single individual in that entire callset, that affected the same small set of genes. The joint calling methodology was essential to the success of that analysis because it enabled the authors to accurately detect ultra-rare variants with confidence, and to do so for enough individuals to accumulate enough statistical power to implicate the corresponding genes. Laura concluded her post with this prediction

“Future breakthroughs in common disease research will continue to come about through the hard work of recruiting large numbers of affected participants; accurately detecting the few, tiny, rare mutations that make them different; and combining as many ultra-rare variants as we can until we can point the finger at genes harboring too many mutations across people who suffer from disease.”

 

It is with that same vision in mind that our engineering team has been working hard to scale up our joint calling capabilities.

Every jump in the scale of the studies we have supported so far has required that we re-engineer various parts of our analysis pipelines to address new barriers related to cost efficiency, computing power, runtime… or the maximum amount of data you can store in a single file on a standard hard drive. (That was a fun one to hit.)

Today, I’m excited to share the solution we developed to get to the next order of magnitude, and to make this kind of scaling accessible to a wider range of researchers and institutions.

 

Aiming for a million

A few years ago, the NIH asked our team to figure out a way to run joint calling on one million human genomes for the All of Us Research Program. For context, at the time, performing joint calling for “just” 15,000 whole genomes was a costly months-long endeavor for a full team of engineers equipped with cutting-edge tools. We knew that to scale to a million genomes, we would need to revisit the engineering design behind the GATK Joint Calling pipeline (again). We also foresaw that many downstream analysis tools would not be able to handle the resulting callset, which would contain trillions of variants. Researchers would need to be able to subset the variant calls to their samples and genomic regions of interest. 

Oh, and we needed to ensure our solution to these problems would be cost- and time-efficient, and scalable to both ends of the spectrum: it should allow us to scale down to 10 samples as easily as scaling up to a million.

So we took joint calling back to the drawing board. We started by designing a core schema for the variant data based on data access patterns for key use cases like training a filtering model, searching to identify samples with a specific variant or variation in a specific location, and extracting data —for all samples or for subsets— into Variant Call Format (VCF). We contemplated which information in the variant files was necessary for joint calling and downstream analysis, and stripped out anything extraneous. We tested many different foundational technologies for processing and storing data, including Spark, Dataflow, and custom developed infrastructure. We ultimately chose Google BigQuery because it is easy to operate, can handle huge data sizes, is cheap to store data and cheap to query, and has excellent security features. 

The result of this three-year development effort: a highly scalable and cost-effective variant storage and processing solution we call the Genomic Variant Store. 

 

Introducing Joint Calling with the Genomic Variant Store

The Genomic Variant Store (GVS) uses Google BigQuery to store variant data according to the core data model. It is designed to function as a persistent data warehouse to which we can add new data over time, simply by importing gVCF files produced by the single-sample calling step of the variant discovery process outlined above. This automatically combines the per-sample variant data across samples and genomic coordinates. We can then query the GVS by sample and/or by genomic coordinates to generate subset callsets of interest in either VCF or Hail VDS file format. 

If you are familiar with the details of the current GATK Best Practices workflow implementation, the GVS data ingestion step is analogous to combining gVCFs into a GenomicsDB data store. However GVS is much more scalable than GenomicsDB due to its use of Google BigQuery. 

The GVS also includes a built-in variant filtering model (GATK VQSR) that determines which variant calls will be considered true variants, as opposed to artifacts. The filtering model is applied when a subset of variant data is extracted to file, and it can be re-trained and improved as new data is added.

 

Overview of the operations supported by the Genomic Variant Store 

 

For convenience, we have developed utility workflows written in the Workflow Description Language (WDL) to perform the import, model training, subsetting and variant search operations. These can be used individually to grow, curate and analyze a persistent variant data store.

We have tested this on a wide range of callset sizes; there is no minimum number of samples, and we anticipate that it will scale to the planned one million genomes for the All of Us Research Program. In fact, we have already used the GVS in production to produce a joint callset from 250,000 human whole genomes for the AoU program, which we believe is the largest joint-called human whole genome callset in the world so far.  

 

Get early access to a self-contained GVS Joint Calling Pipeline

Due to the engineering challenges involved in operating at this scale, there are only a handful of genome centers around the world that are currently able to create joint callsets from tens of thousands of human whole genomes, let alone hundreds of thousands. Yet it is very likely that progress in human genetics could be substantially accelerated if this capability were made more widely accessible.

As we continue to improve the GVS — increasing scalability, adding support for new data types, improving the internal algorithms — we are looking for feedback from external groups to ensure the GVS will work well for a wide audience. To that end, we are starting an early access program for researchers who are interested in trying out the GVS approach for making joint callsets of up to 10,000 human whole genomes. 

We are currently targeting that project size because most population genetics projects today involve less extreme cohorts than the All of Us Research Program; we’re seeing many projects with whole genome sample numbers in the low thousands. 

To enable those projects to use GVS for scalable joint calling without having to manage complex infrastructure and run multiple separate operations, we developed a “one and done” workflow called the GVS Joint Calling pipeline that wraps all the necessary steps. This single self-contained workflow takes in a set of per-sample gVCFs, trains and applies the GATK VQSR filtering model, and extracts the variants into a VCF file containing the complete joint callset. 

We plan to make this pipeline publicly available in a Terra workspace, pre-configured in such a way that anyone can run it out of the box with minimal effort. In our current tests, the GVS Joint Calling pipeline set up in Terra can produce a joint callset from up to 10,000 human whole genomes in less than half a day, at a cost of $0.06 USD per genome.

 

Screenshot of the Terra workflow configuration panel for the GVS Joint Calling Pipeline

 

We invite you to apply to join the early access program by filling out a short form that will help us assess if your callset would be a good fit for the initial release. If you are selected, we will help you get started with the GVS Joint Calling pipeline in Terra, and we will be available to assist you if you run into any problems.   

Even if the Genomic Variant Store doesn’t meet your needs right now, please feel free to use the form to tell us more about your work and what features you be interested in seeing in a future version. We look forward to hearing from you! 

 

 

The post Scaling variant discovery to a million genomes with the Genomic Variant Store appeared first on Terra.

]]>
https://terra.bio/scaling-variant-discovery-to-a-million-genomes-with-the-genomic-variant-store/feed/ 0
Introducing WARP: A collection of cloud-optimized workflows for biological data processing and analysishttps://terra.bio/introducing-warp-a-collection-of-cloud-optimized-workflows-for-biological-data-processing-and-analysis/ https://terra.bio/introducing-warp-a-collection-of-cloud-optimized-workflows-for-biological-data-processing-and-analysis/#respond Thu, 22 Oct 2020 14:21:37 +0000 https://terrabioappdev.wpenginepowered.com/introducing-warp-a-collection-of-cloud-optimized-workflows-for-biological-data-processing-and-analysis/Guest blog post by Kylee Degatano, Product Manager for the Lantern Pipelines team in the Data Sciences Platform at the Broad Institute. I'm very excited to announce the recent release of WDL Analysis Research Pipelines (WARP), a brand new, public GitHub repository of cloud-optimized WDL workflows that are used in production at the [...]

The post Introducing WARP: A collection of cloud-optimized workflows for biological data processing and analysis appeared first on Terra.

]]>
Guest blog post by Kylee Degatano, Product Manager for the Lantern Pipelines team in the Data Sciences Platform at the Broad Institute.

I’m very excited to announce the recent release of WDL Analysis Research Pipelines (WARP), a brand new, public GitHub repository of cloud-optimized WDL workflows that are used in production at the Broad Institute. But before I get into specifics, I want to give you some context that explains why this is significant if you’re not already familiar with our organization.

The Broad Institute of MIT and Harvard — or just “the Broad” for short — is a non-profit biomedical research institution that also happens to be one of the largest sequencing centers in the world. Our Sequencing Platform pumps out hundreds of genomes and exomes every day, amounting to over 45 Tb/day, and the Institute overall is currently managing about 35 Pb of data stored in the cloud.

The teams that develop and operate Terra at the Broad (in collaboration with Verily, an Alphabet company) are part of the Data Sciences Platform (DSP), a sister organization to the Sequencing Platform. The DSP is responsible for processing all the sequencing data generated at the Broad, as well as data produced by partner institutions in the context of large collaborative efforts such as gnomAD, All of Us, the Human Cell Atlas, and the BRAIN Initiative. This includes genomic data of course but also single-cell transcriptomics and epigenomics.

To that end, we have a team of data engineers and bioinformaticians called “Lantern Pipelines” dedicated to building, optimizing, and maintaining workflows, using open-source tools that are either developed in-house (like GATK) or are produced by others in the research community (like STAR). Given the amount of data that needs to be processed, the workflows need to be very robust — we can’t afford to have our production operations screech to a halt anytime a sample comes through with some quirky formatting — and they need to be economical, to make every research dollar count.

The Lantern Pipelines team writes all their workflows in WDL, the Workflow Description Language maintained by the OpenWDL community, and optimizes them to run as quickly and efficiently as possible on Google Cloud using the Cromwell workflow management system — which is the same infrastructure that you have access to in Terra.

Note: Internally we use the term “pipeline” somewhat interchangeably with the term “workflow”, with the nuance that some pipelines are composed of several individual workflow scripts (typically one main workflow importing nested sub-workflows). Here, I use the term workflow throughout for consistency with the terminology used in Terra and its documentation.

As you can imagine, a huge amount of effort goes into developing and testing these workflows, so we’ve been wanting to share them with the world. We’d love to see others benefit from our team’s hard work, especially if that means they avoid having to reinvent wheels that we’ve already built!

That’s why I’m so thrilled that we’ve finally arrived at a good solution for sharing these workflows, which until now had been mostly siloed in private development repositories. We’ve set up a public GitHub repository called WARP, for WDL Analysis Research Pipelines, that collects all of our “blessed” workflows, meaning those that have been fully vetted for use at scale in our production operations. We’ve also registered each of them in Dockstore so they can be easily imported into Terra. In fact, many of these workflows are already available in public Terra workspaces; you can find those by searching for the “warp-pipelines” tag in Terra. We encourage you to clone those workspaces and try out the workflows on the included test data to familiarize yourself with their operation before running them on your own data.

If you have not run workflows in Terra before, check out the Workflows Quickstart video on Youtube for a step-by-step run-through of the process.

We are deeply committed to reproducibility, provenance, and transparency, so every workflow in WARP is released with a semantic version number to provide clear provenance of your data processing, as well as release notes that outline any updates as applicable. They are all open-source under a BSD 3-Clause license, and they call only open-source tools; though keep in mind that the exact licenses of the tools involved may be different, and you are responsible for checking that your intended use is allowed by each tool’s license. In addition, all tools are packaged in public docker containers distributed either by us or by their authors.

Finally, the repo’s documentation portal, which we are currently fleshing out, will include a full list of available workflows and documentation for each of them, so be sure to check it out and see if any of these workflows might provide a solution to some of your analysis needs. Remember also to check the Terra workspaces list in case the one you like is already available in a preconfigured workspace.

On behalf of the entire Lantern Pipelines team, I hope you will find this new resource useful and I look forward to hearing your thoughts about how we could further improve it.

WARP speed ahead!

The post Introducing WARP: A collection of cloud-optimized workflows for biological data processing and analysis appeared first on Terra.

]]>
https://terra.bio/introducing-warp-a-collection-of-cloud-optimized-workflows-for-biological-data-processing-and-analysis/feed/ 0