Genome project to require Google-like computing power

An ambitious project with the goal of producing a more detailed understanding of the link between genetic variations and susceptibility to disease will require an unprecedented amount of computing power and terabytes of data storage, according to the leaders of the project.

The 1,000 Genomes Project, announced earlier this week by an international consortium that includes the National Human Genome Research Institute, part of the National Institutes of Health, plans to examine over a three-year period the human genome at a level of detail never before accomplished.

The project "will greatly expand and further accelerate efforts to find more of the genetic factors involved in human health and disease," said Richard Durbin, deputy director of the Wellcome Trust Sanger Institute in Cambridge, England.

Francis Collins, director of the research institute, said the project will lead to a fivefold increase in the sensitivity of disease discovery efforts across the human genome.

Any two humans are more than 99 percent similar at the genetic level, but the fractional differences can help determine susceptibility to disease and how the body will respond to drugs. The goal of the project is to produce a catalog of variants that are present at 1 percent or greater frequency in the human population across most of the genome. That requires the project to sequence the genes of at least 1,000 people.

The project plans to sequence 8.2 billion DNA base pairs a day -- or the equivalent of more than two human genomes every 24 hours -- during its two-year production phase, for a total of 6 trillion DNA bases, said Gil McVean, co-chair of the analysis committee and professor of mathematical genetics at the University of Oxford.

Managing this massive amount of data will require novel computational methods. Gonçalo Abecasis, a professor of applied statistics and a geneticist who works at the Center for Statistical Genetics at the University of Michigan, said the data produced by the genome project will be so immense that the only process that he can think of that is similar in scope is the search engine Google, which manages billions of Web searches daily.

If the project had to start crunching all the sequence data today, Abecasis estimated it would take a supercomputer with 10,000 massively parallel processors. But, he said, the project is working to develop algorithms and mathematical and computational models that should reduce the computing requirements.

Because the genomes of most people are mostly similar, Abecasis said he is working on models and algorithms designed to process and crunch the fractional differences, much like the way video compression algorithms function when processing power is applied to objects that move and not to static background objects.

The models still are being developed, but the project will require supercomputers to manipulate the data but need far fewer than 10,000 processors, Abecasis said.

The Beijing Genomics Institute in Shenzhen, China, is the other key research organization participating in the project and will perform sequencing along with the Wellcome Trust Sanger Institute and its large-scale sequencing network. That network includes the Broad Institute of MIT and Harvard, the Washington University Genome Sequencing Center at the Washington University School of Medicine in St. Louis, and the Human Genome Sequencing Center at the Baylor College of Medicine in Houston.

Stay up-to-date with federal news alerts and analysis — Sign up for GovExec's email newsletters.
FROM OUR SPONSORS
JOIN THE DISCUSSION
Close [ x ] More from GovExec
 
 

Thank you for subscribing to newsletters from GovExec.com.
We think these reports might interest you:

  • Forecasting Cloud's Future

    Conversations with Federal, State, and Local Technology Leaders on Cloud-Driven Digital Transformation

    Download
  • The Big Data Campaign Trail

    With everyone so focused on security following recent breaches at federal, state and local government and education institutions, there has been little emphasis on the need for better operations. This report breaks down some of the biggest operational challenges in IT management and provides insight into how agencies and leaders can successfully solve some of the biggest lingering government IT issues.

    Download
  • Communicating Innovation in Federal Government

    Federal Government spending on ‘obsolete technology’ continues to increase. Supporting the twin pillars of improved digital service delivery for citizens on the one hand, and the increasingly optimized and flexible working practices for federal employees on the other, are neither easy nor inexpensive tasks. This whitepaper explores how federal agencies can leverage the value of existing agency technology assets while offering IT leaders the ability to implement the kind of employee productivity, citizen service improvements and security demanded by federal oversight.

    Download
  • IT Transformation Trends: Flash Storage as a Strategic IT Asset

    MIT Technology Review: Flash Storage As a Strategic IT Asset For the first time in decades, IT leaders now consider all-flash storage as a strategic IT asset. IT has become a new operating model that enables self-service with high performance, density and resiliency. It also offers the self-service agility of the public cloud combined with the security, performance, and cost-effectiveness of a private cloud. Download this MIT Technology Review paper to learn more about how all-flash storage is transforming the data center.

    Download
  • Ongoing Efforts in Veterans Health Care Modernization

    This report discusses the current state of veterans health care

    Download

When you download a report, your information may be shared with the underwriters of that document.