New! Sign up for our free email newsletter.
Science News
from research organizations

Crowd-sourced competition to create 'big data' diagnostic tools

Date:
June 29, 2016
Source:
Rice University
Summary:
A crowdsourced collaboration/competition known as DREAM 9 set out three years ago to develop ideas for computational tools that would help treat patients with acute myeloid leukemia.
Share:
FULL STORY

Big data has a bright future in personalized medicine, as demonstrated by an international competition centered at Rice University that suggested ways forward for treatment of patients with leukemia.

In the DREAM 9 challenge, 31 teams of computational researchers applied competing methods to a unique set of patient data gathered from hundreds of patients with acute myeloid leukemia at the University of Texas MD Anderson Cancer Center.

Rice bioengineer Amina Qutub is principal investigator of the open-source paper published in PLOS Computational Biology. Rice served as the competition hub, in line with the university's strategic initiative to foster bioscience collaborations with fellow Texas Medical Center institutions.

DREAM, which stands for Dialogue for Reverse Engineering Assessment and Methods, is a platform for crowd-sourced studies that focus on developing computational tools to solve biomedical problems. Essentially, it's a competition that serves as a large, long-standing, international scientific collaboration.

Acute myeloid leukemia presented a worthy challenge since there is no single genetic cause of the disease, which makes it hard to select treatments for patients suffering from the deadly cancer of the blood, Qutub said.

The DREAM 9 patient data set was collected by Steven Kornblau, a leukemia doctor and professor at MD Anderson. The data was distributed to DREAM 9 participants online through Sage Bionetworks' Synapse web portal and through Biowheel, a cloud-based technology launched by the Qutub Lab.

Biowheel is an interactive tool to visualize and group high-dimensional data of all kinds. It was developed by Rice graduate student Chenyue Wendy Hu, undergraduate alumnus Alex Bisberg and Qutub. National Library of Medicine postdoctoral fellow David Noren and research scientist Byron Long, also of the Qutub Lab, are lead authors of the paper.

For DREAM 9, each team was presented with training data from 191 patients that included demographic information like age and gender and more complex proteomic and phosphoprotein data that describes signaling protein pathways believed to play a role in the disease.

The competition used a test data set from 100 patients that didn't include outcomes, such as whether patients responded to therapy, relapsed, survived or died.

The primary challenge was to see how well the teams' algorithms could predict how patients responded to chemotherapy. The eventual goal is to give clinicians a predictive tool to develop individualized treatment plans.

The top-performing models were by Team EvoMed (Li Liu) of Arizona State University and Team Chipmunks (Honglei Xie, Greg Chen, Xihui Lin, Geoffrey Hunter) of the Ontario Institute for Cancer Research, Toronto. They were best able to predict patient response to therapy with an accuracy of close to 80 percent, Qutub said.

She noted that one interesting takeaway was that overall the 31 models found it harder to predict outcomes for patients classified as "resistant to therapy" than for responsive patients. The median model prediction accuracy for resistant patients was 42 percent vs 73 percent for responsive patients. The winning models were impacted by the perturbation of signaling proteins known as phosphoinositide-3-kinase (a cell-cycle regulator) and NPM1 (which contributes to ribosome assembly and chromatin regulation), singling them out as strong candidates for further study.

The Qutub Lab became involved in leading DREAM 9 after the design of Biowheel won a DREAM 8 subchallenge three years ago. Five Qutub Lab members contributed predictive algorithms to the earlier challenge, which focused on proteomic analysis of breast cancer from MD Anderson data.

In discussions with DREAM organizer Gustavo Stolovitzky of IBM, Qutub had suggested a challenge based on one of the leukemia data sets Kornblau and her lab were analyzing to help understand molecular signaling in cancer.

"We used DREAM as a way to get general insight into making more accurate predictive models of clinical outcomes," Qutub said. "Steve (Kornblau), who runs the core banking facility for leukemia patients at MD Anderson Cancer Center, had the foresight to start gathering and banking patient biopsy samples when he was a resident over 25 years ago. The bank is a fantastic resource and a tremendous gift to the public. Genomic and proteomic analysis on a portion of these patient biopsies served as the basis for DREAM."

Because judging the entries was so computationally demanding, the Qutub Lab enlisted Erik Engquist, a co-author of the paper and director of the Center for Research Computing, and Rice's Ken Kennedy Institute for Information Technology (K2I) to help direct data traffic. Engquist helped the lab ensure a level playing field as competitors' algorithms ran on several of the university's high-performance computing platforms. He also helped set up a server to share challenge data via Biowheel, Qutub said.

"We had more than 270 participants and several dozen models to vet. K2I was instrumental in helping us run the challenge," she said.

Before DREAM 9 began, Noren spent considerable time designing the challenge and processing the complex patient data set. During and following the challenge, Noren, Long and the IBM team spent months processing the mountain of output data so the models, which analyzed 40 clinical indicators and 231 gene-expression profiles for each patient, would get a fair comparison. (The Rice lab did not compete because, as administrator, it already knew the results.)

Noren's task was to compare how well each model performed for each patient and to see whether the top-performing models had unique input parameters or features, Qutub said. "This way, we can start to learn which features of patients uniquely predict their outcomes."

The results still only hinted at the complexity of determining an optimal leukemia treatment plan, she said. Qutub's lab is using what it learned from its DREAM experience as a basis for experimentation on leukemia cell lines and test whether targeting specific sets of proteins offer a therapeutic advantage.


Story Source:

Materials provided by Rice University. Note: Content may be edited for style and length.


Journal Reference:

  1. David P. Noren, Byron L. Long, Raquel Norel, Kahn Rrhissorrakrai, Kenneth Hess, Chenyue Wendy Hu, Alex J. Bisberg, Andre Schultz, Erik Engquist, Li Liu, Xihui Lin, Gregory M. Chen, Honglei Xie, Geoffrey A. M. Hunter, Paul C. Boutros, Oleg Stepanov, Thea Norman, Stephen H. Friend, Gustavo Stolovitzky, Steven Kornblau, Amina A. Qutub. A Crowdsourcing Approach to Developing and Assessing Prediction Algorithms for AML Prognosis. PLOS Computational Biology, 2016; 12 (6): e1004890 DOI: 10.1371/journal.pcbi.1004890

Cite This Page:

Rice University. "Crowd-sourced competition to create 'big data' diagnostic tools." ScienceDaily. ScienceDaily, 29 June 2016. <www.sciencedaily.com/releases/2016/06/160629105805.htm>.
Rice University. (2016, June 29). Crowd-sourced competition to create 'big data' diagnostic tools. ScienceDaily. Retrieved November 22, 2024 from www.sciencedaily.com/releases/2016/06/160629105805.htm
Rice University. "Crowd-sourced competition to create 'big data' diagnostic tools." ScienceDaily. www.sciencedaily.com/releases/2016/06/160629105805.htm (accessed November 22, 2024).

Explore More

from ScienceDaily

RELATED STORIES