Big Data Archives - Sanford Burnham Prebys
Institute News

The Cancer Letter covers collaboration between Sanford Burnham Prebys and the National Cancer Institute to precisely prescribe cancer drugs

AuthorGreg Calhoun
Date

May 14, 2024

The May 10 issue of The Cancer Letter details a recent publication explaining the investigation of a new AI tool that may be able to match cancer drugs more precisely to patients.

The Cancer Letter—a news organization and weekly publication based in Washington, D.C., that focuses on cancer research and clinical care—included an article in its May 10 issue about a partnership between scientists at Sanford Burnham Prebys and the National Cancer Institute (NCI).

Authored by Sanju Sinha, PhD, assistant professor in the Cancer Molecular Therapeutics Program at Sanford Burnham Prebys, and the NCI’s Eytan Ruppin, MD, PhD, the “Trials & Tribulations” feature describes a first-of-its-kind computational tool to systematically predict patient response to cancer drugs at single-cell resolution. The study regarding this new tool was published on April 18, 2024, in the journal  Nature Cancer.

The Cancer Letter was founded in 1973 and focuses its coverage on the development of cancer therapies, drug regulation, legislation, cancer research funding, health care finance and public health.

Institute News

NIH director highlights Sanford Burnham Prebys and National Cancer Institute project to improve precision oncology

AuthorGreg Calhoun
Date

May 9, 2024

The NIH director’s blog features a recent publication detailing the study of a new AI tool that may be able to match cancer drugs more precisely to patients.

Monica M. Bertagnolli, MD, director of the National Institutes of Health (NIH), highlighted a collaboration between scientists at Sanford Burnham Prebys and the National Cancer Institute (NCI) on the NIH director’s blog. Bertagnoli noted advances that have been made in precision oncology approaches using a growing array of tests to uncover molecular or genetic profiles of tumors that can help guide treatments. She also recognizes that much more research is needed to realize the full potential of precision oncology.

The spotlighted Nature Cancer study demonstrates the potential to better predict how patients will respond to cancer drugs by using a new AI tool to analyze the sequences of the RNA within each cell of a tumor sample. Current precision oncology methods take an average of the DNA and RNA in all the cells in a tumor sample, which the research team hypothesized could hide certain subpopulations of cells—known as clones—that are more resistant to specific drugs.  

Bertagnoli said, “Interestingly, their research shows that having just one clone in a tumor that is resistant to a particular drug is enough to thwart a response to that drug. As a result, the clone with the worst response in a tumor will best explain a person’s overall treatment response.” 

More of Bertagnoli’s thoughts on this collaboration between scientists at Sanford Burnham Prebys and the NCI are available on the NIH director’s blog

Sanju Sinha, PhD, assistant professor in the Cancer Molecular Therapeutics Program at Sanford Burnham Prebys, is the first author on the featured study. 

Institute News

Media coverage of AI study predicting responses to cancer therapy ranks top 5% among published research

AuthorScott LaFee, Susan Gammon and Greg Calhoun
Date

April 29, 2024

Last week, Sanford Burnham Prebys and the National Cancer Institute shared findings regarding a first-of-its-kind computational tool to systematically predict patient response to cancer drugs at single-cell resolution.

Many news outlets and trade publications took note of this study and the computational tool’s potential future use in hospitals and clinics. This coverage placed the paper in the top 5% of all manuscripts ranked by Altmetric—a service that tracks and analyzes online attention of published research to improve the understanding and value of research and how it affects people and communities.

The results from the highlighted study were published on April 18, 2024, in the journal Nature Cancer.

“Our goal is to create a clinical tool that can predict the treatment response of individual cancer patients in a systematic, data-driven manner. We hope these findings spur more data and more such studies, sooner rather than later,” says first author Sanju Sinha, PhD, assistant professor in the Cancer Molecular Therapeutics Program at Sanford Burnham Prebys.

Here are a few of the venues that helped spread the word about this research: 

  • AP News: “Researchers … suggest that such single-cell RNA sequencing data could one day be used to help doctors more precisely match cancer patients with drugs that will be effective for their cancer.”
  • Politico, fourth story in Future Pulse newsletter: “Our hope is that being able to characterize the tumors on a single-cell resolution will enable us to treat and target potentially the most resistant and aggressive [cells], which are currently missed.”
  • NIH.gov: “The researchers discovered that if just one clone were resistant to a particular drug, the patient would not respond to that drug, even if all the other clones responded.”
  • Inside Precision Medicine: “The model was validated by predicting the response to monotherapy and combination treatment in three independent, recently published clinical trials for multiple myeloma, breast, and lung cancer.”

“I’m very pleased with how many news outlets covered our work,” Sinha says. “It is important and will help us continue improving the tool with more data so it can one day benefit cancer patients.”

Institute News

From postdoc to PI, it’s a journey. Don’t forget to pack some support

AuthorSanju Sinha
Date

December 15, 2023

The journal Nature Cancer asked a dozen early-career investigators to share their thoughts and experiences about starting their own labs in 2023. Among them: Sanju Sinha, PhD, who joined Sanford Burnham Prebys in June. Below is his essay. You can read the rest here.

Don’t forget to pack support 

Against the backdrop of a world emerging from a pandemic, starting my laboratory in 2023 was a whirlwind of excitement and anxiety, against the backdrop of a world emerging from a pandemic.

The goal for my laboratory is to understand cancer initiation and use this knowledge to develop preventative therapies—a goal appreciated by many, yet understudied and underfunded. We are aiming to achieve this by developing computational techniques based on machine-learning and leveraging big data from various sources, such as healthy tissues, pre-cancerous lesions and tumors. This journey has taken several unexpected turns, with its fair share of delights and challenges.

One significant hurdle appeared early: hiring. I recall the advice I received: “Forget it, you can’t hire a postdoc as an early-stage laboratory.” This made me ponder—if I were to choose right now, would I pursue a postdoc? My immediate answer was no. It struck me then: the traditional postdoc route needed a revamp.

Determined to instigate change, I introduced a new role: computational biologist. This position, an alternative to a postdoc, was tailored for transitioning to industry and offered better pay. The response was staggering—more than 400 applications.

Now, I’m proud to lead a fantastic team of three computational biologists from whom I am continually learning. This experience taught me a valuable lesson: crafting roles that serve both the goals of the laboratory and the career aspirations of the applicants can make a world of difference. I urge new principal investigators to shatter norms and design roles that provide fair compensation and smooth industry transition—reflecting the reality of the current job market.

However, the path to establishing a new laboratory was not without setbacks. Rejection is common in this field. I have already experienced a grant rejection and, considering the average grant success rate, I am prepared for many more.

Amid these challenges, my support system proved to be my lifeline. I’m grateful to be part of Sanford Burnham Prebys, which has proved to be more than just a top biomedical research institution. It is a community that provides unparalleled support for early principal investigators through generous startup packages, administrative assistance, hiring and grant-writing guidance, and a network of compassionate peers and mentors.

Equally important is my personal support system—my family, partner and friends who remind me that there is life beyond science, helping me maintain my well-being. This balance, I have realized, is the most crucial tool for anyone on a similar journey—so do not forget to pack support for the ride.

Institute News

Is cloud computing a game changer in cancer research? Three big questions for Lukas Chavez

AuthorMiles Martin
Date

February 22, 2023

As an assistant professor at Sanford Burnham Prebys and director of the Neuro-Oncology Molecular Tumor Board at Rady Children’s Hospital, Lukas Chavez, PhD, leverages modern technology for precision diagnostics and for uncovering new treatment options for the most aggressive childhood brain cancers.

We spoke to Chavez about his work and asked him how modern technology—particularly cloud computing—is shifting the approach to cancer research.

How are you using new technologies to advance your research?

New technologies are helping us generate a huge amount of data as well as many new types of data. All this new information at our disposal has created a pressing need for tools to make sense of it and maximize their benefits. That’s where computational biology and bioinformatics come into play. The childhood brain cancers I work on are very rare, which has historically made it difficult to study large numbers of cases and identify patterns.

Now, data for thousands of cases can be stored in the cloud. By creating data analysis tools, we can reveal insights that we would never have seen otherwise. For example, we’ve developed tools that can use patient data in the cloud to categorize brain cancers into subtypes we’ve never identified before, and we’re learning that there are many more types of brain tumors than we’ve previously understood. We’re basically transforming the classic histo-pathological approach that people have studied for decades by looking at tumor tissues under the microscope and turning that into data science.

How is cloud computing improving cancer research in general?

Assembling big datasets delays everything, so I believe the main idea of cloud computing is really to store data in the cloud, then bring the computational tools to the data, not the other way around.

My team did one study where we assembled publicly available data, and basically downloaded everything locally. The data assembly process alone took at least two to three years because of all the data access agreements and legal offices that were involved.

And that is the burden that cloud computing infrastructures remove. All of this personalized cancer data can be centrally stored in the cloud, which makes it available to more researchers while keeping it secure to protect patient privacy. Researchers can get access without downloading the data, so they are not responsible for data protection anymore. It’s both faster and more secure to just bring your tools to the data.

Are there any risks we need to be aware of?

Like any new technology, we need to be clear about how we use it. The technology is another tool in the toolbox of patient care. It will never entirely replace physicians and researchers, but it can complement and assist them.

Also, because we use costly and sophisticated tools that are being built and trained on very specific patient groups, we need to be careful that these tools are not only helping wealthier segments of society. Ideally, these tools will be expanded worldwide to help everybody affected by cancer.

Institute News

Three big questions for cutting-edge biologist Will Wang

AuthorMiles Martin
Date

January 26, 2023

Will Wang’s spatial omics approach to studying neuromuscular diseases is unique.

He works at the intersection of biology and computer science to study how complex systems of cells interact, specifically focusing on the connections between nerves, muscles, and the immune response and their role in neuromuscular diseases.

We sat down with Wang, who recently joined the Institute as an assistant professor, to discuss his work and how computer technology is shaping the landscape of biomedical research.

How is your team taking advantage of computer technology to study neuromuscular diseases?

No cell exists in isolation. All our cells are organized into complex tissues with different types of cells interacting with each other. We study what happens at these points of interaction, such as where nerves connect to muscle cells. Combining many different types of data such as single cell sequencing, spatial proteomics, and measures of cell-cell signaling helps us get a more holistic look at how interactions between cells determine tissue function, as well as how these interactions are disrupted in injury and disease. Artificial neural networks help us make sense of these different types of data by finding patterns and insights the human brain can’t see on its own. And because computers can learn from the vast modality of data that we gather, we can also use them to help predict how biological systems will behave in the lab. The process goes both ways – from biology to computers and from computers to biology. 

How will these technologies shape the future of biomedical research?

Biology and computer programming are two different languages. There are a lot of mathematicians and programmers who are great at coming up with solutions to process data, but biological questions can get lost in translation and it’s easy to miss the bigger picture. And pure biologists don’t necessarily understand the full scope of what computers can do for them. If we’re going to get the most out of this technology in biomedical research, we need people with enough expertise in both areas that they can bridge the gap, which is what our lab is trying to do. Over time we’re going to see more and more labs that combine traditional biological experiments and data analysis approaches with artificial intelligence and machine learning. 

Are there any potential risks to these new technologies?

Artifical intelligence is here to accelerate discovery. Mundane tasks and measurements that took me weeks to carry out as a graduate student can be automated to a matter of minutes. We can now find patterns in high dimensional images that the human brain can’t easily visualize. However, any kind of artificial intelligence comes with a certain amount of risk if people don’t understand when and how to use the tools. If you just take the absolute word of the algorithm, there will inevitably be times where it’s not correct. As scientists, we use artificial intelligence as a cutting-edge discovery tool, but need to validate the findings in terms of the biology. At the end of the day, it is us, scientists, who are here to drive the discovery process and design real life experiments to make sure our therapies are safe and efficacious. 

Institute News

Why share data from clinical trials? SBP’s CEO Perry Nisen weighs in

AuthorKristen Cusato
Date

September 26, 2016

Sharing clinical trial data with researchers who weren’t involved in the original study maximizes the value of patients’ participation, allowing more research questions to be answered than those of the original study. However, figuring out what data should be shared and how to do it has proven to be difficult.

The most recent issue of the New England Journal of Medicine devoted three perspective articles and an editorial on the topic of data sharing. Perry Nisen, MD, PhD, CEO of Sanford Burnham Prebys Medical Discovery Institute (SBP) and his colleagues discuss efforts to share clinical trial data and the hurdles that investigators still face.

“One of the risks is that there will not be a single simple system where these data can be accessed and analyzed, and the benefits of meta-analyzing data from multiple studies will be limited by cost and complexities,” said Nisen.

GlaxoSmithKline was a first mover in making anonymized patient-level data available from clinical trials. In 2013, the Clinical Study Data Request was established. The site is now managed by the Wellcome Trust, an independent, non-sponsor safe harbor, and includes more than 3,000 trials from 13 industry sponsors.

 

Nisen answers key questions about the future of clinical data sharing:

Q: Why should research sponsors go to the expense of sharing data?

Clinical data sharing is the right thing to do for science and society. First, it increases transparency of clinical trial data. It maximizes the contribution of trial participants to new knowledge and understanding. This allows researchers to confirm or refute findings, and enables them to generate other hypotheses. Scientific research globally is moving toward more transparency in clinical trial reporting and this is an important step toward building trust.

 

Q: What are the challenges to a one-stop shop for sharing all clinical trials data?

Protecting patient privacy and confidentiality is a major concern. Also, ensuring the data are used for valid scientific investigation, preventing erroneous claims of benefit or risk, and controlling the cost associated with anonymizing data in formats investigators can utilize effectively.

Other challenges inherent in data sharing include patient consent, data standards, standards for re-use, conflicts of interest, and intellectual property.

 

The editorial, also co-authored by Frank Rockhold, PhD, professor of biostatistics and bioinformatics at the Duke Clinical Research Institute, and Andrew Freeman, BSc, head of medical policy at GlaxoSmithKline, is available online here.

Institute News

Will you be part of the largest-ever clinical research study?

Authorjmoore
Date

March 23, 2016

It’s called the Precision Medicine Initiative (PMI) Cohort Program, and it was just announced in February by President Obama. If you join the cohort (group of subjects tracked over a long period of time), you can help researchers improve precision medicine, in which doctors select the treatments and preventive strategies that will work best for each patient. This program is just one component of the larger Precision Medicine Initiative announced during last year’s State of the Union address.

What’s the goal? According to NIH Director Francis Collins, the cohort program “seeks to extend precision medicine to all diseases by building a national research cohort of one million or more U.S. participants,” all enrolled by 2019.

Why recruit so many people? Since the program is intended to benefit people affected by many diseases and conditions, it must include large, representative samples of people with each type. Large samples increase the likelihood that studies using these data will find new associations and interactions among genes, environmental factors, and disease risk.

What will participants do? Volunteers will share their health records, complete surveys on lifestyle and environmental exposures, undergo a physical, and provide a biological sample (e.g. blood) for genetic testing.

How will people benefit? Participants will be considered partners in research—they’ll have access to their genetic data and, where possible, how their genes, surroundings, and habits affect their health. They’ll also have a say in how the research is conducted and what questions it should address.

Who’s running it? The NIH is overseeing the whole program, but it will be directly run from multiple institutions (which are currently being selected). The pilot phase will be led by Vanderbilt University and Verily (formerly Google Life Sciences).

What’s the cost? $130 million has been allotted in this fiscal year, but more money will be needed to keep the program going.

Should I be excited about it? Maybe. Some leaders in the health field have criticized the program for throwing money at the latest big thing instead of more low-tech problems like unequal access to healthcare, but such a huge data resource is bound to lead to answers to many important questions. 

What are the challenges for the PMI?

  • Scale—The program will generate one of the largest clinical databases yet, and it’s not clear how difficult it will be to make systems that can store and analyze it.
  • Privacy—Data will be anonymized, but keeping the health information of a million people in one place might represent a target for hackers sophisticated enough to figure out participants’ identities.
  • Interoperability—Health record systems are notoriously incompatible with one another. Though the PMI also has provisions to correct this, it likely won’t be a quick fix.

How can I sign up? Enrollment has not yet begun, but the NIH will announce when the public can get involved. So stay tuned…

 

Institute News

‘Big Data’ used to identify new cancer driver genes

Authorsgammon
Date

October 20, 2015

In a collaborative study led by Sanford Burnham Prebys Medical Discovery Institute (SBP), researchers have combined two publicly available ‘omics’ databases to create a new catalogue of ‘cancer drivers’. Cancer drivers are genes that when altered, are responsible for cancer progression. The researchers used cancer mutation and protein structure databases to identify mutations in patient tumors that alter normal protein-protein interaction (PPI) interfaces. The study, published today in PLoS Computational Biology, identified more than 100 novel cancer driver genes and helps explain how tumors driven by the same gene may lead to different patient outcomes.

“This is the first time that three-dimensional protein features, such as PPIs, have been used to identify driver genes across large cancer datasets,” said lead author Eduard Porta-Pardo, PhD, a postdoctoral fellow at SBP. “We found 71 interfaces in proteins previously unrecognized as cancer drivers, representing potential new cancer predictive markers and/or drug targets. Our analysis also identified several driver interfaces in known cancer genes, such as TP53, HRAS, PI3KCA and EGFR, proving that our method can find relevant cancer driver genes and that alterations in protein interfaces are a common pathogenic mechanism of cancer.”

Cancer is caused by the accumulation of mutations to DNA. Until now, scientists have focused on finding alterations in individual genes and cell pathways that can lead to cancer. But the recent push by the National Institutes of Health (NIH) to encourage data sharing has led to an era of unprecedented ability to systematically analyze large scale genomic, clinical, and molecular data to better explain and predict patient outcomes, as well as finding new drug targets to prevent, treat, and potentially cure cancer.

“For this study we used an extended version of e-Driver, our proprietary computational method of identifying protein regions that drive cancer. We integrated tumor data from almost 6,000 patients in The Cancer Genome Atlas (TCGA) with more than 18,000 three-dimensional protein structures from the Protein Data Bank (PDB),” said Adam Godzik, PhD, director of the Bioinformatics and Structural Biology Program at SBP. “The algorithm analyzes whether structural alterations of PPI interfaces are enriched in cancer mutations, and can therefore identify candidate driver genes.”

“Genes are not monolithic black boxes. They have different regions that code for distinct protein domains that are usually responsible for different functions. It’s possible that a given protein only acts as a cancer driver when a specific region of the protein is mutated,” Godzik explained. “Our method helps identify novel cancer driver genes and propose molecular hypotheses to explain how tumors apparently driven by the same gene have different behaviors, including patient outcomes.”

“Interestingly, we identified some potential cancer drivers that are involved in the immune system. With the growing appreciation of the importance of the immune system in cancer progression, the immunity genes we identified in this study provide new insight regarding which interactions may be most affected,” Godzik added.

The study was performed in collaboration with the European Bioinformatics Institute (UK), Centro de Investigación Principe Felipe (Spain), and CIBER de Enfermedades Raras (Spain).