Artificial Intelligence Accurately Predicts Radiation Treatment Side Effects
|
By HospiMedica International staff writers Posted on 01 Oct 2019 |

Image: New research has shown that a computer model can predict side effects associated with radiation therapy (Photo courtesy of Technology Networks).
Researchers from the University of Texas MD Anderson Cancer Center (Houston, Texas, USA) have demonstrated that a sophisticated computer model can accurately predict two of the most challenging side effects associated with radiation therapy for head and neck cancer. This precision oncology approach has the potential to better identify patients who might benefit from early interventions that could help prevent significant weight loss after treatment or reduce the need for feeding tube placement.
The team of researchers developed models to analyze large sets of data merged from three sources: electronic health records (Epic), an internal web-based charting tool (Brocade) and the record/verify system (Mosaiq). The data included more than 700 clinical and treatment variables for patients with head and neck cancer (75% male/25% female, with a median age of 62 years) who received more than 2,000 courses of radiation therapy (median dose 60 Gy) across five practice sites at MD Anderson from 2016 to 2018.
The researchers used the models to predict three endpoints: significant weight loss, feeding tube placement and unplanned hospitalizations. Results from the best-performing model were then validated against 225 subsequent consecutive radiation therapy treatments. Models with a performance rate that met a pre-specified threshold of area under the curve (AUC) of 0.70 or higher were considered clinically valid (an AUC score of 1.0 would mean the model’s predictions were 100% accurate, while a score of 0.0 would mean the predictions were never accurate). The models predicted the likelihood of significant weight loss (AUC = 0.751) and need for feeding tube placement (AUC = 0.755) with a high degree of accuracy.
“Being able to identify which patients are at greatest risk would allow radiation oncologists to take steps to prevent or mitigate these possible side effects,” said Jay Reddy, MD, PhD, an assistant professor of radiation oncology at The University of Texas MD Anderson Cancer Center and lead author on the study. “If the patient has an intermediate risk, and they might get through treatment without needing a feeding tube, we could take precautions such as setting them up with a nutritionist and providing them with nutritional supplements. If we know their risk for feeding tube placement is extremely high – a better than 50% chance they would need one – we could place it ahead of time so they wouldn’t have to be admitted to the hospital after treatment. We’d know to keep a closer eye on that patient.”
The machine learning approach cannot isolate the single-most predictive factor or combination of factors that lead to negative side effects, but can provide patients and their clinicians with a better understanding of what to expect during the course of treatment. In addition to predicting the likelihood of side effects, machine learning models could potentially predict which treatment plans would be most effective for different types of patients and allow for more personalized approaches to radiation oncology.
“Machine learning can make doctors more efficient and treatment safer by reducing the risk of error,” added Dr. Reddy. “It has the potential for influencing all aspects of radiation oncology today – anything where a computer can look at data and recognize a pattern.”
Related Links:
University of Texas MD Anderson Cancer Center
The team of researchers developed models to analyze large sets of data merged from three sources: electronic health records (Epic), an internal web-based charting tool (Brocade) and the record/verify system (Mosaiq). The data included more than 700 clinical and treatment variables for patients with head and neck cancer (75% male/25% female, with a median age of 62 years) who received more than 2,000 courses of radiation therapy (median dose 60 Gy) across five practice sites at MD Anderson from 2016 to 2018.
The researchers used the models to predict three endpoints: significant weight loss, feeding tube placement and unplanned hospitalizations. Results from the best-performing model were then validated against 225 subsequent consecutive radiation therapy treatments. Models with a performance rate that met a pre-specified threshold of area under the curve (AUC) of 0.70 or higher were considered clinically valid (an AUC score of 1.0 would mean the model’s predictions were 100% accurate, while a score of 0.0 would mean the predictions were never accurate). The models predicted the likelihood of significant weight loss (AUC = 0.751) and need for feeding tube placement (AUC = 0.755) with a high degree of accuracy.
“Being able to identify which patients are at greatest risk would allow radiation oncologists to take steps to prevent or mitigate these possible side effects,” said Jay Reddy, MD, PhD, an assistant professor of radiation oncology at The University of Texas MD Anderson Cancer Center and lead author on the study. “If the patient has an intermediate risk, and they might get through treatment without needing a feeding tube, we could take precautions such as setting them up with a nutritionist and providing them with nutritional supplements. If we know their risk for feeding tube placement is extremely high – a better than 50% chance they would need one – we could place it ahead of time so they wouldn’t have to be admitted to the hospital after treatment. We’d know to keep a closer eye on that patient.”
The machine learning approach cannot isolate the single-most predictive factor or combination of factors that lead to negative side effects, but can provide patients and their clinicians with a better understanding of what to expect during the course of treatment. In addition to predicting the likelihood of side effects, machine learning models could potentially predict which treatment plans would be most effective for different types of patients and allow for more personalized approaches to radiation oncology.
“Machine learning can make doctors more efficient and treatment safer by reducing the risk of error,” added Dr. Reddy. “It has the potential for influencing all aspects of radiation oncology today – anything where a computer can look at data and recognize a pattern.”
Related Links:
University of Texas MD Anderson Cancer Center
Latest AI News
- AI Analysis of Pericardial Fat Refines Long-Term Heart Disease Risk
- Machine Learning Approach Enhances Liver Cancer Risk Stratification
- New AI Approach Monitors Brain Health Using Passive Wearable Data
- AI Tool Maps Early Risk Patterns in Bloodstream Infections
- AI Model Identifies Rare Endocrine Disorder from Hand Images
- AI Tool Promises to Reduce Length of Hospital Stays and Free Up Beds
- Machine Learning Model Cuts Canceled Liver Transplants By 60%
Channels
Artificial Intelligence
view channelAI Analysis of Pericardial Fat Refines Long-Term Heart Disease Risk
Accurately identifying long-term cardiovascular disease risk in asymptomatic adults remains challenging for clinicians. Missed or underestimated risk delays preventive therapy and increases the chance... Read more
Machine Learning Approach Enhances Liver Cancer Risk Stratification
Hepatocellular carcinoma, the most common form of primary liver cancer, is often detected late despite targeted surveillance programs. Current screening guidelines emphasize patients with known cirrhosis,... Read moreCritical Care
view channel
Noninvasive Monitoring Device Enables Earlier Intervention in Heart Failure
Hospitalizations for heart failure with preserved ejection fraction (HFpEF) remain common because lung congestion often worsens before symptoms prompt treatment changes. Missed early decompensation... Read more
Automated IV Labeling Solution Improves Infusion Safety and Efficiency
Medication administration in high-acuity settings is often complicated by multiple concurrent infusions, making accurate line identification essential. In a 10-hospital intensive care unit study, 60% of... Read moreSurgical Techniques
view channel
Ultrasound Technology Aims to Replace Invasive BPH Procedures
Benign prostatic hyperplasia (BPH) is a frequent cause of lower urinary tract symptoms in aging men and often requires invasive procedures or prolonged recovery. With prevalence expected to rise as populations... Read more
Continuous Monitoring with Wearables Enhances Postoperative Patient Safety
Postoperative hypoxemia on general surgical wards is common and often missed by intermittent vital sign checks. Undetected low oxygen levels can delay recovery and raise the risk of complications that... Read morePatient Care
view channel
Wearable Sleep Data Predict Adherence to Pulmonary Rehabilitation
Chronic obstructive pulmonary disease (COPD) is a long-term lung disorder that makes breathing difficult and often disturbs sleep, reducing energy for daily activities. Limited engagement in pulmonary... Read more
Revolutionary Automatic IV-Line Flushing Device to Enhance Infusion Care
More than 80% of in-hospital patients receive intravenous (IV) therapy. Every dose of IV medicine delivered in a small volume (<250 mL) infusion bag should be followed by subsequent flushing to ensure... Read moreHealth IT
view channel
EMR-Based Tool Predicts Graft Failure After Kidney Transplant
Kidney transplantation offers patients with end-stage kidney disease longer survival and better quality of life than dialysis, yet graft failure remains a major challenge. Although a successful transplant... Read more
Printable Molecule-Selective Nanoparticles Enable Mass Production of Wearable Biosensors
The future of medicine is likely to focus on the personalization of healthcare—understanding exactly what an individual requires and delivering the appropriate combination of nutrients, metabolites, and... Read moreBusiness
view channel







