Vienna, Austria

ESTRO 2023

Session Item

Sunday
May 14
09:00 - 10:00
Business Suite 1-2
Head & neck
Jon Cacicedo, Spain
Poster Discussion
Clinical
Systematic review of NTCP models for radiation-induced toxicities in head and neck cancer
Makbule Tambas, The Netherlands
PD-0403

Abstract

Systematic review of NTCP models for radiation-induced toxicities in head and neck cancer
Authors:

Makbule Tambas1, Toshihiko Takada2,3, Enrico Clementel4, Artuur M Leeuwenberg5, Marjan Sharabiani4, Johanna AAG Damen6, Zoë Dunias2, Jan F Nauta1, Karel GM Moons7, Johannes A Langendijk1, Ewoud Schuit2

1University of Groningen, University Medical Center Groningen, Department of Radiation Oncology, Groningen, The Netherlands; 2Utrecht University, University Medical Center Utrecht, Julius Center for Health Sciences and Primary Care, Utrecht, The Netherlands; 3 Fukushima Medical University, Shirakawa Satellite for Teaching And Research (STAR), Department of General Medicine, Shirakawa, Japan; 4European Organisation for Research and Treatment of Cancer (EORTC), Headquarters, Brussels, Belgium; 5Utrecht University, University Medical Center Utrecht, Julius Center for Health Sciences and Primary Care, Utrecht, The Netherlands; 6Cochrane Netherlands, Utrecht University, University Medical Center Utrecht, Julius Center for Health Sciences and Primary Care , Utrecht, The Netherlands; 7Cochrane Netherlands, Utrecht University, University Medical Center Utrecht, Julius Center for Health Sciences and Primary Care, Utrecht, The Netherlands

Show Affiliations
Purpose or Objective

The aim of this Cochrane review was to identify, describe and appraise NTCP-models and their quality and predictive performance in patients with HNC.

Material and Methods

We searched Ovid MEDLINE, Ovid Embase, and registries of ongoing trials from conception to March 2021. Articles reporting on the development and external validation of NTCP models were included to predict any type of radiation-induced side effects. We included all study designs, except for non-nested case-control studies. Data was extracted using CHARM and applicability concern and risk of bias (ROB) were assessed using the PROBAST tool.

Results

A total of 10,119 unique potentially relevant articles and 55 ongoing trials were identified. Of those, 9,419 were excluded based on title and abstract screening, leaving 755 articles for full text screening. During full text screening, 641 articles were excluded. From the 114 remaining articles, data were extracted (Fig 1).

Fig 1. Study flow diagram

Among 333 developed models from 96 articles, 148 models aimed to predict toxicities related to salivary gland function (44%), followed by swallowing-related toxicities (18%), brain and nerve toxicities (7%) and hypothyroidism (6%). Calibration and discrimination of the apparent model were reported for 68 (20%) and 109 (33%) models, respectively. The c-statistics ranged between 0.60 and 0.98. Only 48 (14%) models from 5 articles were judged as low ROB and low concerns for applicability (Fig 2).

Fig 2. Assessment of risk of bias and applicability concerns


No external validation was performed for 244 models (73%). For the remaining 89 models and 4 additional models which were originally developed for patients with different types of cancer, 126 external validations were performed in 22 articles; most often for models that predicted outcomes related to saliva dysfunction (n=51, 40%), followed by swallowing (n=30, 24%) and hypothyroidism (n=24, 19%). The reported c-statistics after external validation ranged between 0.19 and 0.96. The overall ROB and applicability concerns were high in 67% and 10% of the external validations, respectively (Figure 2).

Six models (3 for dysphagia and 3 for hypothyroidism) were identified with ≥ 2 external validations for the same outcome as in the original development study. Those models generally showed good discriminative performance. However, information on their calibration performance was largely lacking and ROB was high. As no model was externally validated at least 5 times, we could not perform any meta-analyses of the performance of models across validation studies.

Conclusion

Among 333 developed models, a limited number of studies had sufficient quality. Only one-fourth of them was externally validated, of which only 6 models at least twice. Most validation studies were judged as high ROB. This review shows the need for more external validation studies of developed models in clinical practice and improvement of the quality of conducting and reporting of prediction model studies in this subject.