Explainable AI model achieves breakthrough accuracy in ovarian cancer detection

The researchers argue that the integration of explainable AI into clinical decision-making pipelines could reshape cancer screening by improving diagnostic precision while maintaining human oversight. Combining these models with imaging data and genomic markers could further enhance predictive performance and personalization of treatment strategies.


CO-EDP, VisionRICO-EDP, VisionRI | Updated: 27-10-2025 09:37 IST | Created: 27-10-2025 09:37 IST
Explainable AI model achieves breakthrough accuracy in ovarian cancer detection
Representative Image. Credit: ChatGPT

Turkish researchers have developed an artificial intelligence model that could transform early ovarian cancer detection by offering both high diagnostic accuracy and clinical transparency.

The study, titled "Explainable Artificial Intelligence for Ovarian Cancer: Biomarker Contributions in Ensemble Models" and published in Biology, introduces an explainable machine learning system that surpasses the performance of conventional clinical indices while providing interpretability crucial for medical decision-making.

Can artificial intelligence outperform traditional ovarian cancer screening methods?

The study addresses one of the prominent challenges in gynecologic oncology, accurate and early differentiation between benign and malignant ovarian tumours. Despite advancements in imaging and biomarker testing, existing tools such as the Risk of Ovarian Malignancy Algorithm (ROMA) often produce ambiguous results, especially in cases where biomarker levels fall near diagnostic thresholds.

To improve clinical precision, the researchers trained five ensemble machine learning models using clinical and laboratory data from 309 female patients. The dataset included 47 features, such as age, tumor markers (CA125, HE4, CEA, CA19-9, AFP), liver enzymes, blood indices, and electrolytes. Using the Boruta feature selection algorithm, the team identified 19 variables that most strongly correlated with malignancy risk.

Among the tested models, Gradient Boosting, CatBoost, XGBoost, LightGBM, and Random Forest, the Gradient Boosting model achieved the highest performance, with an accuracy of 88.99%, AUC-ROC score of 0.934, and Matthews correlation coefficient (MCC) of 0.782. These results not only demonstrated the algorithm's robustness but also showed that the model outperformed the ROMA index, which reached an AUC of only 0.89 on the same dataset.

The Gradient Boosting model also demonstrated better sensitivity at 90% specificity, detecting 82% of malignant cases compared to ROMA's 78%. This improvement is significant for clinical screening, where missed cancers can have serious consequences.

How explainable AI brings trust and clarity to clinical decision-making

The research focuses on explainability, a crucial factor for real-world adoption in healthcare. Many deep learning and complex ensemble methods act as "black boxes," making it difficult for physicians to understand how predictions are derived. To overcome this, the authors incorporated explainable artificial intelligence (XAI) frameworks, specifically SHAP (SHapley Additive exPlanations) and LIME (Local Interpretable Model-Agnostic Explanations), to interpret model outputs at both global and individual levels.

SHAP analysis identified HE4, CEA, globulin (GLO), CA125, and age as the most influential predictors of malignancy. These biomarkers align with established oncological evidence, confirming that AI can reinforce existing medical understanding while revealing additional patterns hidden in clinical data.

On a case-by-case basis, LIME visualizations provided local explanations for each patient, showing how specific feature values influenced the AI's decision toward either benign or malignant classification. For instance, elevated HE4 and CA125 levels or abnormal liver profiles pushed the prediction toward malignancy, while normal globulin and enzyme levels supported benign outcomes.

This level of transparency enables clinicians to validate model decisions, increasing confidence in AI-assisted triage and potentially integrating such tools into everyday diagnostic workflows. By revealing the weight and direction of each variable, the model effectively bridges the gap between data science and clinical reasoning.

What the results mean for future ovarian cancer diagnosis

The findings suggest that machine learning models based on routine clinical and laboratory data can serve as non-invasive, cost-effective screening tools, complementing imaging and traditional biomarkers. The system's high predictive accuracy and interpretability position it as a promising decision-support platform in gynecologic oncology, particularly in settings with limited access to specialized diagnostic infrastructure.

The authors also examined error patterns to identify model limitations. False negatives, cases where malignancy was misclassified as benign, were mainly associated with borderline biomarker values or atypical patient profiles, such as those with low HE4 and CA125 levels. This analysis emphasizes the importance of conservative threshold calibration to minimize missed cancers.

The authors note that the dataset was drawn from a single medical center, which limits generalizability. They recommend external and prospective validation across larger, multi-center cohorts to confirm reproducibility before clinical deployment. Despite this limitation, the study represents a significant advance toward reliable, explainable AI-driven diagnostics in ovarian cancer care.

The researchers argue that the integration of explainable AI into clinical decision-making pipelines could reshape cancer screening by improving diagnostic precision while maintaining human oversight. Combining these models with imaging data and genomic markers could further enhance predictive performance and personalization of treatment strategies.

A step toward trustworthy AI in women's health

The study validates that ensemble learning algorithms can achieve near-expert diagnostic accuracy while preserving interpretability, a vital step for medical AI acceptance.

In the broader context of women's health, the research reinforces the role of data-driven tools in addressing diseases where early detection remains difficult. Ovarian cancer often goes undiagnosed until advanced stages, contributing to its high mortality rate. The model's ability to leverage standard laboratory tests for early risk stratification could make screening more accessible, especially in resource-limited regions.

Looking ahead, the authors advocate a hybrid clinical-AI framework, where explainable models act as second opinions rather than replacements for physicians. With transparent predictive reasoning and rigorous validation, such systems could support more accurate and equitable cancer care globally.

  • FIRST PUBLISHED IN:
  • Devdiscourse

TRENDING

DevShots

Latest News

OPINION / BLOG / INTERVIEW

From Cash to Code: How CBDCs Could Reshape the Future of Social Assistance

Mauritius Adopts IMF’s QPM Model to Strengthen Inflation Targeting and Policy Forecasting

Nutrition with Caution: WHO’s New Rules on Fortifying Oils for Public Health

Greening Health Systems: Lao PDR’s Path to Climate-Ready and Inclusive Primary Care

Connect us on

LinkedIn Quora Youtube RSS
Give Feedback