Wednesday, April 1, 2026
Review finds predictive models for pancreatitis after ERCP show moderate accuracy
Photo by Cht Gsml / Unsplash

Review finds predictive models for pancreatitis after ERCP show moderate accuracy

Plain Language Summary
What this means for you:
Computer models for predicting pancreatitis risk after ERCP show promise but are not ready for routine clinical use.

Researchers reviewed 23 studies that developed computer models to predict a patient's risk of developing pancreatitis after an ERCP procedure. ERCP is a common test for digestive problems. The models used patient information like age, medical history, and details about the procedure to estimate risk.

The review found that the models, when tested on new groups of patients, had moderate accuracy. Machine learning models were slightly more accurate than traditional statistical models. Common factors the models used to predict higher risk included being female, having a history of pancreatitis, and having a difficult procedure.

It's important to know this research is about the models themselves, not a new treatment. The studies had significant limitations, including a lack of real-world testing for many models and potential bias. This means doctors cannot confidently use these models in clinics yet. The findings suggest these tools have potential, but they need much more development and validation to be reliable for patient care.

What this means for you:
Computer models for predicting pancreatitis risk after ERCP show promise but are not ready for routine clinical use.
Read the Full Clinical Summary →
View Original Abstract ↓
Background and aimsPost-ERCP pancreatitis (PEP) is the most common complication following ERCP, leading to significant clinical and economic consequences. Predictive models for PEP can help identify high-risk patients and guide preventive strategies. However, the performance of these models varies, and a comprehensive evaluation is lacking. This study aims to assess the accuracy, reliability, and risk of bias in existing predictive models for PEP.MethodsA comprehensive search was conducted across five databases (PubMed, Embase, Web of Science, Cochrane Library, and CNKI) for studies published until January 2025. Studies that developed or validated predictive models for PEP were included. Models with external validation sets were included in a meta-analysis. Model performance was assessed using the area under the receiver operating characteristic curve (AUC), sensitivity, specificity, and calibration. A random-effects meta-analysis was performed, with heterogeneity assessed using I² statistics. Data extraction and risk of bias were conducted using a standardized template combining the CHARMS and PROBAST tools.ResultsTwenty-three studies (21 model development studies and 2 external validation studies) were included, presenting 21 predictive models for PEP. Nine models incorporated external validation, with one study recalibrating an existing model and another externally validating two prior models. The mean events per variable (EPV) across studies was 10.2 (2.2 to 22.4). The pooled AUC for externally validated models was 0.79 (95% CI: 0.75–0.83). Machine learning models demonstrated higher AUC (0.84) than traditional logistic regression models (0.76). Common predictive factors included difficult cannulation, female sex, pancreatic duct dilation, and a history of pancreatitis.ConclusionsPredictive models for PEP show potential for improving patient risk stratification. However, variability in model performance, lack of external validation, and significant bias in many studies limit their clinical applicability. Further external validation, model refinement, and improved bias control are essential for broader clinical implementation.Systematic Review Registrationhttps://www.crd.york.ac.uk/PROSPERO/view/CRD42024626168, identifier CRD42024626168.