Phase II/III terminations reveal repeatable failure modes. Learn where AI can spot early warning signals and reduce costly late-stage trial stops.

Phase II/III Trial Terminations: Where AI Pays Off
A phase II or phase III trial doesn’t “fail” in a single moment. It usually unravels in slow motion: a safety signal that looks minor at first, an endpoint that stops separating, recruitment that drags, a protocol amendment that changes the population just enough to muddy the readout.
Nature Reviews Drug Discovery published an analysis of phase II and phase III clinical trial terminations from 2013–2023 (from the analyst’s couch, dated 18 December 2025). The practical value isn’t the history lesson—it’s the diagnostic lens. Termination patterns tell you where drug development predictably breaks, and that’s exactly where AI in clinical trials should be held accountable: spotting early warning signals, improving decisions before sponsors burn another year and another nine figures.
Here’s the stance I’ll take: most organizations treat trial termination as a post-mortem topic. The smarter move is to treat it as a design input—something you can model, monitor, and actively prevent.
What trial terminations really measure (and why teams misread them)
Answer first: Trial termination is a lagging indicator of upstream issues in evidence strategy, execution, or asset quality—and it’s more actionable than “success rate” because it exposes failure mechanisms.
Success rates get all the attention because they roll up neatly into portfolio math. But termination is different: it’s the point where a team decides the probability of success no longer justifies the cost, time, or risk. That decision can be triggered by:
- Science (efficacy isn’t there, biomarker hypothesis fails)
- Safety (unacceptable adverse events, class effects, dose limitations)
- Operations (recruitment stalls, site performance lags, supply issues)
- Strategy/finance (competition changes, funding dries up, M&A reprioritization)
Those categories matter because AI can help with some far more than others.
Termination isn’t always “failure”—but it’s always expensive
A terminated phase III can still teach you something clinically meaningful, but you don’t get reimbursed for “learning.” The cost shows up as:
- sunk trial spend (sites, CRO, drug supply)
- opportunity cost (teams tied up, delayed next asset)
- reputational cost (investor narrative, partner confidence)
That’s why termination analysis is so useful: it focuses attention on avoidable waste.
Patterns from 2013–2023: what the last decade signals
Answer first: The decade-long view shows termination is not a rare anomaly; it’s a recurring feature of late-stage development that clusters around a few predictable breakpoints.
The Nature Reviews Drug Discovery piece frames phase II and III success as major drivers of R&D efficiency and uses terminations to surface trends that could inform better trial design and conduct. Even without turning this into a spreadsheet exercise, a few practical implications come through clearly:
- Phase II remains the portfolio “truth serum.” It’s often where biology meets real-world heterogeneity. Many programs look solid in phase I (tolerability, PK/PD) and then stall once endpoints, subpopulations, and standard-of-care realities hit.
- Phase III terminations are frequently execution-amplified. By the time you’re in phase III, the question often shifts from “does it work at all?” to “can we prove it cleanly, fast enough, and safely enough against current care?” Small operational weaknesses become outcome-threatening.
- The reasons aren’t purely scientific. Competitive landscape shifts, recruitment bottlenecks, and protocol complexity can be just as lethal as lack of efficacy.
The uncomfortable takeaway: late-stage risk is partly self-inflicted
A lot of organizations still run trials as if the main uncertainty is the molecule. The reality is that late-stage outcomes are a system property—molecule, endpoint selection, inclusion/exclusion, site network, adherence, regional standard-of-care, data quality, and monitoring all interact.
That’s good news, because systems can be improved.
Where AI can prevent phase II/III terminations (not just explain them)
Answer first: AI prevents terminations when it’s embedded in decision loops—design, feasibility, monitoring, and adaptive actions—not when it’s used as a retrospective dashboard.
If your AI strategy is “we’ll run a model after the interim readout,” you’re already late. The highest-ROI use cases show up earlier and run continuously.
1) Trial design: endpoint, population, and signal-to-noise
Many phase II programs die because the trial can’t separate drug effect from noise. AI helps by quantifying the noise before you commit.
Practical ways AI supports better design:
- Synthetic feasibility cohorts: Use historical EHR/claims + registries to estimate how many patients match criteria as written, not as imagined.
- Eligibility criteria optimization: Model how each criterion changes enrollment speed, event rates, and population representativeness.
- Endpoint sensitivity modeling: Predict which endpoints are most likely to show separation given expected effect size and variability.
Snippet-worthy truth: If your endpoint is fragile, your phase III will be fragile—no matter how good the drug is.
2) Site and enrollment intelligence: operational failures are predictable
Terminations blamed on “slow enrollment” rarely come out of nowhere. The signals show up months earlier: screen failures spike, dropout creeps up, visit windows slip, sites enroll one patient and stop.
AI-enabled approaches that actually work:
- Site performance prediction using prior trial metrics (startup time, screen failure rates, query burden, deviation history)
- Geographic enrollment forecasting that reflects real referral patterns and competing trials
- Patient matching models that estimate “reachable” patients, not just “eligible” patients
If you can predict that half your sites will under-enroll, you can restructure the plan before timelines become fantasy.
3) Safety surveillance: catching weak signals early
Safety-related terminations are the hardest to “optimize,” but they’re not immune to earlier detection. AI can help teams see patterns faster—especially when safety signals are diffuse across subgroups.
High-value techniques include:
- Bayesian signal detection across adverse event categories and severity grades
- Subpopulation risk modeling (comorbidity clusters, concomitant meds, organ function)
- Dose-exposure-response monitoring to detect when tolerability will cap efficacy
This doesn’t replace medical judgment. It changes the speed and clarity of the evidence medical judgment relies on.
4) Data quality and protocol complexity: silent drivers of failure
Some trials don’t terminate because the drug is ineffective; they terminate because the data can’t credibly answer the question.
AI can reduce “avoidable ambiguity” by:
- flagging anomalous data patterns at site/patient level (potential errors or misconduct)
- predicting protocol deviation risk based on visit schedules, assessment burden, and site workload
- optimizing monitoring to focus on the few variables that truly determine interpretability
Opinionated take: A complex protocol is a tax on every site, every visit, every patient. AI should be used to lower that tax—not to cope with it later.
A practical playbook: turning termination trends into an AI roadmap
Answer first: Start with termination drivers, map them to measurable leading indicators, then deploy models that trigger specific actions—not just alerts.
If you want AI to reduce phase II/III terminations, you need a workflow that connects analytics to decisions. Here’s a structure I’ve seen work well.
Step 1: Build a “termination taxonomy” that matches how your company operates
Don’t copy generic categories. Make them decision-relevant. For example:
- efficacy miss (mechanism)
- efficacy miss (endpoint)
- safety/tolerability ceiling
- enrollment infeasible
- data not interpretable
- strategic stop (competition/portfolio)
Then tag past programs. The goal is not perfect labeling—it’s a usable signal.
Step 2: Define leading indicators for each termination type
Examples of leading indicators you can measure early:
- rising screen failure rate (enrollment infeasible)
- increasing missingness in key endpoints (data not interpretable)
- subgroup AE clustering (safety risk)
- drifting standard-of-care in key regions (strategic and endpoint risk)
Step 3: Choose models that match the decision
Not every problem needs deep learning. A simple model that triggers a clear action beats a complex model that produces debate.
- Classification models for site risk tiering
- Time-to-event models for enrollment and dropout forecasting
- Causal inference for endpoint and population choices
- NLP for protocol deviation narratives and AE descriptions
Step 4: Hardwire actions and owners
A prediction without an owner is a report. A prediction with an owner is a control system.
Define in advance:
- what threshold triggers escalation
- who approves mitigation (clinical ops, safety, stats, program lead)
- what mitigation looks like (add sites, revise criteria, enrich population, adjust monitoring)
Common questions pharma teams ask (and the answers that hold up)
“Can AI really predict phase II failure?”
Answer first: It can predict risk of failure modes—especially enrollment, adherence, data quality, and subgroup-specific response—better than it can predict a binary win/lose outcome.
The biggest wins come from predicting what’s controllable. Biology remains biology.
“Won’t regulators push back on AI-driven trial decisions?”
Answer first: Regulators care about transparency, bias control, and patient safety—not whether you used machine learning.
If AI is guiding design or monitoring, document it like any other method: validation, governance, audit trails, and clear human accountability.
“What’s the fastest AI project that reduces termination risk?”
Answer first: Site and enrollment forecasting paired with criteria optimization.
It’s faster because the data is usually available (internal trial ops + external feasibility signals), and the intervention is straightforward.
What this means for AI in Pharmaceuticals & Drug Discovery in 2026
Termination analysis from 2013–2023 should change how you prioritize AI. The flashiest use case isn’t always the highest ROI. If your late-stage trials are terminating due to predictable operational and evidence-strategy breakdowns, then AI belongs in the engine room: feasibility, design optimization, monitoring, and adaptive mitigation.
If you’re building an AI in drug discovery stack but ignoring clinical development execution, you’re leaving the most immediate value on the table.
A good next step is brutally simple: pick one termination driver you see repeatedly (enrollment failure, endpoint fragility, data quality), define two leading indicators you can measure this quarter, and pilot a model that triggers a specific operational decision. Then scale what works.
The open question for every R&D leader heading into 2026 is this: are you using AI to tell better stories about why trials failed—or to stop the next termination before it starts?