Blog

Real-World Evidence, Adaptive Designs & Agentic AI: Bringing It All Together

Share

Imagine you’re six months into a pivotal study and emerging real-world data suggests a specific patient subgroup is responding far better than anticipated. The protocol is locked, the sample size fixed, and the endpoints predetermined. By the time the final analysis confirms the signal—often years later—the window to adapt the study has already closed.

Scenarios like this occur far more often than teams acknowledge. Traditional trial designs still operate on the assumption that everything important can be decided upfront: write the protocol, freeze it, execute, and analyze at the end. That model made sense in an era with slower data cycles and limited external evidence, long before continuous RWE streams and modern computational tools.

But that landscape has shifted. Adaptive designs, real-world evidence, and Agentic AI are now working together to support studies that can respond to new information as it emerges. Instead of relying solely on initial assumptions, modern trials combine controlled clinical data with real-world patient insights—powered by computational systems capable of managing the operational complexity that comes with this level of flexibility.

Why Traditional Protocols Can’t Keep Up

Traditional trial designs assume that most critical decisions—population selection, endpoints, sample size can be finalized upfront. This model was sufficient when patient populations were uniform, treatment options limited, and development timelines extended.

Today, those assumptions no longer hold. Precision medicine, rare diseases, and rapidly evolving therapeutic landscapes demand flexibility. Fixed protocols can slow innovation, expose patients to ineffective therapies, and limit learning.

Adaptive trial designs offer a solution. By incorporating accumulating evidence, they allow mid-study adjustments such as:

  • Dropping underperforming arms
  • Modifying randomization ratios
  • Enriching for responsive patient subgroups

Such flexibility accelerates the identification of effective therapies while minimizing exposure to ineffective interventions. Core adaptive elements include response-adaptive randomization, adaptive enrichment, and sample size re-estimation.

Illustrative examples of adaptive platforms:

  • REMAP-CAP – initially designed for community-acquired pneumonia and later expanded for COVID-19, this global platform trial evaluates multiple interventions simultaneously. Its master protocol allows frequent interim analyses and dynamic modification of treatment arms.
  • RECOVERY – a large-scale adaptive trial assessing repurposed COVID-19 treatments, enabling rapid identification of effective therapies while discontinuing ineffective ones.

Despite their advantages, adaptive designs are operationally demanding. Success requires near-real-time data capture, rapid interim analyses, strong governance, and tight coordination across stakeholders. In practice, operational readiness—not statistical methodology—remains the principal constraint.

Real-World Evidence: From Optional to Essential

Real-world evidence (RWE) is reshaping how trials are designed and justified. Data from EHRs, claims, registries, and wearables now show how treatments perform outside controlled settings.

Regulatory adoption has accelerated sharply. Between 2017–2019, only 13% of oncology submissions to the FDA used RWE for efficacy; by 2019–2021, this rose to 70%. This shift is no longer a trend—it is a structural change. Recent FDA guidance (2023–2024) outlines explicit pathways for RWE-supported submissions, including embedded randomized trials, with EMA closely aligned.

For adaptive trials, RWE adds value across the lifecycle:

  • Planning: Clarifies treatment patterns and population characteristics
  • Execution: Identifies responsive subgroups
  • External comparators: Useful when randomized controls aren’t feasible, especially in rare diseases

A recent example illustrates this value: A multinational analysis of 3,003 metastatic castration-resistant prostate cancer patients showed three major treatments had similar efficacy. This RWE informed a 2024 NICE recommendation when traditional head-to-head trials were infeasible.

Yet RWE remains operationally challenging. Data are fragmented, non-standardized, and massive in scale—over 18 zettabytes generated in 2024 alone. Manual integration with live trial data is simply no longer viable.

Agentic AI: The Missing Orchestration Layer

Adaptive trials that integrate RWE often encounter significant coordination challenges. Interim analyses, evolving external evidence, shifting enrollment dynamics, and ongoing safety assessments rarely progress in parallel. Processes that should be simultaneous become sequential, slowing decision-making.

Agentic AI introduces an orchestration layer that enables continuous, parallel execution of key workflows. Dedicated agents manage specific operational functions, including:

  • Automated data surveillance: real-time data quality checks, endpoint tracking, and safety signal detection
  • Evidence intelligence monitoring: continuous scanning of new RWE studies, regulatory guidance, comparator data, and clinical literature
  • Adaptive simulation modeling: scenario analyses to support arm-dropping, enrichment strategies, and sample size adjustments

A centralized orchestration layer can synthesize these outputs and surface decision-relevant insights with reduced manual coordination.

Consider an adaptive platform trial in a neurological indication. As patient data accumulates, teams must conduct interim analyses, monitor safety across all arms, query external RWE for comparative effectiveness, evaluate subgroups, and assess potential protocol amendments—all at the same time.

Without intelligent orchestration, these activities turn into repeated meetings, manual data compilation, and sequential handoffs, causing significant delays. With Agentic AI, these workflows run in parallel, delivering synchronized, decision-ready insights in real time.

According to McKinsey research, agentic approaches could enable  organizations  to design trials 50% faster with 25% fewer amendments, and potentially double site activation rates while requiring 30-50% fewer operational staff. Industry adoption is growing:

  • Pfizer: Reduced coding time by 50% across 100+ studies
  • Novartis: Uses AI simulations for feasibility and adaptive protocol design
  • Causal AI platforms: Enable what-if scenarios and optimization of ongoing studies

Making It Work: What Development Teams Need to Know

Converging adaptive design, RWE, and AI orchestration offers opportunities but requires careful planning:

  • Data infrastructure is critical: AI effectiveness depends on robust pipelines for structured and unstructured data from trials, EHRs, claims, registries, and wearables.
  • Regulatory engagement early: Autonomous AI raises audit and oversight questions. Decision logic must be traceable and explainable.
  • Cross-functional coordination: Biostatistics, data science, clinical operations, IT, medical affairs, and regulatory teams must collaborate.
  • Skills and mindset: Biostatisticians need machine learning literacy; clinical teams must adopt adaptive decision rules; data managers shift to continuous monitoring.

Practical Starting Point:

  • Start small: Pick one high-value trial where adaptive design, RWE, and orchestration deliver clear benefits.
  • Build a strong data foundation: Clean, standardized, accessible data is non-negotiable.
  • Pilot and validate: Run simulations and shadow analyses before live deployment.
  • Maintain human oversight: Agentic AI augments expertise; critical decisions still require human review.

Looking Ahead:

Adaptive designs allow continuous learning, RWE provides real-world context, and Agentic AI makes this operationally feasible. Trials that reach patients fastest in the future won’t necessarily be the largest—they’ll be the most adaptive, most informed by real-world data, and most effectively orchestrated.

Key Takeaways:

  • Adaptive designs enable continuous trial adjustments
  • RWE provides essential context for design, subgroup discovery, and external comparators
  • Agentic AI orchestrates complex integration at scale
  • Strong data infrastructure must precede AI deployment
  • Early regulatory engagement is critical
  • Begin with one high-value use case rather than a broad transformation

While AI-driven orchestration will accelerate evidence synthesis, final decisions must remain grounded in human judgment, statistical rigor, and regulatory alignment.

SHARE

Author

Fill out the form or email info@maxisIT.com to speak with an Expert

[contact-form-7 id="d523db6" title="Contact form 2"]

Subscribe to blog

Related content

Resolve Clinical Development Data Challenges with RBQM

Webinars

Customized FSP Models to Improve Processes and Promote Clinical Outcomes

Case Study

Have you outgrown your Statistical Computing Environment (SCE)?

Videos

Recent Blogs

Curious About What Maxis AI Can Do for You?

Have questions? Looking to scale your R&D with AI agents? Our experts are just a message away.

[contact-form-7 id="0606806" " html_class="submitting"]