Published on 15/11/2025
Managing Data and Analyses When Protocol Deviations Occur
Why Protocol Deviations Are Data Problems—Not Just Process Problems
Every protocol deviation leaves a footprint in the analysis. Missed windows can shift treatment effect estimates, device firmware changes can alter measurement properties, late SAE clocks can bias safety rates, and unblinding can contaminate endpoint assessments. Treating deviations purely as operational issues creates analytical blind spots; treating them purely as statistical issues ignores root causes. The right approach is dual: operational containment to protect participants and endpoints, paired with analysis-aware handling that preserves
Regulatory anchors for data decisions. The quality-by-design philosophy in the ICH clinical practice framework emphasizes proportionate control of critical-to-quality factors and reliable, retrievable records—principles that translate directly into data strategies for incomplete, off-protocol, or corrupted observations. In the United States, expectations on investigator responsibilities, consent, safety, and trustworthy electronic records/signatures are outlined in FDA resources on human subject protection. These values appear operationally in inspection findings: teams must show why certain data were repeated, imputed, excluded, or retained; how the choices align with the protocol and Statistical Analysis Plan (SAP); and where the evidence lives.
Analytical lens: estimands and intercurrent events. Modern trial planning defines the treatment effect via an estimand that specifies population, variable, intercurrent-event strategy, and summary measure. Deviations often are intercurrent events (e.g., prohibited concomitant therapy, rescue medications, missed assessments, permanent discontinuation). Choices such as treatment-policy (include all data regardless of intercurrent events), hypothetical (assume event did not occur), or composite (count event as outcome) must be applied consistently. When a deviation falls outside the pre-specified playbook, document the rationale and consult the study statistician before irreversible actions (e.g., exclusion).
Safety, efficacy, and integrity. The most consequential deviations typically cluster around: (1) consent and reconsent failures (rights and lawful basis); (2) eligibility misadjudication with dosing (population definition); (3) endpoint timing/standardization (measurement validity); (4) unblinding (assessment bias); (5) investigational product accountability/temperature (dose fidelity); and (6) digital capture (eCOA, wearables) and data interfaces (EDC, IRT, safety, imaging). Each category demands a predictable data handling response tied to risk.
Inspection posture. Inspectors expect clean traceability: a deviation record with risk rationale and containment; linked source entries; a short statistics memo stating whether values are repeated, imputed, excluded, or retained; and SAP references or amendments when general rules change. Above all, they expect ALCOA++ across the chain: attributable, legible, contemporaneous, original, accurate—plus complete, consistent, enduring, and available.
Pre-Specify the Playbook: SAP Rules, Data Structures, and Decision Rights
Clarity prevents forced, case-by-case improvisation later. The SAP and data standards should anticipate common deviation modes and define consistent, reviewable rules. Doing so reduces selective exclusions, accelerates close-out, and improves reproducibility.
Core SAP elements for deviation-prone situations.
- Visit windows and timing: Define allowable windows, rescue assessments, and analytical handling when visits occur outside window. For time-sensitive endpoints, specify whether late measures can substitute (with covariate adjustment) or must be missing; for pharmacokinetics, predefine censoring or modeling choices when samples miss nominal times.
- Unblinding and mis-allocations: Describe how assessments performed post-unblinding are handled (e.g., included in ITT safety but excluded from specific blinded efficacy analyses). For randomization errors, state whether analysis follows ITT allocation or actual treatment received for specific endpoints.
- Digital capture: Predefine how off-clock eCOA entries, device dropouts, or unsynchronized clocks are handled; specify rules for firmware or instrument version changes (e.g., retain if equivalence verified, flag/exclude otherwise).
- Intercurrent events: Map common deviations to estimand strategies (treatment-policy/hypothetical/composite); specify imputation or model-based approaches consistent with those strategies.
- Missing data: Distinguish plausibly missing at random (MAR) from not missing at random (MNAR); require sensitivity analyses (pattern-mixture, selection models, or tipping-point) when MNAR is plausible.
- Per-protocol sets: Define objective, pre-specified criteria (e.g., minimum dosing exposure, visit compliance, no major violations affecting endpoint validity) and ensure these criteria are measurable in data—avoiding subjective, post hoc choices.
Data structures that make rules executable. Prepare analysis flags and variables early so rules run automatically:
- Window flags: on-time/late/early; deviation distance in hours/days.
- Assessment validity flags: correct instrument/firmware version; calibration status; standardized conditions met (fasting, posture, equipment lot).
- Blinding flags: any unblinding before assessment; emergency unblinding details.
- Exposure adherence metrics: dosing dates, interruptions, temperature excursions.
- Intercurrent event variables: start/stop dates for rescue therapy, discontinuation, or prohibited concomitants.
Decision rights and documentation. Establish who can make which calls: the statistician decides analytical inclusion/exclusion and imputation; the PI adjudicates causality/eligibility; data management enforces structural flags; QA checks traceability. In the EU and UK, align operational practice with EMA clinical trial guidance on reliability and robustness of data; ethics considerations should reflect WHO research ethics guidance when decisions affect participant rights (e.g., reconsent before continued data use).
Change control. When patterns force SAP updates (e.g., widespread courier delays shifting visit timing), use versioned amendments and cross-reference the data model changes. Retrospective rule changes require transparent rationale and a sensitivity analysis contrasting old and new rules.
Operationalizing Data Integrity: Provenance, Systems, and Decentralized Realities
Data handling decisions are credible only when the underlying systems can prove who did what, when, and with what configuration. This is where source documentation, audit trails, and system validation meet statistics.
Provenance and auditability. Ensure signature manifestation (printed name, date/time with time zone, and meaning of signature) for consent, adjudications, endpoint confirmations, and analysis decisions. Enforce immutable audit trails and time synchronization across EDC, eCOA, IRT, imaging, and safety systems. The record should allow a reviewer to reconstruct the chain without guessing—consistent with the spirit of Part 11/Annex 11 concepts often examined during inspections.
Digital capture and measurement integrity. For wearables or app-based eCOA, measurement properties can change with firmware. Require vendor release notes, validation checks (bench and, when needed, clinical equivalence), and version variables in the dataset. If equivalence fails or is unknown, classify affected data as “non-comparable” and follow the SAP rule (exclude from primary, include in sensitivity). For home health procedures, document identities, training, and conditions; file photos of kit labels and temperature loggers when applicable.
Reconciliation among systems. Maintain “connection control packs” describing each interface: source/target, frequency, error handling, and owners. Reconcile safety cases with AE pages, IRT dosing with exposure variables, imaging timestamps with visit windows, and eCOA alerts with recorded AEs. Unreconciled mismatches create silent bias; track them as KRIs and escalate early.
Unblinding controls. Document exactly who became unblinded and when. Assess whether assessments after unblinding can remain in ITT analyses or require exclusion/flagging for specific endpoints. If emergency unblinding safeguards failed, record a CAPA and consider an independent reassessment where feasible.
Data review meetings with purpose. Replace perfunctory listings reviews with focused “risk-to-estimand” sessions: which deviations threaten the estimand’s assumptions; which require reconsent; which need rescue assessments; which require SAP updates. Capture decisions as short memos linked to subject records and tables, so the story reads linearly months later.
Regional expectations. When operating in Japan and Australia, align documentation and submission styles to the expectations of PMDA clinical guidance and TGA clinical trial guidance. The underlying principles—participant protection, reliable data, and transparent traceability—are consistent across regions, but the forms and channels differ.
Decentralized (DCT) specifics. Identity and privacy checks during tele-consent or remote assessments must be documented in source; avoid unapproved channels for data transfer. If bandwidth disrupts scheduled assessments, define rescue windows in the SAP. For direct-to-patient shipments, record chain-of-custody and temperature data in analysis-ready structures so affected exposure can be isolated or adjusted.
From Deviation to Table, Figure, and Listing: Statistical Handling, Sensitivities, and Reporting
Ultimately, choices about what to include, exclude, repeat, or impute must flow through to reproducible code, transparent listings, and a narrative that regulators and reviewers can follow in minutes.
Primary analysis sets and robustness. Prespecify ITT, safety, and per-protocol populations with objective criteria; apply them mechanically via flags. When major deviations affect endpoint validity, define a “modified per-protocol” or “validity subset” in advance, then use sensitivity analyses to test whether conclusions depend on those choices.
Missing data approaches. Use model-based analyses aligned to estimands where MAR is reasonable; when MNAR is plausible (e.g., symptom-related dropout), run pattern-mixture or selection-model sensitivities. Tipping-point analyses should vary assumptions until conclusions reverse, with plots that make the inflection obvious. For binary endpoints, consider delta-adjusted imputation; for time-to-event, use competing-risk or treatment-policy strategies as planned.
Timing deviations and censoring. For time-sensitive endpoints, specify whether late assessments count as missing, are adjusted by covariates, or are analyzed in a joint model. For survival analyses, document whether censoring at protocol deviation is appropriate or whether treatment-policy inclusion better reflects clinical reality. Avoid informative censoring that removes worse outcomes disproportionately.
Measurement comparability. If instruments or firmware change, include version-by-treatment interaction checks. When equivalence is confirmed, pool; otherwise, stratify or exclude affected measures from primary while retaining them in supportive analyses. Document equivalence testing method, thresholds, and results in the data-handling memo.
Listings and traceability. Produce review sets that line up deviation records, source references, and analysis flags next to subject-level outcomes. Every exclusion or special handling should be machine-traceable to a row in the deviation log and a rule in the SAP. This is the fastest way to answer, “Why is this value missing or excluded?” during inspection.
Interim analyses and DMC/DSMB. When deviations threaten safety or endpoint reliability, brief the monitoring committee with the data-handling proposal, including operational CAPA. Keep the committee’s recommendations in the TMF and reflect material changes in the SAP amendment.
Quality tolerance limits (QTLs) and KRIs. Link analytical risks to study-level QTLs (e.g., “primary endpoint window misses <1% of randomized participants”) and site-level KRIs (eCOA missingness spikes, repeated firmware changes, unblinding incidents). Crossing a QTL triggers a cross-functional review and may change the analysis or monitoring intensity.
Clinical study report (CSR) narrative. Summarize how deviations were detected, classified, and handled; state how many values were repeated, imputed, excluded, or retained; and provide concise sensitivity results. Reference where detailed memos, flags, and code live. Avoid jargon; write so a scientifically literate reader can follow the logic.
Ethics and transparency. Decisions about excluding data collected without valid consent, or about continuing participation post-deviation, should reflect participant rights and dignity. Discuss how ethics input informed data choices, consistent with international guidance and national expectations. Clear documentation builds trust with reviewers and participants alike.
Practical checklist for the next deviation that touches data
- Confirm containment and create a deviation record with awareness time, risk assessment, and links to source and systems.
- Consult the SAP: identify the pre-specified rule; if none, draft a short data-handling memo and obtain statistician and PI sign-off.
- Set or update analysis flags; ensure traceability from deviation record to dataset variables.
- Decide repeat/impute/exclude/retain; plan and run relevant sensitivity analyses (e.g., tipping point, hypothetical estimand).
- Update listings; rehearse the inspection story (facts → rule → action → evidence → impact on results).
- If recurring, amend SAP and training; add KRIs/QTLs; confirm system changes (e.g., firmware control, scheduling buffers).
Bottom line. Data handling for deviations is where quality systems and statistical practice meet. When rules are pre-specified, systems provide provenance, and analyses include targeted sensitivities, teams can show a coherent, regulator-ready story that protects participants and preserves credible inference—across the USA, EU/UK, Japan, Australia, and other ICH regions.