Published on 16/11/2025
MedDRA Coding and Signal Detection: Precise Terminology, Reliable Analytics, and Audit-Ready Controls
Make the Dictionary Work for You: MedDRA Structure, Governance, and Fit-for-Purpose Use
MedDRA is more than a vocabulary; it is the analytic backbone of pharmacovigilance. Accurate coding determines whether case data roll up correctly, whether Standardised MedDRA Queries (SMQs) retrieve what they should, and whether signals surface—or remain hidden. Regulators across the U.S. FDA, the EMA, Japan’s PMDA, Australia’s TGA, and the harmonization framework of
Know the hierarchy. MedDRA is five-level and multi-axial: Lowest Level Terms (LLTs) roll to Preferred Terms (PTs), which map to High Level Terms (HLTs), then High Level Group Terms (HLGTs), and System Organ Classes (SOCs). Multi-axiality means a PT can belong to multiple SOCs (one “primary” SOC for consistent placement and other “secondary” SOCs for retrieval). Choosing the correct PT (and not just any similar LLT) is the single most important determinant of future analytic quality. For example, “Stevens-Johnson syndrome” should not be coded as “Rash” (loss of specificity) nor as “Toxic epidermal necrolysis” (over-specificity) unless the diagnosis supports it.
SMQs are retrieval tools—treat them as such. SMQs combine PTs/LLTs to support topic-based searching (e.g., anaphylactic reactions, hepatic disorders). Most offer narrow (high specificity) and broad (higher sensitivity) scopes. In clinical development, narrow SMQs reduce false positives for expedited work; for aggregate/signal screening, using both scopes (with transparent counting rules) can be appropriate. Document which scope you use and why—consistency enables trend interpretation and cross-study comparability.
Governance and versions. MedDRA typically updates biannually. Your Safety Management Plan and system configuration should specify: (1) who decides when to upgrade; (2) which cases are migrated vs left in historical versions; (3) how to manage SMQ changes; and (4) how to communicate impacts on trending and signals. Maintain controlled change logs, version tags in the safety database, and a migration validation report. Many inspection findings arise from silent dictionary drift between coding and analysis environments.
Customization without chaos. Company-specific coding conventions (e.g., “Infusion related reaction” vs “Hypersensitivity”) are useful—so long as they are documented and tied to MedDRA intent. Create quick-reference guides for high-impact domains (cardiac, hepatic, hypersensitivity, pregnancy, product quality). Where product-specific Adverse Events of Special Interest (AESIs) exist, define the PT lists and SMQs used for each AESI and keep them under version control alongside your risk management documents.
Combination products and devices. When a device contributes to harm, capture device identifiers (model/lot/serial, software version) and use device problem codes in parallel to MedDRA patient-impact coding. Ensure vigilance obligations for devices align with medicinal product rules in the target region (e.g., EU MDR vigilance alongside EudraVigilance reporting).
Concomitant medication coding matters. Code concomitant drugs with WHO Drug Dictionary (WHO-DD) to support evaluation of interactions and alternative etiologies. Many causal assessments fail because interacting drugs or contraindications are not codified, making automated review unreliable.
Coding Execution Without Regret: Processes, QC, and Migration You Can Defend
From verbatim to PT—the golden path. Good practice routes each verbatim through: (1) medical interpretation (ensure you’re coding the diagnosis when present; otherwise code the most specific sign/symptom); (2) spelling standardization (avoid coding typos); (3) selection of the most specific, defensible PT; (4) LLT choice that preserves mapping; and (5) peer or QC review for high-impact events (death, IMEs, SUSAR candidates). Coders should have rapid access to source narratives and lab data; coding in a vacuum is a root cause of misclassification.
“Seriousness-sensitive” coding control. Create watchlists for terms where a PT switch can flip seriousness recognition (e.g., “Gastrointestinal hemorrhage” vs “Rectal bleeding”). Require secondary review when a coder downgrades a term from an IME or removes a term included in an AESI definition. Record rationale in the case file.
Encoding etiologies and patterns. For syndromes or pathophysiologic clusters (e.g., DILI), code the diagnosis PT when appropriate, but retain key component PTs (e.g., “Alanine aminotransferase increased,” “Bilirubin increased”) so laboratory-driven SMQs fire and quantitative thresholds (Hy’s Law) can be assessed programmatically. For anaphylaxis, code both the syndrome and sentinel features (hypotension, bronchospasm) when documented.
Queries and retrieval rules. Standardize how you pull cases: SMQ narrow vs broad, PT lists for AESIs, date windows (onset vs receipt), and inclusion/exclusion (e.g., exclude study endpoints). Spell out Boolean logic for combined queries and validate them against known positive controls (seed cases) so that both recall and precision are acceptable.
Quality control metrics that matter.
- Coding concordance between primary and QC coders for IMEs/AESIs (target ≥95%).
- Turnaround time from case creation to coded state (median hours; expedite for SUSAR candidates).
- Rework rate (percentage of code changes post-QC or post-signal review).
- Version alignment (percentage of cases and analytics running on the same MedDRA/SMQ versions).
- Drift monitors (alerts if frequency of closely related PTs changes abruptly—often a sign of coder behavior change).
Migration without losing history. When upgrading MedDRA, run a dry-run migration on a representative subset. Quantify impacts: % of PT remaps, AESI/SMQ retrieval changes, signal metric shifts. Retain pre- and post-migration snapshots so you can explain trend breaks to inspectors and in Periodic Reports (DSUR/PBRER). Document how historical cases are treated in aggregate analyses (e.g., re-coded vs analyzed as-is) and label figures accordingly.
Training and coders’ toolkit. Train coders on difficult differentials (e.g., “Syncope” vs “Presyncope”; “Myocardial infarction” vs “Myocardial injury”), on vaccine-specific patterns (e.g., myocarditis/pericarditis), and on device terms. Provide curated decision trees and a searchable coding convention wiki. Gate coding privileges to validated training completion and periodic re-qualification.
Alignment with expectedness and RSI. Because expectedness is evaluated against the Reference Safety Information (IB/label), ensure the coded terms used in RSI tables match the PT granularity used in cases. If RSI lists “Transaminases increased (mild to moderate),” a case coded and supported as “Hepatic failure” will likely be unexpected—make this explicit in the case review to support correct expedited routing.
Signals that Mean Something: Data Sources, Disproportionality, and Medical Context
Signal management transforms coded data into actionable pharmacovigilance intelligence. GVP concepts (EU Module IX) map well to global expectations: signal detection → validation → analysis and prioritization → assessment → recommendation → communication. While algorithms help, clinical judgement is decisive; regulators (e.g., FDA, EMA, PMDA, TGA) expect transparent methods and traceable decisions.
Where signals come from. Internally: clinical trial safety databases (Argus/ARISg), EDC reconcilable SAEs, eCOA adverse symptoms, product quality complaints with AEs, literature, and partner feeds per SDEAs. Externally: FAERS (U.S.), EudraVigilance (EU), VigiBase (WHO global), disease registries, and published observational studies. For early development, internal blinded data are often insufficient for detection; use pooled class data and mechanism-of-action hypotheses to define AESIs and background expectations.
Disproportionality in a nutshell. With spontaneous reporting systems, a product–event pair is compared to other products and events to detect disproportional reporting:
- PRR (Proportional Reporting Ratio) and ROR (Reporting Odds Ratio): simple, transparent, widely used.
- EBGM/EB05 (Empirical Bayes Geometric Mean): shrinks noisy estimates, robust for low counts.
- IC/IC025 (Information Component): Bayesian measure used by WHO-UMC’s BCPNN.
Choose methods appropriate to the data source. Define thresholds, but never treat them as proof. Elevated disproportionality may reflect notoriety bias, stimulated reporting, co-prescription confounding, or duplicate cases. Conversely, real risks may be masked when usage is high or background rates are elevated. Always bring in exposure (denominators), severity, biologic plausibility, and time-to-onset.
For clinical trials—different playbook. Disproportionality is rarely appropriate in blinded development datasets. Instead, rely on pre-specified AESIs, SMQs, and medical review of patterns across dose, time, and risk factors. Use exposure-adjusted incidence rates (EAIR), severity grading (e.g., CTCAE), and adjudication results where available. Unblinded comparative analyses belong in the DSMB/independent statistician lane with strict access control.
Triage and validation. Build weekly/monthly triage meetings where safety scientists screen algorithms, literature hits, and case clusters. A validated signal should have: (1) a coherent case series (quality, temporality, dechallenge/rechallenge); (2) specificity of PTs/SMQs; (3) plausible mechanism; (4) consistency across sources; and (5) consideration of background rates. Document criteria and rationales; keep a living “signal tracker” with timestamps and outcomes.
Analysis and prioritization. Rate signals on impact (seriousness, reversibility, frequency), preventability (labeling/risk minimization potential), and confidence. Use templated workups: case line listings, narratives, time-to-onset plots, dose/exposure relationships, subgroup effects (age/sex/comorbidity), and alternative explanations. Summarize proposed actions (label change, protocol amendment, additional monitoring, targeted follow-up forms, post-marketing study) for governance.
Vaccines and background rates. For high-incidence post-immunization events (e.g., fever, headache) and rare AESIs (e.g., myocarditis, ADEM), anchor interpretation to age/sex-specific background rates and risk windows. The WHO provides immunization safety frameworks; align with national guidance. Use case definition criteria (e.g., Brighton Collaboration) and fast-track medical review pathways.
Communicate and close the loop. Record decisions, timelines, and responsibilities. Update the Investigator’s Brochure (RSI), labels, Risk Management Plan (RMP) or U.S. REMS as needed; inform investigators/IRBs/IECs promptly when participant protection requires it. Reflect significant signals and actions in DSURs/PBRERs with traceability to the underlying cases and analytics.
Inspection-Grade Proof: Evidence Bundle, KPIs, Pitfalls, and a Ready-to-Use Checklist
Rapid-pull evidence for inspectors. Be ready to surface within minutes: (1) dictionary governance SOPs and coding conventions; (2) MedDRA/SMQ and WHO-DD version histories with effective dates; (3) migration validation reports and before/after analytics; (4) AESI definitions and retrieval queries (SMQ scopes, PT lists, Boolean logic) with change logs; (5) coding QC plans and metrics; (6) signal management SOPs, triage minutes, trackers, and completed assessment templates; (7) examples of signal workups leading to labeling/RMP/REMS actions; and (8) role/access controls and training records for coders and safety scientists. Authorities at the FDA, EMA, PMDA, and TGA will look for coherence with ICH expectations and the WHO public-health perspective.
Program-level KPIs.
- Coding concordance for IMEs/AESIs (primary vs QC; target ≥95%).
- On-time coding for expedited-eligible cases (median hours).
- Version alignment (% of cases and analytics on the same MedDRA/SMQ and WHO-DD versions).
- Retrieval precision/recall for AESIs/SMQs validated against seed cases.
- Signal cycle time (detection → validation → assessment → decision).
- Action effectiveness (post-labeling trend changes; adherence to additional risk minimization).
- Duplicate rate in external databases (detected/resolved per quarter).
Common pitfalls—and durable fixes.
- Mismatched granularity (coding generic terms when specific diagnoses exist). → Train on diagnosis-over-symptom rule; require narrative access; QC high-impact domains.
- Dictionary drift between coding and analysis layers. → Centralize version control; lock analytics and re-run change-impact checks after upgrades.
- Untested AESI queries. → Validate on seed cases; monitor precision/recall; keep a change log with governance approval.
- Over-reliance on disproportionality without medical context. → Pair algorithms with case series review, exposure metrics, and biologic plausibility.
- Duplicate cases inflating signals. → Strengthen de-duplication rules and tooling; cross-reference literature and partner feeds; document merges with audit trails.
- Inconsistent RSI/expectedness mapping. → Align PT granularity in RSI tables with case coding; display RSI version in case headers; audit decisions.
- Blind leakage via operational dashboards. → Keep development dashboards arm-agnostic; segregate unblinded comparative analyses to independent statisticians/DSMB lanes.
Study-ready checklist (one page).
- Coding conventions issued and trained; access to narratives/labs for accurate PT selection.
- MedDRA/SMQ and WHO-DD versions documented; migration plan validated with before/after analytics.
- Seriousness-sensitive watchlist active; secondary review required for IMEs/AESI-linked PT changes.
- AESI definitions published with retrieval logic (SMQ scopes/PT lists); precision/recall validated.
- SMQ usage documented (narrow/broad) and consistent by purpose (expedited vs aggregate).
- Signal triage cadence set; tracker active with validation criteria, assessments, and decisions.
- Disproportionality methods and thresholds defined for external databases; confounding and duplicates addressed.
- Clinical trial oversight uses EAIR, adjudication, and medical review; unblinded analyses restricted per charter.
- KPI dashboard in place (concordance, timeliness, version alignment, signal cycle time, action effectiveness).
- Inspection pack prepared: SOPs, change logs, migration reports, QC outputs, signal dossiers, and training records.
Bottom line. Great safety decisions start with great terminology. When MedDRA coding is specific, governed, and validated—and when signal detection marries algorithms with clinical judgement and exposure context—your program can find true risks early, respond proportionately, and demonstrate control to assessors at the FDA, EMA, PMDA, and TGA, consistent with ICH principles and the WHO commitment to public health.