Statistics Made Simple: Sensitivity, Specificity & Power (MRCP Part 1)
- Crack Medicine

- Jan 30
- 4 min read
TL;DR:
Sensitivity, specificity, predictive values, and statistical power are repeatedly tested in MRCP Part 1, usually through interpretation rather than calculation. This article explains what each term means in practice, how examiners frame questions, and how to avoid common traps. A short MCQ-style case and a practical revision checklist are included to support efficient exam preparation.
Why this topic matters for MRCP Part 1
Statistics is one of the most consistently examined areas in MRCP Part 1, yet it remains a common source of avoidable marks. The exam does not test advanced mathematics. Instead, it assesses whether candidates can interpret diagnostic tests and research findings in a clinically meaningful way.
Questions on sensitivity, specificity, predictive values, and power are designed to look simple while probing conceptual understanding. Candidates who rely only on memorised formulas often struggle when prevalence changes, when a screening context is implied, or when study validity is questioned.
This article supports the core statistics syllabus within the MRCP Part 1 overview and complements applied practice available in the MRCP question bank and mock tests.
Scope of statistics tested in MRCP Part 1
Within the statistics domain, the exam focuses on applied clinical epidemiology. The most commonly tested areas include:
Diagnostic test performance
Screening versus diagnostic testing
Interpretation of study results
Basic clinical trial design
Bias, error, and limitations of evidence
Among these, sensitivity, specificity, predictive values, and power are among the highest-yield topics and appear repeatedly across exam diets.

Core definitions you must know
Sensitivity
Sensitivity is the probability that a test is positive when the disease is truly present.
Formula: True positives / (True positives + False negatives)
A highly sensitive test produces very few false negatives
Exam principle: A highly sensitive test is useful for ruling out disease when the result is negative (SnNout).
Specificity
Specificity is the probability that a test is negative when the disease is truly absent.
Formula: True negatives / (True negatives + False positives)
A highly specific test produces very few false positives
Exam principle: A highly specific test is useful for ruling in disease when the result is positive (SpPin).
Positive predictive value (PPV)
PPV is the probability that a patient actually has the disease given a positive test result.
Formula: True positives / (True positives + False positives)
Strongly influenced by disease prevalence
Negative predictive value (NPV)
NPV is the probability that a patient does not have the disease given a negative test result.
Formula: True negatives / (True negatives + False negatives)
Also influenced by disease prevalence
Statistical power
Power is the probability that a study will detect a true effect if one genuinely exists.
Power = 1 − β (beta error)
Commonly set at 80–90% in clinical research
Exam principle: Low power increases the risk of a false-negative study.
One table to consolidate understanding
Measure | Depends on prevalence? | Typical MRCP use |
Sensitivity | No | Screening tests, ruling out disease |
Specificity | No | Confirmatory tests, ruling in disease |
PPV | Yes | Clinical interpretation of positive results |
NPV | Yes | Reassurance after a negative result |
Power | No | Assessing validity of trial conclusions |
The 5 most tested subtopics
1. Screening versus diagnostic tests
Screening tests prioritise high sensitivity to minimise missed cases. Diagnostic confirmation prioritises high specificity to avoid false positives. MRCP questions often imply this distinction without stating it explicitly.
2. Effect of prevalence on predictive values
As prevalence increases:
PPV increases
NPV decreases
As prevalence falls:
PPV decreases
NPV increases
Sensitivity and specificity remain unchanged. This contrast is a frequent exam focus.
3. False positives and false negatives
Examiners often ask which error is more clinically significant in a given scenario. False negatives are particularly serious in life-threatening but treatable conditions, while false positives may cause harm through unnecessary investigation or treatment.
4. Power and sample size
Power increases with:
Larger sample size
Larger effect size
Lower data variability
Questions commonly describe a negative trial and ask whether inadequate power could explain the findings.
5. Confidence intervals versus p values
Confidence intervals describe precision, not just significance. A wide interval suggests uncertainty, even if the p value is significant. If the confidence interval crosses the null value, the result is not statistically significant.
Mini MCQ-style case
Question: A screening test for Disease X has a sensitivity of 98% and a specificity of 70%. The disease prevalence in the screened population is 0.5%. Which statement is most accurate?
Answer: The positive predictive value will be low.
Explanation: Despite excellent sensitivity, the very low prevalence means most positive results are false positives. This illustrates why PPV depends heavily on prevalence, a classic MRCP Part 1 testing point. Similar interpretation-based questions are common in the MRCP question bank.
Common pitfalls (high-yield)
Confusing sensitivity with positive predictive value
Forgetting that predictive values change with prevalence
Assuming power relates to statistical significance rather than study design
Overinterpreting a non-significant result without considering power
Ignoring clinical consequences of false results in question stems
Practical revision checklist
Memorise definitions with clinical context
Always ask whether prevalence is relevant
Link sensitivity to screening and specificity to confirmation
When power is mentioned, think false negatives
Focus on interpretation rather than calculations
Test understanding under exam conditions using a mock test
FAQs
What is the difference between sensitivity and specificity?
Sensitivity measures detection of disease when it is present, while specificity measures correct exclusion when disease is absent. Both are intrinsic test properties.
Why does prevalence affect PPV but not sensitivity?
PPV reflects real-world probability after a test result and depends on how common the disease is. Sensitivity is calculated only within diseased individuals.
Is power the same as statistical significance?
No. Power relates to a study’s ability to detect a true effect, while significance describes whether an observed result is unlikely due to chance.
How is power increased in a clinical trial?
Primarily by increasing sample size, but also by increasing effect size or reducing variability.
Ready to start?
For systematic coverage of statistics and other core topics, return to the MRCP Part 1 overview hub. Consolidate understanding with targeted practice using the Free MRCP MCQs and assess readiness with a Start a mock test under exam conditions.
Sources
MRCP(UK). Examination Regulations and Blueprint. https://www.mrcpuk.org
Altman DG. Practical Statistics for Medical Research. Chapman & Hall/CRC.
BMJ. Statistics Notes Series. https://www.bmj.com/specialties/statistics-notes



Comments