In medicine and statistics, sensitivity and specificity mathematically describe the accuracy of a test that reports the presence or absence of a medical condition. If individuals who have the condition are considered "positive" and those who do not are considered "negative", then sensitivity is a measure of how well a test can identify true positives and specificity is a measure of how well a test can identify true negatives:
If the true status of the condition cannot be known, sensitivity and specificity can be defined relative to a "gold standard test" which is assumed correct. For all testing, both diagnoses and screening, there is usually a tradeoff between sensitivity and specificity, such that higher sensitivities will mean lower specificities and vice versa.
A test which reliably detects the presence of a condition, resulting in a high number of true positives and low number of false negatives, will have a high sensitivity. This is especially important when the consequence of failing to treat the condition is serious and/or the treatment is very effective and has minimal side effects.
A test which reliably excludes individuals who do not have the condition, resulting in a high number of true negatives and low number of false positives, will have a high specificity. This is especially important when people who are identified as having a condition may be subjected to more testing, expense, stigma, anxiety, etc.
The terms "sensitivity" and "specificity" were introduced by American biostatistician Jacob Yerushalmy in 1947.^{[1]}
There are different definitions within laboratory quality control, wherein "analytical sensitivity" is defined as the smallest amount of substance in a sample that can accurately be measured by an assay (synonymously to detection limit), and "analytical specificity" is defined as the ability of an assay to measure one particular organism or substance, rather than others.^{[2]} However, this article deals with diagnostic sensitivity and specificity as defined at top.
Imagine a study evaluating a test that screens people for a disease. Each person taking the test either has or does not have the disease. The test outcome can be positive (classifying the person as having the disease) or negative (classifying the person as not having the disease). The test results for each subject may or may not match the subject's actual status. In that setting:
After getting the numbers of true positives, false positives, true negatives, and false negatives, the sensitivity and specificity for the test can be calculated. If it turns out that the sensitivity is high then any person who has the disease is likely to be classified as positive by the test. On the other hand, if the specificity is high, any person who does not have the disease is likely to be classified as negative by the test. An NIH web site has a discussion of how these ratios are calculated.^{[3]}
Consider the example of a medical test for diagnosing a condition. Sensitivity (sometimes also named the detection rate in a clinical setting) refers to the test's ability to correctly detect ill patients out of those who do have the condition.^{[4]} Mathematically, this can be expressed as:
A negative result in a test with high sensitivity can be useful for "ruling out" disease,^{[4]} since it rarely misdiagnoses those who do have the disease. A test with 100% sensitivity will recognize all patients with the disease by testing positive. In this case, a negative test result would definitively rule out the presence of the disease in a patient. However, a positive result in a test with high sensitivity is not necessarily useful for "ruling in" disease. Suppose a 'bogus' test kit is designed to always give a positive reading. When used on diseased patients, all patients test positive, giving the test 100% sensitivity. However, sensitivity does not take into account false positives. The bogus test also returns positive on all healthy patients, giving it a false positive rate of 100%, rendering it useless for detecting or "ruling in" the disease.^{[citation needed]}
The calculation of sensitivity does not take into account indeterminate test results. If a test cannot be repeated, indeterminate samples either should be excluded from the analysis (the number of exclusions should be stated when quoting sensitivity) or can be treated as false negatives (which gives the worstcase value for sensitivity and may therefore underestimate it).^{[citation needed]}
A test with a higher sensitivity has a lower type II error rate.
Consider the example of a medical test for diagnosing a disease. Specificity refers to the test's ability to correctly reject healthy patients without a condition. Mathematically, this can be written as:
A positive result in a test with high specificity can be useful for "ruling in" disease, since the test rarely gives positive results in healthy patients.^{[5]} A test with 100% specificity will recognize all patients without the disease by testing negative, so a positive test result would definitively rule in the presence of the disease. However, a negative result from a test with high specificity is not necessarily useful for "ruling out" disease. For example, a test that always returns a negative test result will have a specificity of 100% because specificity does not consider false negatives. A test like that would return negative for patients with the disease, making it useless for "ruling out" the disease.
A test with a higher specificity has a lower type I error rate.
The above graphical illustration is meant to show the relationship between sensitivity and specificity. The black, dotted line in the center of the graph is where the sensitivity and specificity are the same. As one moves to the left of the black dotted line, the sensitivity increases, reaching its maximum value of 100% at line A, and the specificity decreases. The sensitivity at line A is 100% because at that point there are zero false negatives, meaning that all the negative test results are true negatives. When moving to the right, the opposite applies, the specificity increases until it reaches the B line and becomes 100% and the sensitivity decreases. The specificity at line B is 100% because the number of false positives is zero at that line, meaning all the positive test results are true positives.
The middle solid line in both figures that show the level of sensitivity and specificity is the test cutoff point. As previously described, moving this line results in a tradeoff between the level of sensitivity and specificity. The lefthand side of this line contains the data points that tests below the cut off point and are considered negative (the blue dots indicate the False Negatives (FN), the white dots True Negatives (TN)). The righthand side of the line shows the data points that tests above the cut off point and are considered positive (red dots indicate False Positives (FP)). Each side contains 40 data points.
For the figure that shows high sensitivity and low specificity, there are 3 FN and 8 FP. Using the fact that positive results = true positives (TP) + FP, we get TP = positive results  FP, or TP = 40  8 = 32. The number of sick people in the data set is equal to TP + FN, or 32 + 3 = 35. The sensitivity is therefore 32 / 35 = 91.4%. Using the same method, we get TN = 40  3 = 37, and the number of healthy people 37 + 8 = 45, which results in a specificity of 37 / 45 = 82.2 %.
For the figure that shows low sensitivity and high specificity, there are 8 FN and 3 FP. Using the same method as the previous figure, we get TP = 40  3 = 37. The number of sick people is 37 + 8 = 45, which gives a sensitivity of 37 / 45 = 82.2 %. There are 40  8 = 32 TN. The specificity therefore comes out to 32 / 35 = 91.4%.
The red dot indicates the patient with the medical condition. The red background indicates the area where the test predicts the data point to be positive. The true positive in this figure is 6, and false negatives of 0 (because all positive condition is correctly predicted as positive). Therefore, the sensitivity is 100% (from 6 / (6 + 0)). This situation is also illustrated in the previous figure where the dotted line is at position A (the lefthand side is predicted as negative by the model, the righthand side is predicted as positive by the model). When the dotted line, test cutoff line, is at position A, the test correctly predicts all the population of the true positive class, but it will fail to correctly identify the data point from the true negative class.
Similar to the previously explained figure, the red dot indicates the patient with the medical condition. However, in this case, the green background indicates that the test predicts that all patients are free of the medical condition. The number of data point that is true negative is then 26, and the number of false positives is 0. This result in 100% specificity (from 26 / (26 + 0)). Therefore, sensitivity or specificity alone cannot be used to measure the performance of the test.
In medical diagnosis, test sensitivity is the ability of a test to correctly identify those with the disease (true positive rate), whereas test specificity is the ability of the test to correctly identify those without the disease (true negative rate). If 100 patients known to have a disease were tested, and 43 test positive, then the test has 43% sensitivity. If 100 with no disease are tested and 96 return a completely negative result, then the test has 96% specificity. Sensitivity and specificity are prevalenceindependent test characteristics, as their values are intrinsic to the test and do not depend on the disease prevalence in the population of interest.^{[6]} Positive and negative predictive values, but not sensitivity or specificity, are values influenced by the prevalence of disease in the population that is being tested. These concepts are illustrated graphically in this applet Bayesian clinical diagnostic model which show the positive and negative predictive values as a function of the prevalence, sensitivity and specificity.
It is often claimed that a highly specific test is effective at ruling in a disease when positive, while a highly sensitive test is deemed effective at ruling out a disease when negative.^{[7]}^{[8]} This has led to the widely used mnemonics SPPIN and SNNOUT, according to which a highly specific test, when positive, rules in disease (SPPIN), and a highly sensitive test, when negative, rules out disease (SNNOUT). Both rules of thumb are, however, inferentially misleading, as the diagnostic power of any test is determined by the prevalence of the condition being tested, the test's sensitivity and its specificity.^{[9]}^{[10]}^{[11]} The SNNOUT mnemonic has some validity when the prevalence of the condition in question is extremely low in the tested sample.
The tradeoff between specificity and sensitivity is explored in ROC analysis as a trade off between TPR and FPR (that is, recall and fallout).^{[12]} Giving them equal weight optimizes informedness = specificity + sensitivity − 1 = TPR − FPR, the magnitude of which gives the probability of an informed decision between the two classes (> 0 represents appropriate use of information, 0 represents chancelevel performance, < 0 represents perverse use of information).^{[13]}
The sensitivity index or d′ (pronounced "deeprime") is a statistic used in signal detection theory. It provides the separation between the means of the signal and the noise distributions, compared against the standard deviation of the noise distribution. For normally distributed signal and noise with mean and standard deviations and , and and , respectively, d′ is defined as:
An estimate of d′ can be also found from measurements of the hit rate and falsealarm rate. It is calculated as:
where function Z(p), p ∈ [0, 1], is the inverse of the cumulative Gaussian distribution.
d′ is a dimensionless statistic. A higher d′ indicates that the signal can be more readily detected.
Main article: Confusion matrix 
The relationship between sensitivity, specificity, and similar terms can be understood using the following table. Consider a group with P positive instances and N negative instances of some condition. The four outcomes can be formulated in a 2×2 contingency table or confusion matrix, as well as derivations of several metrics using the four outcomes, as follows:
Predicted condition  ^{Sources: }^{[16]}^{[17]} ^{[18]}^{[19]}^{[20]}^{[21]}^{[22]}^{[23]} ^{.mwparseroutput .hlist dl,.mwparseroutput .hlist ol,.mwparseroutput .hlist ul{margin:0;padding:0}.mwparseroutput .hlist dd,.mwparseroutput .hlist dt,.mwparseroutput .hlist li{margin:0;display:inline}.mwparseroutput .hlist.inline,.mwparseroutput .hlist.inline dl,.mwparseroutput .hlist.inline ol,.mwparseroutput .hlist.inline ul,.mwparseroutput .hlist dl dl,.mwparseroutput .hlist dl ol,.mwparseroutput .hlist dl ul,.mwparseroutput .hlist ol dl,.mwparseroutput .hlist ol ol,.mwparseroutput .hlist ol ul,.mwparseroutput .hlist ul dl,.mwparseroutput .hlist ul ol,.mwparseroutput .hlist ul ul{display:inline}.mwparseroutput .hlist .mwemptyli{display:none}.mwparseroutput .hlist dt::after{content:": "}.mwparseroutput .hlist dd::after,.mwparseroutput .hlist li::after{content:" · ";fontweight:bold}.mwparseroutput .hlist dd:lastchild::after,.mwparseroutput .hlist dt:lastchild::after,.mwparseroutput .hlist li:lastchild::after{content:none}.mwparseroutput .hlist dd dd:firstchild::before,.mwparseroutput .hlist dd dt:firstchild::before,.mwparseroutput .hlist dd li:firstchild::before,.mwparseroutput .hlist dt dd:firstchild::before,.mwparseroutput .hlist dt dt:firstchild::before,.mwparseroutput .hlist dt li:firstchild::before,.mwparseroutput .hlist li dd:firstchild::before,.mwparseroutput .hlist li dt:firstchild::before,.mwparseroutput .hlist li li:firstchild::before{content:" (";fontweight:normal}.mwparseroutput .hlist dd dd:lastchild::after,.mwparseroutput .hlist dd dt:lastchild::after,.mwparseroutput .hlist dd li:lastchild::after,.mwparseroutput .hlist dt dd:lastchild::after,.mwparseroutput .hlist dt dt:lastchild::after,.mwparseroutput .hlist dt li:lastchild::after,.mwparseroutput .hlist li dd:lastchild::after,.mwparseroutput .hlist li dt:lastchild::after,.mwparseroutput .hlist li li:lastchild::after{content:")";fontweight:normal}.mwparseroutput .hlist ol{counterreset:listitem}.mwparseroutput .hlist ol>li{counterincrement:listitem}.mwparseroutput .hlist ol>li::before{content:" "counter(listitem)"\a0 "}.mwparseroutput .hlist dd ol>li:firstchild::before,.mwparseroutput .hlist dt ol>li:firstchild::before,.mwparseroutput .hlist li ol>li:firstchild::before{content:" ("counter(listitem)"\a0 "}.mwparseroutput .navbar{display:inline;fontsize:88%;fontweight:normal}.mwparseroutput .navbarcollapse{float:left;textalign:left}.mwparseroutput .navbarboxtext{wordspacing:0}.mwparseroutput .navbar ul{display:inlineblock;whitespace:nowrap;lineheight:inherit}.mwparseroutput .navbarbrackets::before{marginright:0.125em;content:"[ "}.mwparseroutput .navbarbrackets::after{marginleft:0.125em;content:" ]"}.mwparseroutput .navbar li{wordspacing:0.125em}.mwparseroutput .navbar a>span,.mwparseroutput .navbar a>abbr{textdecoration:inherit}.mwparseroutput .navbarmini abbr{fontvariant:smallcaps;borderbottom:none;textdecoration:none;cursor:inherit}.mwparseroutput .navbarctfull{fontsize:114%;margin:0 7em}.mwparseroutput .navbarctmini{fontsize:114%;margin:0 4em}viewtalkedit}  
Total population = P + N 
Predicted Positive (PP)  Predicted Negative (PN)  Informedness, bookmaker informedness (BM) = TPR + TNR − 1 
Prevalence threshold (PT) = √TPR × FPR  FPR/TPR  FPR  
Actual condition

Positive (P) ^{[a]}  True positive (TP), hit^{[b]} 
False negative (FN), miss, underestimation 
True positive rate (TPR), recall, sensitivity (SEN), probability of detection, hit rate, power = TP/P = 1 − FNR 
False negative rate (FNR), miss rate type II error ^{[c]} = FN/P = 1 − TPR 
Negative (N)^{[d]}  False positive (FP), false alarm, overestimation 
True negative (TN), correct rejection^{[e]} 
False positive rate (FPR), probability of false alarm, fallout type I error ^{[f]} = FP/N = 1 − TNR 
True negative rate (TNR), specificity (SPC), selectivity = TN/N = 1 − FPR  
Prevalence = P/P + N 
Positive predictive value (PPV), precision = TP/PP = 1 − FDR 
False omission rate (FOR) = FN/PN = 1 − NPV 
Positive likelihood ratio (LR+) = TPR/FPR 
Negative likelihood ratio (LR−) = FNR/TNR  
Accuracy (ACC) = TP + TN/P + N 
False discovery rate (FDR) = FP/PP = 1 − PPV 
Negative predictive value (NPV) = TN/PN = 1 − FOR 
Markedness (MK), deltaP (Δp) = PPV + NPV − 1 
Diagnostic odds ratio (DOR) = LR+/LR−  
Balanced accuracy (BA) = TPR + TNR/2 
F_{1} score = 2 PPV × TPR/PPV + TPR = 2 TP/2 TP + FP + FN 
Fowlkes–Mallows index (FM) = √PPV × TPR 
Matthews correlation coefficient (MCC) = √TPR × TNR × PPV × NPV  √FNR × FPR × FOR × FDR 
Threat score (TS), critical success index (CSI), Jaccard index = TP/TP + FN + FP 
Fecal occult blood screen test outcome  ^{viewtalkedit}  
Total population (pop.) = 2030 
Test outcome positive  Test outcome negative  Accuracy (ACC) = (TP + TN) / pop.
= (20 + 1820) / 2030 ≈ 90.64% 
F_{1} score = 2 × precision × recall/precision + recall
≈ 0.174  
Patients with bowel cancer (as confirmed on endoscopy) 
Actual condition positive (AP) = 30 (2030 × 1.48%) 
True positive (TP) = 20 (2030 × 1.48% × 67%) 
False negative (FN) = 10 (2030 × 1.48% × (100% − 67%)) 
True positive rate (TPR), recall, sensitivity = TP / AP
= 20 / 30 ≈ 66.7% 
False negative rate (FNR), miss rate = FN / AP
= 10 / 30 ≈ 33.3% 
Actual condition negative (AN) = 2000 (2030 × (100% − 1.48%)) 
False positive (FP) = 180 (2030 × (100% − 1.48%) × (100% − 91%)) 
True negative (TN) = 1820 (2030 × (100% − 1.48%) × 91%) 
False positive rate (FPR), fallout, probability of false alarm = FP / AN
= 180 / 2000 = 9.0% 
Specificity, selectivity, true negative rate (TNR) = TN / AN
= 1820 / 2000 = 91%  
Prevalence = AP / pop.
= 30 / 2030 ≈ 1.48% 
Positive predictive value (PPV), precision = TP / (TP + FP)
= 20 / (20 + 180) = 10% 
False omission rate (FOR) = FN / (FN + TN)
= 10 / (10 + 1820) ≈ 0.55% 
Positive likelihood ratio (LR+) = TPR/FPR
= (20 / 30) / (180 / 2000) ≈ 7.41 
Negative likelihood ratio (LR−) = FNR/TNR
= (10 / 30) / (1820 / 2000) ≈ 0.366  
False discovery rate (FDR) = FP / (TP + FP)
= 180 / (20 + 180) = 90.0% 
Negative predictive value (NPV) = TN / (FN + TN)
= 1820 / (10 + 1820) ≈ 99.45% 
Diagnostic odds ratio (DOR) = LR+/LR−
≈ 20.2 
Related calculations
This hypothetical screening test (fecal occult blood test) correctly identified twothirds (66.7%) of patients with colorectal cancer.^{[a]} Unfortunately, factoring in prevalence rates reveals that this hypothetical test has a high false positive rate, and it does not reliably identify colorectal cancer in the overall population of asymptomatic people (PPV = 10%).
On the other hand, this hypothetical test demonstrates very accurate detection of cancerfree individuals (NPV ≈ 99.5%). Therefore, when used for routine colorectal cancer screening with asymptomatic adults, a negative result supplies important data for the patient and doctor, such as ruling out cancer as the cause of gastrointestinal symptoms or reassuring patients worried about developing colorectal cancer.
Sensitivity and specificity values alone may be highly misleading. The 'worstcase' sensitivity or specificity must be calculated in order to avoid reliance on experiments with few results. For example, a particular test may easily show 100% sensitivity if tested against the gold standard four times, but a single additional test against the gold standard that gave a poor result would imply a sensitivity of only 80%. A common way to do this is to state the binomial proportion confidence interval, often calculated using a Wilson score interval.
Confidence intervals for sensitivity and specificity can be calculated, giving the range of values within which the correct value lies at a given confidence level (e.g., 95%).^{[26]}
In information retrieval, the positive predictive value is called precision, and sensitivity is called recall. Unlike the Specificity vs Sensitivity tradeoff, these measures are both independent of the number of true negatives, which is generally unknown and much larger than the actual numbers of relevant and retrieved documents. This assumption of very large numbers of true negatives versus positives is rare in other applications.^{[13]}
The Fscore can be used as a single measure of performance of the test for the positive class. The Fscore is the harmonic mean of precision and recall:
In the traditional language of statistical hypothesis testing, the sensitivity of a test is called the statistical power of the test, although the word power in that context has a more general usage that is not applicable in the present context. A sensitive test will have fewer Type II errors.
Similarly to the domain of information retrieval, in the research area of gene prediction, the number of true negatives (nongenes) in genomic sequences is generally unknown and much larger than the actual number of genes (true positives). The convenient and intuitively understood term specificity in this research area has been frequently used with the mathematical formula for precision and recall as defined in biostatistics. The pair of thus defined specificity (as positive predictive value) and sensitivity (true positive rate) represent major parameters characterizing the accuracy of gene prediction algorithms. ^{[27]} ^{[28]} ^{[29]} ^{[30]} Conversely, the term specificity in a sense of true negative rate would have little, if any, application in the genome analysis research area.