Youden Index Calculator

Evaluate diagnostic test performance - enter sensitivity and specificity or raw counts to get J, LR+, LR−, PPV, and NPV.

J Youden Index Calculator

Core Metrics

Sensitivity
-
Specificity
-
Youden's J
-

Likelihood Ratios

LR+ = Sens/(1−Spec)
-
LR− = (1−Sens)/Spec
-

Predictive Values

PPV (precision)
-
NPV
-
Accuracy
-

Classification Metrics

F1 Score
-
MCC
-

📖 What is the Youden Index?

The Youden Index (also called Youden's J statistic or informedness) is a summary measure of diagnostic test performance that combines sensitivity and specificity into a single number: J = Sensitivity + Specificity − 1. Introduced by William J. Youden in 1950, it ranges from 0 (no better than chance) to 1 (perfect test). The index is widely used in clinical research, epidemiology, and machine learning to evaluate binary classifiers and to find the optimal cut-off point on a ROC curve.

Youden's J is geometrically the maximum vertical distance from the ROC curve to the 45-degree chance diagonal. Maximising J simultaneously optimises sensitivity and specificity under equal-weight assumptions. When the costs of false positives and false negatives differ (as in most real clinical situations), a weighted version of J or a different optimisation criterion may be preferred.

This calculator goes beyond just J to provide a complete diagnostic test evaluation: likelihood ratios (LR+ and LR−) quantify how much a positive or negative test result shifts the probability of disease; PPV and NPV give the probability of disease given the test result (requiring prevalence); accuracy is the overall proportion of correct classifications; F1 score is the harmonic mean of precision and recall; and the Matthews Correlation Coefficient (MCC) is the most balanced single metric for binary classification quality.

Applications span all of medicine and biostatistics: evaluating cancer screening tests, assessing machine learning model quality, optimising fraud detection thresholds, setting diagnostic cut-offs for blood biomarkers, comparing competing diagnostic protocols, and designing clinical decision support algorithms. Understanding all these metrics together - not just accuracy or sensitivity alone - is essential for responsible diagnostic test evaluation.

📐 Formulas

J = Sensitivity + Specificity − 1

Sensitivity (Recall, TPR): Sens = TP / (TP + FN)

Specificity (TNR): Spec = TN / (TN + FP)

Positive Likelihood Ratio: LR+ = Sensitivity / (1 − Specificity)

Negative Likelihood Ratio: LR− = (1 − Sensitivity) / Specificity

PPV (with prevalence π): PPV = (Sens × π) / (Sens × π + (1 − Spec) × (1 − π))

NPV (with prevalence π): NPV = (Spec × (1 − π)) / (Spec × (1 − π) + (1 − Sens) × π)

Accuracy: Acc = (TP + TN) / (TP + FP + TN + FN)

F1 Score: F1 = 2TP / (2TP + FP + FN)

Matthews Correlation Coefficient: MCC = (TP×TN − FP×FN) / √[(TP+FP)(TP+FN)(TN+FP)(TN+FN)]

📖 How to Use This Calculator

1
Choose your input mode: Manual if you already know sensitivity and specificity as decimal proportions (0–1). Counts if you have the raw confusion matrix (TP, FP, TN, FN).
2
In Manual mode, optionally enter the disease prevalence in your target population to get meaningful PPV and NPV values. Without prevalence, PPV/NPV cannot be calculated from sensitivity and specificity alone.
3
Click Calculate. The verdict gives an immediate quality rating for J. Check LR+ and LR− to assess clinical utility - LR+ > 10 and LR− < 0.1 are gold-standard thresholds.

💡 Example Calculations

Example 1 - Cancer screening test evaluation

1
Test performance: Sensitivity = 0.92, Specificity = 0.85. Prevalence in screening population = 0.01 (1%).
2
J = 0.92 + 0.85 − 1 = 0.77 - excellent. LR+ = 0.92/(1−0.85) = 6.13. LR− = (1−0.92)/0.85 = 0.094 < 0.1 - clinically useful negative LR.
3
PPV at 1% prevalence: (0.92×0.01)/(0.92×0.01 + 0.15×0.99) = 0.0092/0.1577 = 5.8% - only 1 in 17 positive tests is a true positive. NPV = 99.9% - almost all negatives are truly negative.
4
Implication: Despite excellent J, the low prevalence means most positives are false alarms. A confirmatory high-specificity test is needed after a positive screening result.
J = 0.77 (Excellent), LR+ = 6.13, LR− = 0.094, PPV = 5.8%
Try this example →

Example 2 - Fraud detection model (counts mode)

1
Confusion matrix: TP = 850, FN = 150, FP = 500, TN = 8500. Total = 10,000 transactions (1% fraud rate).
2
Sensitivity = 850/1000 = 0.85. Specificity = 8500/9000 = 0.944. J = 0.85 + 0.944 − 1 = 0.794 - excellent fraud detector.
3
PPV = 850/(850+500) = 63% - of transactions flagged as fraud, 63% are genuine fraud. F1 = 2×850/(2×850+500+150) = 0.72. MCC = (850×8500 − 500×150)/√(…) = 0.73.
4
Conclusion: Good model. The business must decide whether 63% precision (37% false alarm rate for investigators) and 85% recall (catching 85% of fraud) is acceptable, or whether to tune the threshold.
J = 0.794 (Excellent), PPV = 63%, F1 = 0.72, MCC = 0.73
Try this example →

❓ Frequently Asked Questions

What is the Youden Index?+
The Youden Index (also called Youden's J statistic or informedness) is a single summary measure of diagnostic test performance that combines sensitivity and specificity into one number. It is defined as J = Sensitivity + Specificity − 1. The index was introduced by William J. Youden in 1950 as a way to find the optimal cut-off point for a continuous diagnostic variable - the threshold that maximises J is the point on the ROC curve farthest from the diagonal chance line. J ranges from 0 (no diagnostic value, equivalent to random guessing) to 1 (perfect sensitivity and specificity). Negative values indicate the test performs worse than chance.
How is the Youden Index used to find an optimal cut-off?+
When you have a continuous biomarker (e.g. PSA level, blood pressure, HbA1c), you must choose a threshold above which the test is 'positive'. Choosing this threshold involves a trade-off: higher thresholds increase specificity but reduce sensitivity. The optimal cut-off is the threshold that maximises Youden's J = Sensitivity + Specificity − 1, which corresponds to the point on the ROC curve with the maximum vertical distance from the 45-degree line. This approach assumes equal costs for false positives and false negatives - if misses are more costly than false alarms, a lower threshold (higher sensitivity) is preferred.
What is the difference between sensitivity and specificity?+
Sensitivity (true positive rate, recall) = TP/(TP+FN) - the proportion of actual positives correctly identified by the test. High sensitivity means few false negatives (few missed cases). Specificity (true negative rate) = TN/(TN+FP) - the proportion of actual negatives correctly identified. High specificity means few false positives (few false alarms). There is typically a trade-off: increasing sensitivity (by lowering the threshold) decreases specificity and vice versa. A mnemonic: Sensitivity rules OUT (SnNout - a highly Sensitive test, when Negative, rules Out the diagnosis). Specificity rules IN (SpPin - a highly Specific test, when Positive, rules In the diagnosis).
What are likelihood ratios and how do I use them?+
Positive likelihood ratio LR+ = Sensitivity / (1 − Specificity) - how many times more likely a positive test result is in someone with the disease than someone without it. Negative likelihood ratio LR− = (1 − Sensitivity) / Specificity - how many times more likely a negative test result is in someone with the disease compared to someone without. Clinical rules of thumb: LR+ > 10 or LR− < 0.1 provide large, often conclusive, diagnostic shifts. LR+ 5–10 or LR− 0.1–0.2 provide moderate shifts. LR+ 2–5 or LR− 0.2–0.5 are small but sometimes clinically useful. LR near 1 means the test provides almost no diagnostic information.
What is PPV and NPV and why does prevalence matter?+
Positive predictive value (PPV) is the probability that a patient with a positive test truly has the disease: PPV = TP/(TP+FP). Negative predictive value (NPV) = TN/(TN+FN). Both depend critically on disease prevalence. A test with 95% sensitivity and 95% specificity sounds excellent, but in a population with 1% prevalence: PPV = (0.95×0.01)/(0.95×0.01 + 0.05×0.99) ≈ 16% - 84% of positive tests are false positives! In a population with 30% prevalence: PPV rises to 89%. This is why screening tests in low-prevalence populations require very high specificity to be useful.
What is the Matthews Correlation Coefficient (MCC)?+
The Matthews Correlation Coefficient (MCC) is a balanced measure of binary classification performance that accounts for all four cells of the confusion matrix: MCC = (TP×TN − FP×FN) / √[(TP+FP)(TP+FN)(TN+FP)(TN+FN)]. It ranges from −1 (perfect inverse prediction) to +1 (perfect prediction), with 0 indicating random chance. MCC is considered more informative than accuracy, F1, or Youden's J for imbalanced datasets because it uses all four confusion matrix quadrants. A perfect classifier has MCC = 1, J = 1, and F1 = 1 simultaneously.
How does F1 score differ from Youden's J?+
F1 score = 2×TP / (2×TP + FP + FN) = 2 × Precision × Recall / (Precision + Recall). It focuses on the positive class only and does not account for true negatives. Youden's J = Sensitivity + Specificity − 1 accounts for both the positive and negative classes symmetrically. For balanced datasets, both tend to agree. For imbalanced datasets, F1 is better when true negatives are very numerous (F1 ignores them, correctly for rare disease screening). J is better when you care about both classes equally. In clinical diagnostics, J is generally preferred because both false positives (unnecessary treatment) and false negatives (missed cases) have real costs.
What Youden Index values are considered clinically acceptable?+
There are no universally agreed thresholds, but commonly used benchmarks are: J ≥ 0.75 - excellent diagnostic test; J 0.50–0.75 - good, useful in clinical practice; J 0.25–0.50 - fair, limited clinical value without other information; J 0–0.25 - poor, close to chance; J < 0 - performs worse than chance, test is inverting positive/negative. These are guidelines, not absolute rules - the acceptable J depends on the clinical context, the consequences of false positives vs false negatives, and the availability of alternative tests.