Sensitivity and specificity are statistical measures of the performance of a binary classification test, also known in statistics as classification function:

  • Sensitivity (also called the true positive rate, or the recall
     in some fields) measures the proportion of positives that are correctly
     identified as such (e.g., the percentage of sick people who are
    correctly identified as having the condition).
  • Specificity (also called the true negative rate)
     measures the proportion of negatives that are correctly identified as
    such (e.g., the percentage of healthy people who are correctly
    identified as not having the condition).
  • Thus sensitivity quantifies the avoiding of false negatives, as specificity does for false positives. For any test, there is usually a trade-off between the measures. For instance, in an airport security
     setting in which one is testing for potential threats to safety,
    scanners may be set to trigger on low-risk items like belt buckles and
    keys (low specificity), in order to reduce the risk of missing objects
    that do pose a threat to the aircraft and those aboard (high
    sensitivity). This trade-off can be represented graphically as a receiver operating characteristic
     curve. A perfect predictor would be described as 100% sensitive (e.g.,
    all sick are identified as sick) and 100% specific (e.g., no healthy are
     identified as sick); however, theoretically any predictor will possess a
     minimum error bound known as the Bayes error rate.