skip to main content
10.1145/3354031.3354040acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicbipConference Proceedingsconference-collections
research-article

Relationships of Cohen's Kappa, Sensitivity, and Specificity for Unbiased Annotations

Authors Info & Claims
Published:13 August 2019Publication History

ABSTRACT

For the binary classification tasks in supervised learning, the labels of data have to be available for classifier development. Cohen's kappa is usually employed as a quality measure for data annotation, which is inconsistent with its true functionality of assessing the inter-annotator consistency. However, the derived relationship functions of Cohen's kappa, sensitivity, and specificity in the literature are complicated, thus are unable to be employed to interpret classification performance from kappa values. In this study, based on an annotation generation model, we develop simple relationships of kappa, sensitivity, and specificity when there is no bias in the annotations. A relationship between kappa and Youden's J statistic, a performance metric for binary classification, is further obtained. The derived relationships are evaluated on a synthetic dataset using linear regression analysis. The results demonstrate the accuracy of the derived relationships. It suggests the potential of estimating classification performance from kappa values when bias is absent in the annotations.

References

  1. Aickin, M., 1990. Maximum likelihood estimation of agreement in the constant predictive probability model, and its relation to Cohen's kappa. Biometrics, (June 1990), 293--302.Google ScholarGoogle ScholarCross RefCross Ref
  2. Beigman Klebanov, B. and Beigman, E., 2009. From annotator agreement to noise models. Computational Linguistics, 35, 4 (Dec. 2009), pp. 495--503.Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Byrt, T., Bishop, J. and Carlin, J.B., 1993. Bias, prevalence and kappa. Journal of clinical epidemiology, 46, 5 (May 1993), 423--429.Google ScholarGoogle ScholarCross RefCross Ref
  4. Carletta, J., 1996. Assessing agreement on classification tasks: the kappa statistic. Computational linguistics, 22, 2 (June 1996), 249--254.Google ScholarGoogle Scholar
  5. Cohen, J., 1960. A coefficient of agreement for nominal scales. Educational and psychological measurement, 20, 1 (April 1960), 37--46.Google ScholarGoogle Scholar
  6. Eugenio, B.D. and Glass, M., 2004. The kappa statistic: A second look. Computational linguistics, 30, 1 (March 2004), 95--101.Google ScholarGoogle Scholar
  7. Feuerman, M. and Miller, A.R., 2005. The kappa statistic as a function of sensitivity and specificity. International Journal of Mathematical Education in Science and Technology, 36,5 (July 2005), 517--527.Google ScholarGoogle ScholarCross RefCross Ref
  8. Feuerman, M. and Miller, A.R., 2007. Critical points for certain statistical measures of agreement. International Journal of Mathematical Education in Science and Technology, 38, 6, (Sep. 2007), 739--748.Google ScholarGoogle ScholarCross RefCross Ref
  9. Feuerman, M. and Miller, A.R., 2008. Relationships between statistical measures of agreement: sensitivity, specificity and kappa. Journal of evaluation in clinical practice, 14, 5(Oct. 2008), pp. 930--933.Google ScholarGoogle ScholarCross RefCross Ref
  10. Frénay, B. and Verleysen, M., 2014. Classification in the presence of label noise: a survey. IEEE Transactions on Neural Networks and Learning Systems, 25, 5 (May 2014), 845--869.Google ScholarGoogle ScholarCross RefCross Ref
  11. Ghosh, A., Manwani, N. and Sastry, P.S., 2015. Making risk minimization tolerant to label noise. Neurocomputing, 160 (July 2015), 93--107.Google ScholarGoogle Scholar
  12. Natarajan, N., Dhillon, I.S., Ravikumar, P.K. and Tewari, A., 2013. Learning with noisy labels. In Advances in Neural Information Processing Systems (Lake Tahoe, USA, Dec. 05-10, 2013), Curran Associates, Inc., New York, NY, 1196--1204.Google ScholarGoogle Scholar
  13. Passonneau, R.J. and Carpenter, B., 2014. The benefits of a model of annotation. Transactions of the Association for Computational Linguistics, 2 (Dec. 2014), 311--326.Google ScholarGoogle Scholar
  14. Schisterman, E.F., Perkins, N.J., Liu, A. and Bondell, H., 2005. Optimal cut-point and its corresponding Youden Index to discriminate individuals using pooled blood samples. Epidemiology, (Jan. 2005), 73--81.Google ScholarGoogle Scholar
  15. Thompson, W.D. and Walter, S.D., 1988. A reappraisal of the kappa coefficient. Journal of clinical epidemiology, 41, 10, (Jan. 1988), 949--958.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Relationships of Cohen's Kappa, Sensitivity, and Specificity for Unbiased Annotations

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Other conferences
      ICBIP '19: Proceedings of the 4th International Conference on Biomedical Signal and Image Processing
      August 2019
      149 pages
      ISBN:9781450372244
      DOI:10.1145/3354031

      Copyright © 2019 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 13 August 2019

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Research
      • Refereed limited

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader