会议专题

Inter- Annotator Agreement and the Upper Limit on Machine Performance: Evidence from Biomedical Natural Language Processing

  Human-annotated data is a fundamental part of natural language processing system development and evaluation. The quality of that data is typically assessed by calculating the agreement between the annotators. It is widely assumed that this agreement between annotators is the upper limit on system performance in natural language processing: if humans cant agree with each other about the classification more than some percentage of the time, we dont expect a computer to do any better. We trace the logical positivist roots of the motivation for measuring inter-annotator agreement, demonstrate the prevalence of the widely-held assumption about the relationship between inter-annotator agreement and system performance, and present data that suggest that interannotator agreement is not, in fact, an upper bound on language processing system performance.

Natural Language Processing Supervised Machine Learning Evaluation Studies

Mayla Boguslav Kevin Bretonnel Cohen

Computational Bioscience Program,University Colorado School of Medicine,Aurora,CO,USA

国际会议

第十六届世界医药健康信息学大会((MEDINFO2017)、第二届世界医药健康信息学华语论坛(WCHIS 2017)、第15届全国医药信息学大会(CMIA 2017)

苏州

英文

298-302

2017-08-21(万方平台首次上网日期,不代表论文的发表时间)