Constructing Evaluation Corpora for Automated Clinical Named Entity Recognition. S, P., V., O., M., D, G., K., S., P., & H, C., G., C., M., D., P. In Proceedings of the 6th international Conference on Language Resources and Evaluation (2008), pages 3143-3150, 2008. European Language Resources Association (ELRA).
Constructing Evaluation Corpora for Automated Clinical Named Entity Recognition [link]Website  abstract   bibtex   
We report on the construction of a gold-standard dataset consisting of annotated clinical notes suitable for evaluating our biomedical named entity recognition system. The dataset is the result of consensus between four human annotators and contains 1,556 annotations on 160 clinical notes using 658 unique concept codes from SNOMED-CT corresponding to human disorders. Inter-annotator agreement was calculated on annotations from 100 of the documents for span (90.9%), concept code (81.7%), context (84.8%), and status (86.0%) agreement. Complete agreement for span, concept code, context, and status was 74.6%. We found that creating a consensus set based on annotations from two independently-created annotation sets can reduce inter-annotator disagreement by 32.3%. We found little benefit to pre-annotating the corpus with a third-party named entity recognizer.
@inProceedings{
 title = {Constructing Evaluation Corpora for Automated Clinical Named Entity Recognition},
 type = {inProceedings},
 year = {2008},
 identifiers = {[object Object]},
 pages = {3143-3150},
 websites = {http://www.lrec-conf.org/proceedings/lrec2008/},
 publisher = {European Language Resources Association (ELRA)},
 editors = {[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object]},
 id = {d9b0db98-b54f-3b3c-8b4c-90884f2aa3ce},
 created = {2011-12-28T07:04:55.000Z},
 file_attached = {false},
 profile_id = {5284e6aa-156c-3ce5-bc0e-b80cf09f3ef6},
 group_id = {066b42c8-f712-3fc3-abb2-225c158d2704},
 last_modified = {2017-03-14T14:36:19.698Z},
 tags = {named entities},
 read = {false},
 starred = {false},
 authored = {false},
 confirmed = {true},
 hidden = {false},
 citation_key = {S2008},
 private_publication = {false},
 abstract = {We report on the construction of a gold-standard dataset consisting of annotated clinical notes suitable for evaluating our biomedical named entity recognition system. The dataset is the result of consensus between four human annotators and contains 1,556 annotations on 160 clinical notes using 658 unique concept codes from SNOMED-CT corresponding to human disorders. Inter-annotator agreement was calculated on annotations from 100 of the documents for span (90.9%), concept code (81.7%), context (84.8%), and status (86.0%) agreement. Complete agreement for span, concept code, context, and status was 74.6%. We found that creating a consensus set based on annotations from two independently-created annotation sets can reduce inter-annotator disagreement by 32.3%. We found little benefit to pre-annotating the corpus with a third-party named entity recognizer.},
 bibtype = {inProceedings},
 author = {S, Philip V Ogren M and D, Guergana K Savova Ph and H, Christopher G Chute M D P},
 booktitle = {Proceedings of the 6th international Conference on Language Resources and Evaluation (2008)}
}

Downloads: 0