Community annotation experiment for ground truth generation for the i2b2 medication challenge.

J Am Med Inform Assoc

Department of Information Studies, University at Albany, State University of New York, Albany, NY, USA.

Published: November 2010

AI Article Synopsis

  • The i2b2 medication challenge team conducted a community annotation experiment during the Third i2b2 Workshop to evaluate annotation quality for clinical discharge summaries.
  • Participants were tasked with annotating 10 discharge summaries each, using guidelines provided by the authors, with disagreements resolved by a third annotator.
  • Results showed that the community's inter-annotator agreement was on par with expert annotators, achieving high-quality ground truth scores (F-measures above 0.90) and confirming the community's effectiveness in performing complex clinical annotations.

Article Abstract

Objective: Within the context of the Third i2b2 Workshop on Natural Language Processing Challenges for Clinical Records, the authors (also referred to as 'the i2b2 medication challenge team' or 'the i2b2 team' for short) organized a community annotation experiment.

Design: For this experiment, the authors released annotation guidelines and a small set of annotated discharge summaries. They asked the participants of the Third i2b2 Workshop to annotate 10 discharge summaries per person; each discharge summary was annotated by two annotators from two different teams, and a third annotator from a third team resolved disagreements.

Measurements: In order to evaluate the reliability of the annotations thus produced, the authors measured community inter-annotator agreement and compared it with the inter-annotator agreement of expert annotators when both the community and the expert annotators generated ground truth based on pooled system outputs. For this purpose, the pool consisted of the three most densely populated automatic annotations of each record. The authors also compared the community inter-annotator agreement with expert inter-annotator agreement when the experts annotated raw records without using the pool. Finally, they measured the quality of the community ground truth by comparing it with the expert ground truth.

Results And Conclusions: The authors found that the community annotators achieved comparable inter-annotator agreement to expert annotators, regardless of whether the experts annotated from the pool. Furthermore, the ground truth generated by the community obtained F-measures above 0.90 against the ground truth of the experts, indicating the value of the community as a source of high-quality ground truth even on intricate and domain-specific annotation tasks.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC2995684PMC
http://dx.doi.org/10.1136/jamia.2010.004200DOI Listing

Publication Analysis

Top Keywords

ground truth
24
inter-annotator agreement
20
agreement expert
12
expert annotators
12
community
9
community annotation
8
i2b2 medication
8
medication challenge
8
third i2b2
8
i2b2 workshop
8

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!