Font Size: a A A

Evaluating the effects of noninteractive and machine-assisted interactive manual clinical text annotation approaches on the quality of reference standards

Posted on:2015-10-03Degree:Ph.DType:Thesis
University:The University of UtahCandidate:South, Brett RayFull Text:PDF
GTID:2478390017999820Subject:Information Technology
Abstract/Summary:
Manual annotation of clinical texts is often used as a method of generating reference standards that provide data for training and evaluation of Natural Language Processing (NLP) systems. Manually annotating clinical texts is time consuming, expensive, and requires considerable cognitive effort on the part of human reviewers. Furthermore, reference standards must be generated in ways that produce consistent and reliable data but must also be valid in order to adequately evaluate the performance of those systems. The amount of labeled data necessary varies depending on the level of analysis, the complexity of the clinical use case, and the methods that will be used to develop automated machine systems for information extraction and classification. Evaluating methods that potentially reduce cost, manual human workload, introduce task efficiencies, and reduce the amount of labeled data necessary to train NLP tools for specific clinical use cases are active areas of research inquiry in the clinical NLP domain.;This dissertation integrates a mixed methods approach using methodologies from cognitive science and artificial intelligence with manual annotation of clinical texts. Aim 1 of this dissertation identifies factors that affect manual annotation of clinical texts. These factors are further explored by evaluating approaches that may introduce efficiencies into manual review tasks applied to two different NLP development areas---semantic annotation of clinical concepts and identification of information representing Protected Health Information (PHI) as defined by HIPAA. Both experiments integrate different priming mechanisms using noninteractive and machine-assisted methods. The main hypothesis for this research is that integrating pre-annotation or other machine-assisted methods within manual annotation workflows will improve efficiency of manual annotation tasks without diminishing the quality of generated reference standards.
Keywords/Search Tags:Manual, Reference standards, Annotation, Clinical texts, Methods, Machine-assisted, Evaluating, NLP
Related items