DocumentCode :
2398545
Title :
Cheap, Fast, and Good Enough for the Non-biomedical Domain but is It Usable for Clinical Natural Language Processing? Evaluating Crowdsourcing for Clinical Trial Announcement Named Entity Annotations
Author :
Zhai, Haijun ; Lingren, Todd ; Deleger, Louise ; Li, Qi ; Kaiser, Megan ; Stoutenborough, Laura ; Solti, Imre
Author_Institution :
Div. of Biomed. Inf., Cincinnati Children´´s Hosp. Med. Center, Cincinnati, OH, USA
fYear :
2012
fDate :
27-28 Sept. 2012
Firstpage :
106
Lastpage :
106
Abstract :
Building upon previous work from the general crowdsourcing research, this study investigates the usability of crowdsourcing in the clinical NLP domain for annotating medical named entities and entity linkages in a clinical trial announcement (CTA) corpus. The results indicate that crowdsourcing is a feasible, inexpensive, fast, and practical approach to annotate clinical text (without PHI) on large scale for medical named entities. The crowdsourcing program code was released publicly.
Keywords :
information retrieval; medical computing; natural language processing; outsourcing; text analysis; CTA corpus; clinical NLP domain; clinical natural language processing; clinical text annotation; clinical trial announcement corpus; crowdsourcing evaluation; crowdsourcing program code; crowdsourcing usability; entity linkages; medical named entity annotation; Biomedical imaging; Clinical trials; Hospitals; Joining processes; Natural language processing; Pediatrics; Usability;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Healthcare Informatics, Imaging and Systems Biology (HISB), 2012 IEEE Second International Conference on
Conference_Location :
San Diego, CA
Print_ISBN :
978-1-4673-4803-4
Type :
conf
DOI :
10.1109/HISB.2012.31
Filename :
6366196
Link To Document :
بازگشت