Overview of the Medical Question Answering Task at TREC 2017 LiveQA

View Researcher's Other Codes

Disclaimer: The provided code links for this paper are external links. Science Nest has no responsibility for the accuracy, legality or content of these links. Also, by downloading this code(s), you agree to comply with the terms of use as set out by the author(s) of the code(s).

Please contact us in case of a broken link from here

Authors Asma Ben Abacha, Eugene Agichtein, Yuval Pinter, Dina Demner-Fushman
Journal/Conference Name TREC
Paper Category
Paper Abstract We present an overview of the medical question answering task organized at the TREC 2017 LiveQA track. The task addresses the automatic answering of consumer health questions received by the US National Library of Medicine. We provided both training question-answer pairs, and test questions with reference answers 1. All questions were manually annotated with the main entities (foci) and question types. The medical task received eight runs from five participating teams. Different approaches have been applied, including classical answer retrieval based on question analysis and similar question retrieval. In particular, several deep learning approaches were tested, including attentional encoder-decoder networks, long short-term memory networks and convolutional neural networks. The training datasets were both from the open domain and the medical domain. We discuss the obtained results and give some insights for future research in medical question answering.
Date of publication 2017
Code Programming Language dataset
Comment

Copyright Researcher 2021