Kia, Mahsa Abazari and Garifullina, Aygul and Kern, Mathias and Chamberlain, Jon and Jameel, Shoaib (2022) Adaptable Closed-Domain Question Answering Using Contextualized CNN-Attention Models and Question Expansion. IEEE Access, 10. pp. 45080-45092. DOI https://doi.org/10.1109/access.2022.3170466
Kia, Mahsa Abazari and Garifullina, Aygul and Kern, Mathias and Chamberlain, Jon and Jameel, Shoaib (2022) Adaptable Closed-Domain Question Answering Using Contextualized CNN-Attention Models and Question Expansion. IEEE Access, 10. pp. 45080-45092. DOI https://doi.org/10.1109/access.2022.3170466
Kia, Mahsa Abazari and Garifullina, Aygul and Kern, Mathias and Chamberlain, Jon and Jameel, Shoaib (2022) Adaptable Closed-Domain Question Answering Using Contextualized CNN-Attention Models and Question Expansion. IEEE Access, 10. pp. 45080-45092. DOI https://doi.org/10.1109/access.2022.3170466
Abstract
In closed-domain Question Answering (QA), the goal is to retrieve answers to questions within a specific domain. The main challenge of closed-domain QA is to develop a model that only requires small datasets for training since large-scale corpora may not be available. One approach is a flexible QA model that can adapt to different closed domains and train on their corpora. In this paper, we present a novel versatile reading comprehension style approach for closed-domain QA (called CA-AcdQA). The approach is based on pre-trained contextualized language models, Convolutional Neural Network (CNN), and a self-attention mechanism. The model captures the relevance between the question and context sentences at different levels of granularity by exploring the dependencies between the features extracted by the CNN. Moreover, we include candidate answer identification and question expansion techniques for context reduction and rewriting ambiguous questions. The model can be tuned to different domains with a small training dataset for sentence-level QA. The approach is tested on four publicly-available closed-domain QA datasets: Tesla (person), California (region), EU-law (system), and COVID-QA (biomedical) against nine other QA approaches. Results show that the ALBERT model variant outperforms all approaches on all datasets with a significant increase in Exact Match and F1 score. Furthermore, for the Covid-19 QA in which the text is complicated and specialized, the model is improved considerably with additional biomedical training resources (an F1 increase of 15.9 over the next highest baseline).
Item Type: | Article |
---|---|
Uncontrolled Keywords: | Closed-domain question answering; convolutional neural network; question expansion; self-attention |
Divisions: | Faculty of Science and Health Faculty of Science and Health > Computer Science and Electronic Engineering, School of |
SWORD Depositor: | Unnamed user with email elements@essex.ac.uk |
Depositing User: | Unnamed user with email elements@essex.ac.uk |
Date Deposited: | 20 Jun 2022 12:18 |
Last Modified: | 30 Oct 2024 19:33 |
URI: | http://repository.essex.ac.uk/id/eprint/32814 |
Available files
Filename: Adaptable_Closed-Domain_Question_Answering_Using_Contextualized_CNN-Attention_Models_and_Question_Expansion.pdf
Licence: Creative Commons: Attribution 3.0