Sun, Xia and Gao, Yi and Sutcliffe, Richard and Guo, Shou-Xi and Wang, Xin and Feng, Jun (2021) Word Representation Learning Based on Bidirectional GRUs With Drop Loss for Sentiment Classification. IEEE Transactions on Systems, Man, and Cybernetics: Systems, 51 (7). pp. 4532-4542. DOI https://doi.org/10.1109/tsmc.2019.2940097
Sun, Xia and Gao, Yi and Sutcliffe, Richard and Guo, Shou-Xi and Wang, Xin and Feng, Jun (2021) Word Representation Learning Based on Bidirectional GRUs With Drop Loss for Sentiment Classification. IEEE Transactions on Systems, Man, and Cybernetics: Systems, 51 (7). pp. 4532-4542. DOI https://doi.org/10.1109/tsmc.2019.2940097
Sun, Xia and Gao, Yi and Sutcliffe, Richard and Guo, Shou-Xi and Wang, Xin and Feng, Jun (2021) Word Representation Learning Based on Bidirectional GRUs With Drop Loss for Sentiment Classification. IEEE Transactions on Systems, Man, and Cybernetics: Systems, 51 (7). pp. 4532-4542. DOI https://doi.org/10.1109/tsmc.2019.2940097
Abstract
Sentiment classification is a fundamental task in many natural language processing applications. Neural networks have achieved great successes on the sentiment classification task in recent years, since recurrent neural networks and long-short-term memory networks have the ability to deal with sequences of different lengths and to capture contextual semantic information. However, the effectiveness of these methods is limited when used to extract contextual information from relatively long texts. Therefore, in our model, we apply bidirectional gated recurrent units to capture contextual information as far as possible when learning word representations, which may effectively reduce the noise compared to other methods. We also propose a novel loss function namely drop loss (DL) which makes the model focus on the hard examples - examples which are easily classified incorrectly - in order to improve the accuracy of the model. We experiment on four commonly used datasets, and the results show that the proposed method has a good performance on four datasets, and needs fewer parameters compared with recent benchmarks, such as CoVe, ULMFiT, embeddings from language models, and bidirectional encoder representations from transformers. Furthermore, we demonstrate that the classification performance of existing shallow network models can be significantly improved by using DL. In particular, the accuracy of the CNN+LSTM model improves 9% on the IMDB-10 dataset.
Item Type: | Article |
---|---|
Uncontrolled Keywords: | Loss function; neural networks; sentiment classification; word presentation |
Divisions: | Faculty of Science and Health Faculty of Science and Health > Computer Science and Electronic Engineering, School of |
SWORD Depositor: | Unnamed user with email elements@essex.ac.uk |
Depositing User: | Unnamed user with email elements@essex.ac.uk |
Date Deposited: | 11 Oct 2023 14:50 |
Last Modified: | 11 Oct 2023 14:50 |
URI: | http://repository.essex.ac.uk/id/eprint/36617 |
Available files
Filename: Word_Representation_Learning_Based_on_Bidirectional_GRUs_With_Drop_Loss_for_Sentiment_Classification.pdf
Licence: Creative Commons: Attribution 4.0