Hu, Jingzhao and Zhang, Hao and Liu, Yang and Sutcliffe, Richard and Feng, Jun (2022) BBW: a batch balance wrapper for training deep neural networks on extremely imbalanced datasets with few minority samples. Applied Intelligence, 52 (6). pp. 6723-6738. DOI https://doi.org/10.1007/s10489-021-02623-9
Hu, Jingzhao and Zhang, Hao and Liu, Yang and Sutcliffe, Richard and Feng, Jun (2022) BBW: a batch balance wrapper for training deep neural networks on extremely imbalanced datasets with few minority samples. Applied Intelligence, 52 (6). pp. 6723-6738. DOI https://doi.org/10.1007/s10489-021-02623-9
Hu, Jingzhao and Zhang, Hao and Liu, Yang and Sutcliffe, Richard and Feng, Jun (2022) BBW: a batch balance wrapper for training deep neural networks on extremely imbalanced datasets with few minority samples. Applied Intelligence, 52 (6). pp. 6723-6738. DOI https://doi.org/10.1007/s10489-021-02623-9
Abstract
In recent years, Deep Neural Networks (DNNs) have achieved excellent performance on many tasks, but it is very difficult to train good models from imbalanced datasets. Creating balanced batches either by majority data down-sampling or by minority data up-sampling can solve the problem in certain cases. However, it may lead to learning process instability and overfitting. In this paper, we propose the Batch Balance Wrapper (BBW), a novel framework which can adapt a general DNN to be well trained from extremely imbalanced datasets with few minority samples. In BBW, two extra network layers are added to the start of a DNN. The layers prevent overfitting of minority samples and improve the expressiveness of the sample distribution of minority samples. Furthermore, Batch Balance (BB), a class-based sampling algorithm, is proposed to make sure the samples in each batch are always balanced during the learning process. We test BBW on three well-known extremely imbalanced datasets with few minority samples. The maximum imbalance ratio reaches 1167:1 with only 16 positive samples. Compared with existing approaches, BBW achieves better classification performance. In addition, BBW-wrapped DNNs are 16.39 times faster, relative to unwrapped DNNs. Moreover, BBW does not require data pre-processing or additional hyper-parameter tuning, operations that may require additional processing time. The experiments prove that BBW can be applied to common applications of extremely imbalanced data with few minority samples, such as the classification of EEG signals, medical images and so on.
Item Type: | Article |
---|---|
Uncontrolled Keywords: | deep learning; deep neural networks; imbalanced dataset; batch balance wrapper framework |
Divisions: | Faculty of Science and Health > Computer Science and Electronic Engineering, School of |
SWORD Depositor: | Unnamed user with email elements@essex.ac.uk |
Depositing User: | Unnamed user with email elements@essex.ac.uk |
Date Deposited: | 09 Aug 2024 14:10 |
Last Modified: | 09 Aug 2024 14:10 |
URI: | http://repository.essex.ac.uk/id/eprint/36883 |
Available files
Filename: hu_jingzhao_batch_balance_wrapper_2022.pdf
Licence: Creative Commons: Attribution 4.0