Xue, Chenhoa and Hu, Weitao and Chakraborty, Joyraj and Guo, Zhijin and Li, Kang and Shi, Tianyu and Reed, Martin and Thomos, Nikolaos (2026) Constructing composite features for interpretable music-tagging. In: IEEE International Conference on Acoustics, Speech, and Signal Processing, 2026-05-04 - 2026-05-08, Barcelona, Spain. (In Press)
Xue, Chenhoa and Hu, Weitao and Chakraborty, Joyraj and Guo, Zhijin and Li, Kang and Shi, Tianyu and Reed, Martin and Thomos, Nikolaos (2026) Constructing composite features for interpretable music-tagging. In: IEEE International Conference on Acoustics, Speech, and Signal Processing, 2026-05-04 - 2026-05-08, Barcelona, Spain. (In Press)
Xue, Chenhoa and Hu, Weitao and Chakraborty, Joyraj and Guo, Zhijin and Li, Kang and Shi, Tianyu and Reed, Martin and Thomos, Nikolaos (2026) Constructing composite features for interpretable music-tagging. In: IEEE International Conference on Acoustics, Speech, and Signal Processing, 2026-05-04 - 2026-05-08, Barcelona, Spain. (In Press)
Abstract
Combining multiple audio features can improve the performance of music tagging, but common deep learning-based feature fusion methods often lack interpretability. To address this problem, we propose a Genetic Programming (GP) pipeline that automatically evolves composite features by mathematically combining base music features, thereby capturing synergistic interactions while preserving interpretability. This approach provides representational benefits similar to deep feature fusion without sacrificing interpretability. Experiments on the MTG-Jamendo and GTZAN datasets demonstrate consistent improvements compared to state-of-the-art systems across base feature sets at different abstraction levels. It should be noted that most of the performance gains are noticed within the first few hundred GP evaluations, indicating that effective feature combinations can be identified under modest search budgets. The top evolved expressions include linear, nonlinear, and conditional forms, with various low-complexity solutions at top performance aligned with parsimony pressure to prefer simpler expressions. Analyzing these composite features further reveals which interactions and transformations tend to be beneficial for tagging, offering insights that remain opaque in black-box deep models.
| Item Type: | Conference or Workshop Item (Paper) |
|---|---|
| Additional Information: | Published proceedings: _not provided_ |
| Uncontrolled Keywords: | Music tagging; Feature construction; Genetic programming; Interpretability; Music information retrieval |
| Subjects: | Z Bibliography. Library Science. Information Resources > ZR Rights Retention |
| Divisions: | Faculty of Science and Health Faculty of Science and Health > Computer Science and Electronic Engineering, School of |
| SWORD Depositor: | Unnamed user with email elements@essex.ac.uk |
| Depositing User: | Unnamed user with email elements@essex.ac.uk |
| Date Deposited: | 02 Apr 2026 16:02 |
| Last Modified: | 02 Apr 2026 16:02 |
| URI: | http://repository.essex.ac.uk/id/eprint/43070 |
Available files
Filename: ICASSP_2026_Music_Tagging.pdf
Licence: Creative Commons: Attribution 4.0