Radman, Moein and Podmore, Joshua and Poli, Riccardo and Paulmann, Silke and Daly, Ian (2026) Decoding semantic categories: Insights from an fMRI ALE meta analysis. Journal of Neural Engineering, 22 (6). 061006-061006. DOI https://doi.org/10.1088/1741-2552/ae302b
Radman, Moein and Podmore, Joshua and Poli, Riccardo and Paulmann, Silke and Daly, Ian (2026) Decoding semantic categories: Insights from an fMRI ALE meta analysis. Journal of Neural Engineering, 22 (6). 061006-061006. DOI https://doi.org/10.1088/1741-2552/ae302b
Radman, Moein and Podmore, Joshua and Poli, Riccardo and Paulmann, Silke and Daly, Ian (2026) Decoding semantic categories: Insights from an fMRI ALE meta analysis. Journal of Neural Engineering, 22 (6). 061006-061006. DOI https://doi.org/10.1088/1741-2552/ae302b
Abstract
Objective.The human brain organizes conceptual knowledge into semantic categories; however, the extent to which these categories share common or distinct neural representations remains unclear. This study aims to clarify this organizational structure by identifying consistent, modality-controlled activation patterns across several widely used and frequently investigated semantic domains in functional magnetic resonance imaging (fMRI) research. By quantifying the distinctiveness and overlap among these patterns, we provide a more precise foundation for understanding the brain's semantic architecture, as well as for applications such as semantic brain-computer interfaces (BCI).Approach.Following PRISMA guidelines, we conducted a systematic review and meta-analysis of 75 fMRI studies covering six semantic categories: animals, tools, food, music, body parts, and pain. Using activation likelihood estimation, we identified convergent activation patterns for each category while controlling for stimulus modality (visual, auditory, tactile, and written). Subsequently, Jaccard-based overlap analyses were performed to quantify the degree of neural commonality and separability across concept-modality pairs, thereby revealing the underlying structure of representational similarity.Main results.Distinct yet partially overlapping activation networks were identified for each semantic category. Tools and animals showed shared activity in the lateral occipital and ventral temporal regions, reflecting common object-based visual processing. In contrast, food-related stimuli primarily recruited limbic and subcortical structures associated with affective and motivational processing. Music and animal sounds overlapped within the superior temporal and insular cortices, whereas body parts and pain engaged occipito-parietal and cingulo-insular networks, respectively. Together, these findings reveal a hierarchically organized and modality-dependent semantic architecture in the human brain.Significance.This meta-analysis offers a quantitative and integrative characterization of how semantic knowledge is distributed and differentiated across cortical systems. By demonstrating how conceptual content and sensory modality jointly shape neural organization, the study refines theoretical models of semantic cognition and provides a methodological basis for evaluating conceptual separability. These insights have direct implications for semantic neural decoding and for the development of BCI systems grounded in meaning-based neural representations.
| Item Type: | Article |
|---|---|
| Uncontrolled Keywords: | Brain; Humans; Magnetic Resonance Imaging; Brain Mapping; Semantics; Brain-Computer Interfaces |
| Subjects: | Z Bibliography. Library Science. Information Resources > ZR Rights Retention |
| Divisions: | Faculty of Science and Health > Computer Science and Electronic Engineering, School of Faculty of Science and Health > Psychology, Department of |
| SWORD Depositor: | Unnamed user with email elements@essex.ac.uk |
| Depositing User: | Unnamed user with email elements@essex.ac.uk |
| Date Deposited: | 13 Jan 2026 14:30 |
| Last Modified: | 13 Jan 2026 14:37 |
| URI: | http://repository.essex.ac.uk/id/eprint/42434 |
Available files
Filename: Manuscript_final.pdf
Licence: Creative Commons: Attribution 4.0