Si, Weiyong and Wang, Ning and Harris, Rebecca and Yang, Chenguang (2025) Deep Multimodal Imitation Learning-Based Framework for Robot-Assisted Medical Examination. IEEE Transactions on Industrial Electronics. pp. 1-9. DOI https://doi.org/10.1109/tie.2025.3589442
Si, Weiyong and Wang, Ning and Harris, Rebecca and Yang, Chenguang (2025) Deep Multimodal Imitation Learning-Based Framework for Robot-Assisted Medical Examination. IEEE Transactions on Industrial Electronics. pp. 1-9. DOI https://doi.org/10.1109/tie.2025.3589442
Si, Weiyong and Wang, Ning and Harris, Rebecca and Yang, Chenguang (2025) Deep Multimodal Imitation Learning-Based Framework for Robot-Assisted Medical Examination. IEEE Transactions on Industrial Electronics. pp. 1-9. DOI https://doi.org/10.1109/tie.2025.3589442
Abstract
Medical ultrasound examination is a challenging dexterous manipulation task for robots. Even for experienced sonographers, it involves real-time decision-making, motion control, and force regulation based on ultrasound images and patient feedback. In this article, we propose a unified framework for robot-assisted medical examination, specifically for the initial registration in artery scanning, by leveraging deep multimodal imitation learning, compliant control, and trajectory optimization. To process multimodal inputs during the initial registration phase, we investigate a deep imitation learning model that fuses RGB and ultrasound images, contact force, and proprioceptive data. The deep imitation learning model predicts the desired motion and contact force. We design a compliant controller in Cartesian space to track the desired trajectory and force. To smooth the trajectory and ensure safety, we employ a trajectory optimization planner between the deep imitation learning module and the low-level compliant controller. The generalization capability of the deep multimodal imitation learning module, control performance, and the quality of the acquired ultrasound images on both the Phantom and human subjects were evaluated. Experimental results show that the proposed approach significantly improves the success rate of autonomous ultrasound scanning from 75% to 90%, while also reducing the completion time.
| Item Type: | Article |
|---|---|
| Additional Information: | © 2025 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. |
| Uncontrolled Keywords: | Deep multimodal learning; force and com-pliant control; learning from demonstration (LfD); robot-assisted sonography; robot-assisted sonography; robot-assisted sonography |
| Divisions: | Faculty of Science and Health Faculty of Science and Health > Computer Science and Electronic Engineering, School of |
| SWORD Depositor: | Unnamed user with email elements@essex.ac.uk |
| Depositing User: | Unnamed user with email elements@essex.ac.uk |
| Date Deposited: | 05 Nov 2025 10:42 |
| Last Modified: | 05 Nov 2025 10:44 |
| URI: | http://repository.essex.ac.uk/id/eprint/41868 |
Available files
Filename: TIE_Deep_Multimodal_Imitation_Learning_based_Framework_for_Robot_assisted_Medical_Examination_final.pdf