Liu, Kezhong and Zhou, Yiwen and Chen, Mozi and He, Jianhua (2026) MmWave Radar Perception Learning using Pervasive Visual-Inertial Supervision. IEEE Transactions on Intelligent Transportation Systems. (In Press)
Liu, Kezhong and Zhou, Yiwen and Chen, Mozi and He, Jianhua (2026) MmWave Radar Perception Learning using Pervasive Visual-Inertial Supervision. IEEE Transactions on Intelligent Transportation Systems. (In Press)
Liu, Kezhong and Zhou, Yiwen and Chen, Mozi and He, Jianhua (2026) MmWave Radar Perception Learning using Pervasive Visual-Inertial Supervision. IEEE Transactions on Intelligent Transportation Systems. (In Press)
Abstract
This paper introduces a radar perception learning framework guided by data collected from commonly equipped visual-inertial (VI) sensor suites on smart vehicles. Unlike existing approaches that rely on dense point clouds from 3D LiDARs, which are costly and not widely deployed, this method leverages the broader availability of VI data. However, visual images alone lack the ability to capture the three-dimensional motion of moving targets, which limits their effectiveness in supervising motion-related tasks. To overcome this limitation, the framework integrates multiple perception tasks such as odometry estimation, motion segmentation, and scene flow prediction into a unified learning process. The first component is an odometry estimation module that combines deterministic ego-motion models with data-driven learning results. This fusion helps accurately infer the scene flow of static background points while minimizing drift. The second component is a supervision signal extraction module that aligns optical and millimeter-wave radar measurements to guide the learning of radar scene flow and rigid transformations. This module improves the reliability of dynamic point supervision through joint constraints across sensing modalities. The third component introduces a feature-selection module designed for cross-modal learning. It enhances the accuracy of motion segmentation and enforces consistency between odometry and scene flow, resulting in more coherent radar perception outputs. Experimental evaluations show that this framework achieves superior performance in challenging conditions such as smoke-obscured environments. It surpasses state-of-the-art (SOTA) methods that depend on high-cost LiDAR systems.
| Item Type: | Article |
|---|---|
| Subjects: | Z Bibliography. Library Science. Information Resources > ZR Rights Retention |
| Divisions: | Faculty of Science and Health Faculty of Science and Health > Computer Science and Electronic Engineering, School of |
| SWORD Depositor: | Unnamed user with email elements@essex.ac.uk |
| Depositing User: | Unnamed user with email elements@essex.ac.uk |
| Date Deposited: | 26 Jan 2026 11:33 |
| Last Modified: | 26 Jan 2026 11:33 |
| URI: | http://repository.essex.ac.uk/id/eprint/42647 |
Available files
Filename: MmWave Radar Perception Learning using Pervasive Visual-Inertial Supervision.pdf
Licence: Creative Commons: Attribution 4.0