BEVCalib: LiDAR-Camera Calibration via Geometry-Guided Bird’s-Eye View Representation

Weiduo Yuan, Jerry Li, Justin Yue, Divyank Shah, Konstantinos Karydis, Hang Qiu
Proceedings of The 9th Conference on Robot Learning, PMLR 305:4824-4836, 2025.

Abstract

Accurate LiDAR-camera calibration is the foundation of accurate multimodal fusion environmental perception for autonomous driving and robotic systems. Traditional calibration methods require extensive data collection in controlled environments and cannot compensate for the transformation changes during the vehicle/robot movement. In this paper, we propose the first model that uses bird’s-eye view (BEV) features to perform LiDAR camera calibration from raw data, termed BEVCalib. To achieve this, we extract camera BEV features and LiDAR BEV features separately and fuse them into a shared BEV feature space. To fully utilize the geometry information from the BEV feature, we introduce a novel feature selector to choose the most important feature in the transformation decoder, which reduces memory consumption and enables efficient training. Extensive evaluations in various datasets demonstrate that BEVCalib establishes a new state-of-the-art; improving the best open-source baseline by two orders of magnitude on KITTI, Nuscenes, and our dynamic extrinsic dataset, respectively, and outperforming the best baseline in literature by 72% on KITTI dataset, and 69% on Nuscenes dataset. All source code and checkpoints will be released.

Cite this Paper


BibTeX
@InProceedings{pmlr-v305-yuan25a, title = {BEVCalib: LiDAR-Camera Calibration via Geometry-Guided Bird’s-Eye View Representation}, author = {Yuan, Weiduo and Li, Jerry and Yue, Justin and Shah, Divyank and Karydis, Konstantinos and Qiu, Hang}, booktitle = {Proceedings of The 9th Conference on Robot Learning}, pages = {4824--4836}, year = {2025}, editor = {Lim, Joseph and Song, Shuran and Park, Hae-Won}, volume = {305}, series = {Proceedings of Machine Learning Research}, month = {27--30 Sep}, publisher = {PMLR}, pdf = {https://raw.githubusercontent.com/mlresearch/v305/main/assets/yuan25a/yuan25a.pdf}, url = {https://proceedings.mlr.press/v305/yuan25a.html}, abstract = {Accurate LiDAR-camera calibration is the foundation of accurate multimodal fusion environmental perception for autonomous driving and robotic systems. Traditional calibration methods require extensive data collection in controlled environments and cannot compensate for the transformation changes during the vehicle/robot movement. In this paper, we propose the first model that uses bird’s-eye view (BEV) features to perform LiDAR camera calibration from raw data, termed BEVCalib. To achieve this, we extract camera BEV features and LiDAR BEV features separately and fuse them into a shared BEV feature space. To fully utilize the geometry information from the BEV feature, we introduce a novel feature selector to choose the most important feature in the transformation decoder, which reduces memory consumption and enables efficient training. Extensive evaluations in various datasets demonstrate that BEVCalib establishes a new state-of-the-art; improving the best open-source baseline by two orders of magnitude on KITTI, Nuscenes, and our dynamic extrinsic dataset, respectively, and outperforming the best baseline in literature by 72% on KITTI dataset, and 69% on Nuscenes dataset. All source code and checkpoints will be released.} }
Endnote
%0 Conference Paper %T BEVCalib: LiDAR-Camera Calibration via Geometry-Guided Bird’s-Eye View Representation %A Weiduo Yuan %A Jerry Li %A Justin Yue %A Divyank Shah %A Konstantinos Karydis %A Hang Qiu %B Proceedings of The 9th Conference on Robot Learning %C Proceedings of Machine Learning Research %D 2025 %E Joseph Lim %E Shuran Song %E Hae-Won Park %F pmlr-v305-yuan25a %I PMLR %P 4824--4836 %U https://proceedings.mlr.press/v305/yuan25a.html %V 305 %X Accurate LiDAR-camera calibration is the foundation of accurate multimodal fusion environmental perception for autonomous driving and robotic systems. Traditional calibration methods require extensive data collection in controlled environments and cannot compensate for the transformation changes during the vehicle/robot movement. In this paper, we propose the first model that uses bird’s-eye view (BEV) features to perform LiDAR camera calibration from raw data, termed BEVCalib. To achieve this, we extract camera BEV features and LiDAR BEV features separately and fuse them into a shared BEV feature space. To fully utilize the geometry information from the BEV feature, we introduce a novel feature selector to choose the most important feature in the transformation decoder, which reduces memory consumption and enables efficient training. Extensive evaluations in various datasets demonstrate that BEVCalib establishes a new state-of-the-art; improving the best open-source baseline by two orders of magnitude on KITTI, Nuscenes, and our dynamic extrinsic dataset, respectively, and outperforming the best baseline in literature by 72% on KITTI dataset, and 69% on Nuscenes dataset. All source code and checkpoints will be released.
APA
Yuan, W., Li, J., Yue, J., Shah, D., Karydis, K. & Qiu, H.. (2025). BEVCalib: LiDAR-Camera Calibration via Geometry-Guided Bird’s-Eye View Representation. Proceedings of The 9th Conference on Robot Learning, in Proceedings of Machine Learning Research 305:4824-4836 Available from https://proceedings.mlr.press/v305/yuan25a.html.

Related Material