ActLoc: Learning to Localize on the Move via Active Viewpoint Selection

Jiajie Li, Boyang Sun, Luca Di Giammarino, Hermann Blum, Marc Pollefeys
Proceedings of The 9th Conference on Robot Learning, PMLR 305:1225-1245, 2025.

Abstract

Reliable localization is critical for robot navigation, yet many existing systems assume that all viewpoints along a trajectory are equally informative. In practice, localization becomes unreliable when the robot observes unmapped, ambiguous, or uninformative regions. To address this, we present ActLoc, an active viewpoint-aware planning framework for enhancing localization accuracy for general robot navigation tasks. At the core of ActLoc is an attention-based model trained at scale for viewpoint selection. This model encodes a metric map of the scene, along with camera poses used during map construction, and estimates localization accuracy over camera pitch and yaw directions at arbitrary 3D waypoint in space. This per-point accuracy distribution is integrated into the path planning process, allowing the robot to actively choose camera orientation that maximize localization robustness while respecting task and motion constraints. ActLoc achieves state-of-the-art performance in single-viewpoint selection task, and generalizes effectively to full-trajectory planning. It provides a modular enhancement to a wide range of navigation and inspection tasks in structured environments.

Cite this Paper


BibTeX
@InProceedings{pmlr-v305-li25b, title = {ActLoc: Learning to Localize on the Move via Active Viewpoint Selection}, author = {Li, Jiajie and Sun, Boyang and Giammarino, Luca Di and Blum, Hermann and Pollefeys, Marc}, booktitle = {Proceedings of The 9th Conference on Robot Learning}, pages = {1225--1245}, year = {2025}, editor = {Lim, Joseph and Song, Shuran and Park, Hae-Won}, volume = {305}, series = {Proceedings of Machine Learning Research}, month = {27--30 Sep}, publisher = {PMLR}, pdf = {https://raw.githubusercontent.com/mlresearch/v305/main/assets/li25b/li25b.pdf}, url = {https://proceedings.mlr.press/v305/li25b.html}, abstract = {Reliable localization is critical for robot navigation, yet many existing systems assume that all viewpoints along a trajectory are equally informative. In practice, localization becomes unreliable when the robot observes unmapped, ambiguous, or uninformative regions. To address this, we present ActLoc, an active viewpoint-aware planning framework for enhancing localization accuracy for general robot navigation tasks. At the core of ActLoc is an attention-based model trained at scale for viewpoint selection. This model encodes a metric map of the scene, along with camera poses used during map construction, and estimates localization accuracy over camera pitch and yaw directions at arbitrary 3D waypoint in space. This per-point accuracy distribution is integrated into the path planning process, allowing the robot to actively choose camera orientation that maximize localization robustness while respecting task and motion constraints. ActLoc achieves state-of-the-art performance in single-viewpoint selection task, and generalizes effectively to full-trajectory planning. It provides a modular enhancement to a wide range of navigation and inspection tasks in structured environments.} }
Endnote
%0 Conference Paper %T ActLoc: Learning to Localize on the Move via Active Viewpoint Selection %A Jiajie Li %A Boyang Sun %A Luca Di Giammarino %A Hermann Blum %A Marc Pollefeys %B Proceedings of The 9th Conference on Robot Learning %C Proceedings of Machine Learning Research %D 2025 %E Joseph Lim %E Shuran Song %E Hae-Won Park %F pmlr-v305-li25b %I PMLR %P 1225--1245 %U https://proceedings.mlr.press/v305/li25b.html %V 305 %X Reliable localization is critical for robot navigation, yet many existing systems assume that all viewpoints along a trajectory are equally informative. In practice, localization becomes unreliable when the robot observes unmapped, ambiguous, or uninformative regions. To address this, we present ActLoc, an active viewpoint-aware planning framework for enhancing localization accuracy for general robot navigation tasks. At the core of ActLoc is an attention-based model trained at scale for viewpoint selection. This model encodes a metric map of the scene, along with camera poses used during map construction, and estimates localization accuracy over camera pitch and yaw directions at arbitrary 3D waypoint in space. This per-point accuracy distribution is integrated into the path planning process, allowing the robot to actively choose camera orientation that maximize localization robustness while respecting task and motion constraints. ActLoc achieves state-of-the-art performance in single-viewpoint selection task, and generalizes effectively to full-trajectory planning. It provides a modular enhancement to a wide range of navigation and inspection tasks in structured environments.
APA
Li, J., Sun, B., Giammarino, L.D., Blum, H. & Pollefeys, M.. (2025). ActLoc: Learning to Localize on the Move via Active Viewpoint Selection. Proceedings of The 9th Conference on Robot Learning, in Proceedings of Machine Learning Research 305:1225-1245 Available from https://proceedings.mlr.press/v305/li25b.html.

Related Material