Revisiting Online Learning Approach to Inverse Linear Optimization: A Fenchel–Young Loss Perspective and Gap-Dependent Regret Analysis

Shinsaku Sakaue, Han Bao, Taira Tsuchiya
Proceedings of The 28th International Conference on Artificial Intelligence and Statistics, PMLR 258:46-54, 2025.

Abstract

This paper revisits the online learning approach to inverse linear optimization studied by B{ä}rmann et al. (2017), where the goal is to infer an unknown linear objective function of an agent from sequential observations of the agent’s input-output pairs. First, we provide a simple understanding of the online learning approach through its connection to online convex optimization of \emph{Fenchel–Young losses}. As a byproduct, we present an offline guarantee on the \emph{suboptimality loss}, which measures how well predicted objective vectors explain the agent’s choices, without assuming the optimality of the agent’s choices. Second, assuming that there is a gap between optimal and suboptimal objective values in the agent’s decision problems, we obtain an upper bound independent of the time horizon $T$ on the sum of suboptimality and \emph{estimate losses}, where the latter measures the quality of solutions recommended by predicted objective vectors. Interestingly, our gap-dependent analysis achieves a faster rate than the standard $O(\sqrt{T})$ regret bound by exploiting structures specific to inverse linear optimization, even though neither the loss functions nor their domains possess desirable properties, such as strong convexity.

Cite this Paper


BibTeX
@InProceedings{pmlr-v258-sakaue25a, title = {Revisiting Online Learning Approach to Inverse Linear Optimization: A Fenchel–Young Loss Perspective and Gap-Dependent Regret Analysis}, author = {Sakaue, Shinsaku and Bao, Han and Tsuchiya, Taira}, booktitle = {Proceedings of The 28th International Conference on Artificial Intelligence and Statistics}, pages = {46--54}, year = {2025}, editor = {Li, Yingzhen and Mandt, Stephan and Agrawal, Shipra and Khan, Emtiyaz}, volume = {258}, series = {Proceedings of Machine Learning Research}, month = {03--05 May}, publisher = {PMLR}, pdf = {https://raw.githubusercontent.com/mlresearch/v258/main/assets/sakaue25a/sakaue25a.pdf}, url = {https://proceedings.mlr.press/v258/sakaue25a.html}, abstract = {This paper revisits the online learning approach to inverse linear optimization studied by B{ä}rmann et al. (2017), where the goal is to infer an unknown linear objective function of an agent from sequential observations of the agent’s input-output pairs. First, we provide a simple understanding of the online learning approach through its connection to online convex optimization of \emph{Fenchel–Young losses}. As a byproduct, we present an offline guarantee on the \emph{suboptimality loss}, which measures how well predicted objective vectors explain the agent’s choices, without assuming the optimality of the agent’s choices. Second, assuming that there is a gap between optimal and suboptimal objective values in the agent’s decision problems, we obtain an upper bound independent of the time horizon $T$ on the sum of suboptimality and \emph{estimate losses}, where the latter measures the quality of solutions recommended by predicted objective vectors. Interestingly, our gap-dependent analysis achieves a faster rate than the standard $O(\sqrt{T})$ regret bound by exploiting structures specific to inverse linear optimization, even though neither the loss functions nor their domains possess desirable properties, such as strong convexity.} }
Endnote
%0 Conference Paper %T Revisiting Online Learning Approach to Inverse Linear Optimization: A Fenchel–Young Loss Perspective and Gap-Dependent Regret Analysis %A Shinsaku Sakaue %A Han Bao %A Taira Tsuchiya %B Proceedings of The 28th International Conference on Artificial Intelligence and Statistics %C Proceedings of Machine Learning Research %D 2025 %E Yingzhen Li %E Stephan Mandt %E Shipra Agrawal %E Emtiyaz Khan %F pmlr-v258-sakaue25a %I PMLR %P 46--54 %U https://proceedings.mlr.press/v258/sakaue25a.html %V 258 %X This paper revisits the online learning approach to inverse linear optimization studied by B{ä}rmann et al. (2017), where the goal is to infer an unknown linear objective function of an agent from sequential observations of the agent’s input-output pairs. First, we provide a simple understanding of the online learning approach through its connection to online convex optimization of \emph{Fenchel–Young losses}. As a byproduct, we present an offline guarantee on the \emph{suboptimality loss}, which measures how well predicted objective vectors explain the agent’s choices, without assuming the optimality of the agent’s choices. Second, assuming that there is a gap between optimal and suboptimal objective values in the agent’s decision problems, we obtain an upper bound independent of the time horizon $T$ on the sum of suboptimality and \emph{estimate losses}, where the latter measures the quality of solutions recommended by predicted objective vectors. Interestingly, our gap-dependent analysis achieves a faster rate than the standard $O(\sqrt{T})$ regret bound by exploiting structures specific to inverse linear optimization, even though neither the loss functions nor their domains possess desirable properties, such as strong convexity.
APA
Sakaue, S., Bao, H. & Tsuchiya, T.. (2025). Revisiting Online Learning Approach to Inverse Linear Optimization: A Fenchel–Young Loss Perspective and Gap-Dependent Regret Analysis. Proceedings of The 28th International Conference on Artificial Intelligence and Statistics, in Proceedings of Machine Learning Research 258:46-54 Available from https://proceedings.mlr.press/v258/sakaue25a.html.

Related Material