A Bandit Model for Human-Machine Decision Making with Private Information and Opacity

Sebastian Bordt, Ulrike Von Luxburg
Proceedings of The 25th International Conference on Artificial Intelligence and Statistics, PMLR 151:7300-7319, 2022.

Abstract

Applications of machine learning inform human decision makers in a broad range of tasks. The resulting problem is usually formulated in terms of a single decision maker. We argue that it should rather be described as a two-player learning problem where one player is the machine and the other the human. While both players try to optimize the final decision, the setup is often characterized by (1) the presence of private information and (2) opacity, that is imperfect understanding between the decision makers. We prove that both properties can complicate decision making considerably. A lower bound quantifies the worst-case hardness of optimally advising a decision maker who is opaque or has access to private information. An upper bound shows that a simple coordination strategy is nearly minimax optimal. More efficient learning is possible under certain assumptions on the problem, for example that both players learn to take actions independently. Such assumptions are implicit in existing literature, for example in medical applications of machine learning, but have not been described or justified theoretically.

Cite this Paper


BibTeX
@InProceedings{pmlr-v151-bordt22a, title = { A Bandit Model for Human-Machine Decision Making with Private Information and Opacity }, author = {Bordt, Sebastian and Von Luxburg, Ulrike}, booktitle = {Proceedings of The 25th International Conference on Artificial Intelligence and Statistics}, pages = {7300--7319}, year = {2022}, editor = {Camps-Valls, Gustau and Ruiz, Francisco J. R. and Valera, Isabel}, volume = {151}, series = {Proceedings of Machine Learning Research}, month = {28--30 Mar}, publisher = {PMLR}, pdf = {https://proceedings.mlr.press/v151/bordt22a/bordt22a.pdf}, url = {https://proceedings.mlr.press/v151/bordt22a.html}, abstract = { Applications of machine learning inform human decision makers in a broad range of tasks. The resulting problem is usually formulated in terms of a single decision maker. We argue that it should rather be described as a two-player learning problem where one player is the machine and the other the human. While both players try to optimize the final decision, the setup is often characterized by (1) the presence of private information and (2) opacity, that is imperfect understanding between the decision makers. We prove that both properties can complicate decision making considerably. A lower bound quantifies the worst-case hardness of optimally advising a decision maker who is opaque or has access to private information. An upper bound shows that a simple coordination strategy is nearly minimax optimal. More efficient learning is possible under certain assumptions on the problem, for example that both players learn to take actions independently. Such assumptions are implicit in existing literature, for example in medical applications of machine learning, but have not been described or justified theoretically. } }
Endnote
%0 Conference Paper %T A Bandit Model for Human-Machine Decision Making with Private Information and Opacity %A Sebastian Bordt %A Ulrike Von Luxburg %B Proceedings of The 25th International Conference on Artificial Intelligence and Statistics %C Proceedings of Machine Learning Research %D 2022 %E Gustau Camps-Valls %E Francisco J. R. Ruiz %E Isabel Valera %F pmlr-v151-bordt22a %I PMLR %P 7300--7319 %U https://proceedings.mlr.press/v151/bordt22a.html %V 151 %X Applications of machine learning inform human decision makers in a broad range of tasks. The resulting problem is usually formulated in terms of a single decision maker. We argue that it should rather be described as a two-player learning problem where one player is the machine and the other the human. While both players try to optimize the final decision, the setup is often characterized by (1) the presence of private information and (2) opacity, that is imperfect understanding between the decision makers. We prove that both properties can complicate decision making considerably. A lower bound quantifies the worst-case hardness of optimally advising a decision maker who is opaque or has access to private information. An upper bound shows that a simple coordination strategy is nearly minimax optimal. More efficient learning is possible under certain assumptions on the problem, for example that both players learn to take actions independently. Such assumptions are implicit in existing literature, for example in medical applications of machine learning, but have not been described or justified theoretically.
APA
Bordt, S. & Von Luxburg, U.. (2022). A Bandit Model for Human-Machine Decision Making with Private Information and Opacity . Proceedings of The 25th International Conference on Artificial Intelligence and Statistics, in Proceedings of Machine Learning Research 151:7300-7319 Available from https://proceedings.mlr.press/v151/bordt22a.html.

Related Material