Imitation Learning Based on Disentangled Representation Learning of Behavioral Characteristics

Ryoga Oishi, Sho Sakaino, Toshiaki Tsuji
Proceedings of The 9th Conference on Robot Learning, PMLR 305:4625-4640, 2025.

Abstract

In the field of robot learning, it is becoming possible to coordinate robot action through language instructions. On the other hand, it is still a difficult task to adjust the action based on human instructions because human instructions are often qualitative, and there are cases where there is no one-to-one correspondence between the behavior and the instructions. In this paper, we propose a motion generation model that can adjust actions in response to qualitative human instructions during task execution. The core of the proposed method is a learning architecture that maps qualitative human instructions to actions. Specifically, the demonstration is divided into short action sequences, and labels reflecting human qualitative senses are assigned to these sequences to realize learning that links human qualitative instructions and robot actions. In evaluation experiments, we verified the effectiveness of the method in two tasks: a pick-and-place task and a wiping task. Experimental results showed that the proposed method is able to generate motions in response to human qualitative instructions during task execution, whereas the conventional method generates trajectories all at once, making it impossible to adjust motions during task execution.

Cite this Paper


BibTeX
@InProceedings{pmlr-v305-oishi25a, title = {Imitation Learning Based on Disentangled Representation Learning of Behavioral Characteristics}, author = {Oishi, Ryoga and Sakaino, Sho and Tsuji, Toshiaki}, booktitle = {Proceedings of The 9th Conference on Robot Learning}, pages = {4625--4640}, year = {2025}, editor = {Lim, Joseph and Song, Shuran and Park, Hae-Won}, volume = {305}, series = {Proceedings of Machine Learning Research}, month = {27--30 Sep}, publisher = {PMLR}, pdf = {https://raw.githubusercontent.com/mlresearch/v305/main/assets/oishi25a/oishi25a.pdf}, url = {https://proceedings.mlr.press/v305/oishi25a.html}, abstract = {In the field of robot learning, it is becoming possible to coordinate robot action through language instructions. On the other hand, it is still a difficult task to adjust the action based on human instructions because human instructions are often qualitative, and there are cases where there is no one-to-one correspondence between the behavior and the instructions. In this paper, we propose a motion generation model that can adjust actions in response to qualitative human instructions during task execution. The core of the proposed method is a learning architecture that maps qualitative human instructions to actions. Specifically, the demonstration is divided into short action sequences, and labels reflecting human qualitative senses are assigned to these sequences to realize learning that links human qualitative instructions and robot actions. In evaluation experiments, we verified the effectiveness of the method in two tasks: a pick-and-place task and a wiping task. Experimental results showed that the proposed method is able to generate motions in response to human qualitative instructions during task execution, whereas the conventional method generates trajectories all at once, making it impossible to adjust motions during task execution.} }
Endnote
%0 Conference Paper %T Imitation Learning Based on Disentangled Representation Learning of Behavioral Characteristics %A Ryoga Oishi %A Sho Sakaino %A Toshiaki Tsuji %B Proceedings of The 9th Conference on Robot Learning %C Proceedings of Machine Learning Research %D 2025 %E Joseph Lim %E Shuran Song %E Hae-Won Park %F pmlr-v305-oishi25a %I PMLR %P 4625--4640 %U https://proceedings.mlr.press/v305/oishi25a.html %V 305 %X In the field of robot learning, it is becoming possible to coordinate robot action through language instructions. On the other hand, it is still a difficult task to adjust the action based on human instructions because human instructions are often qualitative, and there are cases where there is no one-to-one correspondence between the behavior and the instructions. In this paper, we propose a motion generation model that can adjust actions in response to qualitative human instructions during task execution. The core of the proposed method is a learning architecture that maps qualitative human instructions to actions. Specifically, the demonstration is divided into short action sequences, and labels reflecting human qualitative senses are assigned to these sequences to realize learning that links human qualitative instructions and robot actions. In evaluation experiments, we verified the effectiveness of the method in two tasks: a pick-and-place task and a wiping task. Experimental results showed that the proposed method is able to generate motions in response to human qualitative instructions during task execution, whereas the conventional method generates trajectories all at once, making it impossible to adjust motions during task execution.
APA
Oishi, R., Sakaino, S. & Tsuji, T.. (2025). Imitation Learning Based on Disentangled Representation Learning of Behavioral Characteristics. Proceedings of The 9th Conference on Robot Learning, in Proceedings of Machine Learning Research 305:4625-4640 Available from https://proceedings.mlr.press/v305/oishi25a.html.

Related Material