quotation:[Copy]
Bosen Lian1,Wenqian Xue2,Frank L. Lewis3.[en_title][J].Control Theory and Technology,2023,21(3):281~291.[Copy]
【Print page】 【Online reading】【Download 【PDF Full text】 View/Add CommentDownload reader Close

←Previous page|Page Next →

Back Issue    Advanced search

This Paper:Browse 236   Download 0 本文二维码信息
码上扫一扫!
Heterogeneousmulti-player imitation learning
BosenLian1,WenqianXue2,FrankL.Lewis3
0
(1 Department of Electrical and Computer Engineering, Auburn University, Auburn, AL 36849, USA;2 State Key Laboratory of Synthetical Automation for Process Industries and International Joint Research Laboratory of Integrated Automation, Northeastern University, Shenyang 110819, Liaoning, China;3 UTA Research Institute, University of Texas at Arlington (UTA), Fort Worth, TX 76118, USA)
摘要:
This paper studies imitation learning in nonlinear multi-player game systems with heterogeneous control input dynamics. We propose a model-free data-driven inverse reinforcement learning (RL) algorithm for a leaner to find the cost functions of a N-player Nash expert system given the expert’s states and control inputs. This allows us to address the imitation learning problem without prior knowledge of the expert’s system dynamics. To achieve this, we provide a basic model-based algorithm that is built upon RL and inverse optimal control. This serves as the foundation for our final model-free inverse RL algorithm which is implemented via neural network-based value function approximators. Theoretical analysis and simulation examples verify the methods.
关键词:  Imitation learning · Inverse reinforcement learning · Heterogeneous multi-player games · Data-driven model-free control
DOI:https://doi.org/10.1007/s11768-023-00171-w
基金项目:
Heterogeneousmulti-player imitation learning
Bosen Lian1,Wenqian Xue2,Frank L. Lewis3
(1 Department of Electrical and Computer Engineering, Auburn University, Auburn, AL 36849, USA;2 State Key Laboratory of Synthetical Automation for Process Industries and International Joint Research Laboratory of Integrated Automation, Northeastern University, Shenyang 110819, Liaoning, China;3 UTA Research Institute, University of Texas at Arlington (UTA), Fort Worth, TX 76118, USA)
Abstract:
This paper studies imitation learning in nonlinear multi-player game systems with heterogeneous control input dynamics. We propose a model-free data-driven inverse reinforcement learning (RL) algorithm for a leaner to find the cost functions of a N-player Nash expert system given the expert’s states and control inputs. This allows us to address the imitation learning problem without prior knowledge of the expert’s system dynamics. To achieve this, we provide a basic model-based algorithm that is built upon RL and inverse optimal control. This serves as the foundation for our final model-free inverse RL algorithm which is implemented via neural network-based value function approximators. Theoretical analysis and simulation examples verify the methods.
Key words:  Imitation learning · Inverse reinforcement learning · Heterogeneous multi-player games · Data-driven model-free control