Please use this identifier to cite or link to this item:
https://hdl.handle.net/20.500.11851/11702
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Xue, Wenqian | - |
dc.contributor.author | Lian, Bosen | - |
dc.contributor.author | Kartal, Yusuf | - |
dc.contributor.author | Fan, Jialu | - |
dc.contributor.author | Chai, Tianyou | - |
dc.contributor.author | Lewis, Frank L. | - |
dc.date.accessioned | 2024-08-18T17:23:05Z | - |
dc.date.available | 2024-08-18T17:23:05Z | - |
dc.date.issued | 2024 | - |
dc.identifier.issn | 1545-5955 | - |
dc.identifier.issn | 1558-3783 | - |
dc.identifier.uri | https://doi.org/10.1109/TASE.2024.3427657 | - |
dc.identifier.uri | https://hdl.handle.net/20.500.11851/11702 | - |
dc.description.abstract | This paper proposes a data-driven model-free inverse reinforcement learning (IRL) algorithm tailored for solving an inverse H-infinity control problem. In the problem, both an expert and a learner engage in H-infinity control to reject disturbances and the learner's objective is to imitate the expert's behavior by reconstructing the expert's performance function through IRL techniques. Introducing zero-sum game principles, we first formulate a model-based single-loop IRL policy iteration algorithm that includes three key steps: updating the policy, action, and performance function using a new correction formula and the standard inverse optimal control principles. Building upon the model-based approach, we propose a model-free single-loop off-policy IRL algorithm that eliminates the need for initial stabilizing policies and prior knowledge of the dynamics of expert and learner. Also, we provide rigorous proof of convergence, stability, and Nash optimality to guarantee the effectiveness and reliability of the proposed algorithms. Furthermore, we show-case the efficiency of our algorithm through simulations and experiments, highlighting its advantages compared to the existing methods. | en_US |
dc.description.sponsorship | NSFC [61991404, 62394342, U22A2049]; Liaoning Revitalization Talents Program [XLYC2007135]; Science and Technology Major Project of Liaoning Province [2020JH1, 10100008]; Key Research and Development Program of Liaoning Province [2023JH26, 10200011]; Research Program of the Liaoning Liaohe Laboratory [LLL23ZZ-05-01] | en_US |
dc.description.sponsorship | This work was supported in part by NSFC under Grant 61991404, Grant 62394342, and Grant U22A2049; in part by Liaoning Revitalization Talents Program under Grant XLYC2007135; in part by the 2020 Science and Technology Major Project of Liaoning Province under Grant 2020JH1/10100008; in part by the Key Research and Development Program of Liaoning Province under Grant 2023JH26/10200011; and in part by the Research Program of the Liaoning Liaohe Laboratory under Grant LLL23ZZ-05-01. | en_US |
dc.language.iso | en | en_US |
dc.publisher | Ieee-Inst Electrical Electronics Engineers Inc | en_US |
dc.relation.ispartof | Ieee Transactions on Automation Science and Engineering | en_US |
dc.rights | info:eu-repo/semantics/closedAccess | en_US |
dc.subject | Game theory | en_US |
dc.subject | Games | en_US |
dc.subject | Trajectory | en_US |
dc.subject | Cost function | en_US |
dc.subject | Mathematical models | en_US |
dc.subject | Reinforcement learning | en_US |
dc.subject | Optimal control | en_US |
dc.subject | Inverse reinforcement learning | en_US |
dc.subject | inverse H-infinity control | en_US |
dc.subject | reinforcement learning | en_US |
dc.subject | zero-sum games | en_US |
dc.subject | imitation learning | en_US |
dc.title | Model-Free Inverse H-Infinity Control for Imitation Learning | en_US |
dc.type | Article | en_US |
dc.type | Article; Early Access | en_US |
dc.department | TOBB ETÜ | en_US |
dc.identifier.wos | WOS:001279014600001 | en_US |
dc.institutionauthor | … | - |
dc.identifier.doi | 10.1109/TASE.2024.3427657 | - |
dc.relation.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | en_US |
item.openairetype | Article | - |
item.openairetype | Article; Early Access | - |
item.languageiso639-1 | en | - |
item.grantfulltext | none | - |
item.fulltext | No Fulltext | - |
item.openairecristype | http://purl.org/coar/resource_type/c_18cf | - |
item.openairecristype | http://purl.org/coar/resource_type/c_18cf | - |
item.cerifentitytype | Publications | - |
item.cerifentitytype | Publications | - |
Appears in Collections: | WoS İndeksli Yayınlar Koleksiyonu / WoS Indexed Publications Collection |
CORE Recommender
Items in GCRIS Repository are protected by copyright, with all rights reserved, unless otherwise indicated.