dc.contributor.author | Saldaña Enderica, Carlos Alberto | |
dc.contributor.author | Llata García, José Ramón | |
dc.contributor.author | Torre Ferrero, Carlos | |
dc.contributor.other | Universidad de Cantabria | es_ES |
dc.date.accessioned | 2025-09-09T14:57:29Z | |
dc.date.available | 2025-09-09T14:57:29Z | |
dc.date.issued | 2025-06 | |
dc.identifier.issn | 2218-6581 | |
dc.identifier.uri | https://hdl.handle.net/10902/37095 | |
dc.description.abstract | This study proposes a robust methodology for vibration suppression and trajec tory tracking in rotary flexible-link systems by leveraging guided reinforcement learning (GRL). The approach integrates the twin delayed deep deterministic policy gradient (TD3) algorithm with a linear quadratic regulator (LQR) acting as a guiding controller during training. Flexible-link mechanisms common in advanced robotics and aerospace systems exhibit oscillatory behavior that complicates precise control. To address this, the system is first identified using experimental input-output data from a Quanser® virtual plant, gener ating an accurate state-space representation suitable for simulation-based policy learning. The hybrid control strategy enhances sample efficiency and accelerates convergence by incorporating LQR-generated trajectories during TD3 training. Internally, the TD3 agent benefits from architectural features such as twin critics, delayed policy updates, and target action smoothing, which collectively improve learning stability and reduce overestimation bias. Comparative results show that the guided TD3 controller achieves superior perfor mance in terms of vibration damping, transient response, and robustness, when compared to conventional LQR, fuzzy logic, neural networks, and GA-LQR approaches. Although the controller was validated using a high-fidelity digital twin, it has not yet been deployed on the physical plant. Future work will focus on real-time implementation and structural robustness testing under parameter uncertainty. Overall, this research demonstrates that guided reinforcement learning can yield stable and interpretable policies that comply with classical control criteria, offering a scalable and generalizable framework for intelligent control of flexible mechanical systems. | es_ES |
dc.description.sponsorship | This research was funded by Universidad Estatal Península de Santa Elena, Ecuador, as part of its Academic Improvement Plan. This funding is internal and specific to the university, and no additional external public, commercial, or non-profit funding was received. | es_ES |
dc.format.extent | 28 p. | es_ES |
dc.language.iso | eng | es_ES |
dc.publisher | MPDI | es_ES |
dc.rights | © 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license. | es_ES |
dc.rights.uri | http://creativecommons.org/licenses/by/4.0/ | * |
dc.source | Robotics, 2025, 14(6), 76 | es_ES |
dc.subject.other | Guided reinforcement learning | es_ES |
dc.subject.other | Deep reinforcement learning | es_ES |
dc.subject.other | TD3 | es_ES |
dc.subject.other | Linear quadratic regulator | es_ES |
dc.subject.other | Hybrid control | es_ES |
dc.subject.other | Vibration suppression | es_ES |
dc.subject.other | Flexible link systems | es_ES |
dc.subject.other | Robotics | es_ES |
dc.title | Guided reinforcement learning with twin delayed deep deterministic policy gradient for a rotary flexible-link system | es_ES |
dc.type | info:eu-repo/semantics/article | es_ES |
dc.rights.accessRights | openAccess | es_ES |
dc.identifier.DOI | 10.3390/robotics14060076 | |
dc.type.version | publishedVersion | es_ES |