Deep Reinforcement Learning-Based Control of Stewart Platform With Parametric Simulation in ROS and Gazebo
dc.authorid | Tavakol Aghaei, Vahid/0000-0002-4876-1015 | |
dc.authorid | İkizoğlu, Serhat/0000-0003-2394-7988 | |
dc.authorwosid | Tavakol Aghaei, Vahid/AES-9479-2022 | |
dc.authorwosid | İkizoğlu, Serhat/ABB-1783-2020 | |
dc.contributor.author | Yadavari, Hadi | |
dc.contributor.author | Aghaei, Vahid Tavakol | |
dc.contributor.author | Ikizoglu, Serhat | |
dc.date.accessioned | 2024-05-19T14:46:29Z | |
dc.date.available | 2024-05-19T14:46:29Z | |
dc.date.issued | 2023 | |
dc.department | İstinye Üniversitesi | en_US |
dc.description.abstract | The Stewart platform is an entirely parallel robot with mechanical differences from typical serial robotic manipulators, which has a wide application area ranging from flight and driving simulators to structural test platforms. This work concentrates on learning to control a complex model of the Stewart platform using state-of-the-art deep reinforcement learning (DRL) algorithms. In this regard, to enhance the reliability of the learning performance and to have a test bed capable of mimicking the behavior of the system completely, a precisely designed simulation environment is presented. Therefore, we first design a parametric representation for the kinematics of the Stewart platform in Gazebo and robot operating system (ROS) and integrate it with a Python class to conveniently generate the structures in simulation description format (SDF). Then, to control the system, we benefit from three DRL algorithms: the asynchronous advantage actor-critic (A3C), the deep deterministic policy gradient (DDPG), and the proximal policy optimization (PPO) to learn the control gains of a proportional integral derivative (PID) controller for a given reaching task. We chose to apply these algorithms due to the Stewart platform's continuous action and state spaces, making them well-suited for our problem, where exact controller tuning is a crucial task. The simulation results show that the DRL algorithms can successfully learn the controller gains, resulting in satisfactory control performance. | en_US |
dc.identifier.doi | 10.1115/1.4056971 | |
dc.identifier.issn | 1942-4302 | |
dc.identifier.issn | 1942-4310 | |
dc.identifier.issue | 3 | en_US |
dc.identifier.scopus | 2-s2.0-85151427194 | en_US |
dc.identifier.scopusquality | Q1 | en_US |
dc.identifier.uri | https://doi.org10.1115/1.4056971 | |
dc.identifier.uri | https://hdl.handle.net/20.500.12713/5531 | |
dc.identifier.volume | 15 | en_US |
dc.identifier.wos | WOS:000988824300014 | en_US |
dc.identifier.wosquality | N/A | en_US |
dc.indekslendigikaynak | Web of Science | en_US |
dc.indekslendigikaynak | Scopus | en_US |
dc.language.iso | en | en_US |
dc.publisher | Asme | en_US |
dc.relation.ispartof | Journal of Mechanisms and Robotics-Transactions of the Asme | en_US |
dc.relation.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | en_US |
dc.rights | info:eu-repo/semantics/closedAccess | en_US |
dc.snmz | 20240519_ka | en_US |
dc.subject | Reinforcement Learning | en_US |
dc.subject | Deep Learning | en_US |
dc.subject | Control | en_US |
dc.subject | Ros | en_US |
dc.subject | Gazebo | en_US |
dc.subject | Stewart Platform | en_US |
dc.subject | Parallel Platforms | en_US |
dc.title | Deep Reinforcement Learning-Based Control of Stewart Platform With Parametric Simulation in ROS and Gazebo | en_US |
dc.type | Article | en_US |