Energy optimization of wind turbines via a neural control policy based on reinforcement learning Markov chain Monte Carlo algorithm

dc.authoridTavakol Aghaei, Vahid/0000-0002-4876-1015
dc.authoridYıldırım, Sinan/0000-0001-7980-8990
dc.authoridYesilyurt, Serhat/0000-0001-5425-1532
dc.authoridOnat, Ahmet/0000-0003-0491-1074
dc.authorwosidTavakol Aghaei, Vahid/AES-9479-2022
dc.authorwosidYıldırım, Sinan/Y-4290-2019
dc.authorwosidYesilyurt, Serhat/H-9546-2013
dc.contributor.authorAghaei, Vahid Tavakol
dc.contributor.authorAgababaoglu, Arda
dc.contributor.authorBawo, Biram
dc.contributor.authorNaseradinmousavi, Peiman
dc.contributor.authorYildirim, Sinan
dc.contributor.authorYesilyurt, Serhat
dc.contributor.authorOnat, Ahmet
dc.date.accessioned2024-05-19T14:46:28Z
dc.date.available2024-05-19T14:46:28Z
dc.date.issued2023
dc.departmentİstinye Üniversitesien_US
dc.description.abstractThis study focuses on the numerical analysis and optimal control of vertical-axis wind turbines (VAWT) using Bayesian reinforcement learning (RL). We specifically address small-scale wind turbines, which are well-suited to local and compact production of electrical energy on a small scale, such as urban and rural infrastructure installations. Existing literature concentrates on large scale wind turbines which run in unobstructed, mostly constant wind profiles. However urban installations generally must cope with rapidly changing wind patterns. To bridge this gap, we formulate and implement an RL strategy using the Markov chain Monte Carlo (MCMC) algorithm to optimize the long-term energy output of a wind turbine. Our MCMC-based RL algorithm is a model-free and gradient-free algorithm, in which the designer does not have to know the precise dynamics of the plant and its uncertainties. Our method addresses the uncertainties by using a multiplicative reward structure, in contrast with additive reward used in conventional RL approaches. We have shown numerically that the method specifically overcomes the shortcomings typically associated with conventional solutions, including, but not limited to, component aging, modeling errors, and inaccuracies in the estimation of wind speed patterns. Our results show that the proposed method is especially successful in capturing power from wind transients; by modulating the generator load and hence the rotor torque load, so that the rotor tip speed quickly reaches the optimum value for the anticipated wind speed. This ratio of rotor tip speed to wind speed is known to be critical in wind power applications. The wind to load energy efficiency of the proposed method was shown to be superior to two other methods; the classical maximum power point tracking method and a generator controlled by deep deterministic policy gradient (DDPG) method.en_US
dc.identifier.doi10.1016/j.apenergy.2023.121108
dc.identifier.issn0306-2619
dc.identifier.issn1872-9118
dc.identifier.scopus2-s2.0-85152966093en_US
dc.identifier.scopusqualityQ1en_US
dc.identifier.urihttps://doi.org10.1016/j.apenergy.2023.121108
dc.identifier.urihttps://hdl.handle.net/20.500.12713/5530
dc.identifier.volume341en_US
dc.identifier.wosWOS:000984716500001en_US
dc.identifier.wosqualityN/Aen_US
dc.indekslendigikaynakWeb of Scienceen_US
dc.indekslendigikaynakScopusen_US
dc.language.isoenen_US
dc.publisherElsevier Sci Ltden_US
dc.relation.ispartofApplied Energyen_US
dc.relation.publicationcategoryMakale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanıen_US
dc.rightsinfo:eu-repo/semantics/openAccessen_US
dc.snmz20240519_kaen_US
dc.subjectReinforcement Learning (Rl)en_US
dc.subjectWind Turbine Energy Maximizationen_US
dc.subjectNeural Controlen_US
dc.subjectMarkov Chain Monte Carlo (Mcmc)en_US
dc.subjectBayesian Learningen_US
dc.subjectDeep Deterministic Policy Gradient (Ddpg)en_US
dc.titleEnergy optimization of wind turbines via a neural control policy based on reinforcement learning Markov chain Monte Carlo algorithmen_US
dc.typeArticleen_US

Dosyalar