TY - GEN
T1 - A Model-Free Multi-Objective Deep Reinforcement Learning based Controller for Modular Multilevel Converters
AU - Serhan, Abdulrahman
AU - Alquennah, Alamera Nouran
AU - Trabelsi, Mohamed
AU - Ghrayeb, Ali
AU - Zribi, Mohamed
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025
Y1 - 2025
N2 - The increasing deployment of renewable energy systems, electric vehicles, and high-voltage direct current transmission infrastructures has intensified interest in Modular Multilevel Converters (MMCs), which offer superior scalability, waveform quality, and fault tolerance. However, MMC control remains challenging due to its nonlinear dynamics and multi-objective requirements, including output current tracking, capacitor-voltage balancing, and circulating current suppression. This paper presents a model-free control strategy based on Deep Reinforcement Learning (DRL), employing the Proximal Policy Optimization algorithm to achieve these control objectives in a 3-level single-phase MMC. The proposed DRL-based controller learns an optimal switching policy directly from interaction data, eliminating the need for an accurate system model or manual tuning. Simulation results from MATLAB/Simulink confirm that the trained agent achieves low total harmonic distortion, maintains capacitor voltages around the desired values, and minimizes the circulating current, while demonstrating robustness under load variations and dynamic transients. These findings highlight the effectiveness of the proposed DRL approach as a scalable and adaptive solution for complex multilevel inverter control problems.
AB - The increasing deployment of renewable energy systems, electric vehicles, and high-voltage direct current transmission infrastructures has intensified interest in Modular Multilevel Converters (MMCs), which offer superior scalability, waveform quality, and fault tolerance. However, MMC control remains challenging due to its nonlinear dynamics and multi-objective requirements, including output current tracking, capacitor-voltage balancing, and circulating current suppression. This paper presents a model-free control strategy based on Deep Reinforcement Learning (DRL), employing the Proximal Policy Optimization algorithm to achieve these control objectives in a 3-level single-phase MMC. The proposed DRL-based controller learns an optimal switching policy directly from interaction data, eliminating the need for an accurate system model or manual tuning. Simulation results from MATLAB/Simulink confirm that the trained agent achieves low total harmonic distortion, maintains capacitor voltages around the desired values, and minimizes the circulating current, while demonstrating robustness under load variations and dynamic transients. These findings highlight the effectiveness of the proposed DRL approach as a scalable and adaptive solution for complex multilevel inverter control problems.
KW - Modular Multilevel Converter
KW - Multilevel Inverter
KW - Reinforcement Learning
UR - https://www.scopus.com/pages/publications/105024684046
U2 - 10.1109/IECON58223.2025.11221152
DO - 10.1109/IECON58223.2025.11221152
M3 - Conference contribution
AN - SCOPUS:105024684046
T3 - IECON Proceedings (Industrial Electronics Conference)
BT - IECON 2025 - 51st Annual Conference of the IEEE Industrial Electronics Society
PB - IEEE Computer Society
T2 - 51st Annual Conference of the IEEE Industrial Electronics Society, IECON 2025
Y2 - 14 October 2025 through 17 October 2025
ER -