Deep Reinforcement Learning-Based Control Architectures for Autonomous Maritime Renewable Energy Platforms
Abstract
Autonomous vessels driven by renewable energy are increasingly envisioned as vital for sustainable ocean?operations such as environmental monitoring, offshore power generation, and long-haul unmanned surface vehicles. Implementing fine-scale control of these systems has proven challenging however,?due to time-varying sea-state dynamics, sporadic energy inputs, the possibility of failure at the component level, and the requirement for coordination between multiple agents. In the article, an end-to-end deep reinforcement learning-based hierarchical control solution with real-time navigation and?its synthesis for energy optimization is proposed. It combines high-level energy regulation with low-level actuator scheduling so as to react to the variations of?the environment and internal perturbations. Simulations using actual wave realizations, sensor failures, actuator outages, and network communication variation were used?to demonstrate the performance of the control system in the following 5 performance aspects: energy saving, navigation accuracy, communication reliability, fault tolerant and multi-agent coordination. Results indicate that the architecture sustained over 80% of the performance and achieved energy efficiencies up to 54.5% in the?best case under failure scenarios. Performance-measures demonstrated reasonable scalability?up to 5–7 agents without significant communication overhead. The findings support the applicability of deep reinforcement learning for real-time maritime control under uncertainty, offering a viable alternative to conventional rule-based or predictive control strategies. The framework’s modular design allows for future integration with federated learning, hybrid control models, or autonomous deployment. The article contributes to the growing field of intelligent marine systems by providing a robust and adaptable control strategy for sustainable and scalable operations in autonomous maritime environments.
Keywords
References
Wang, Z., et al., Adversarial deep reinforcement learning based robust depth tracking control for underactuated autonomous underwater vehicle. Engineering Applications of Artificial Intelligence, 2024. 130: p. 107728.
Guo, S., et al. An Autonomous Path Planning Model for Unmanned Ships Based on Deep Reinforcement Learning. Sensors, 2020. 20, DOI: 10.3390/s20020426.
Song, D., et al., Guidance and control of autonomous surface underwater vehicles for target tracking in ocean environment by deep reinforcement learning. Ocean Engineering, 2022. 250: p. 110947.
Wang, S., et al., A Data-Driven Multi-Agent Autonomous Voltage Control Framework Using Deep Reinforcement Learning. IEEE Transactions on Power Systems, 2020. 35(6): p. 4644-4654.
Feng, B., et al., Robust federated deep reinforcement learning for optimal control in multiple virtual power plants with electric vehicles. Applied Energy, 2023. 349: p. 121615.
Han, X., et al., An autonomous control technology based on deep reinforcement learning for optimal active power dispatch. International Journal of Electrical Power & Energy Systems, 2023. 145: p. 108686.
Zhang, X., et al. Decision-Making for the Autonomous Navigation of Maritime Autonomous Surface Ships Based on Scene Division and Deep Reinforcement Learning. Sensors, 2019. 19, DOI: 10.3390/s19184055.
Wang, W., et al. Deep Reinforcement Learning Based Tracking Control of an Autonomous Surface Vessel in Natural Waters. in 2023 IEEE International Conference on Robotics and Automation (ICRA). 2023.
Guo, J., et al., Deep Reinforcement Learning-Based Hierarchical Energy Control Strategy of a Platoon of Connected Hybrid Electric Vehicles Through Cloud Platform. IEEE Transactions on Transportation Electrification, 2024. 10(1): p. 305-315.
Li, Q., et al. Review of Deep Reinforcement Learning and Its Application in Modern Renewable Power System Control. Energies, 2023. 16, DOI: 10.3390/en16104143.
Duan, J., et al., Deep-Reinforcement-Learning-Based Autonomous Voltage Control for Power Grid Operations. IEEE Transactions on Power Systems, 2020. 35(1): p. 814-817.
Tu, Z., et al., Development of Deep Reinforcement Learning Co-Simulation Platforms for Power System Control. IEEE Transactions on Automation Science and Engineering, 2025. 22: p. 4780-4789.
Waltz, M. and O. Okhrin, Spatial–temporal recurrent reinforcement learning for autonomous ships. Neural Networks, 2023. 165: p. 634-653.
Wang, Y.L., et al., Networked and Deep Reinforcement Learning-Based Control for Autonomous Marine Vehicles: A Survey. IEEE Transactions on Systems, Man, and Cybernetics: Systems, 2025. 55(1): p. 4-17.
Liu, T., Y. Hu, and H. Xu, Deep Reinforcement Learning for Vectored Thruster Autonomous Underwater Vehicle Control. Complexity, 2021. 2021(1): p. 6649625.
Huang, F., et al., A general motion control architecture for an autonomous underwater vehicle with actuator faults and unknown disturbances through deep reinforcement learning. Ocean Engineering, 2022. 263: p. 112424.
E. D. Lusiana, S. Astutik, Nurjannah, dan A. B. Sambah, “Using Machine Learning Approach to Cluster Marine Environmental Features of Lesser Sunda Island,” J. Appl. Data Sci., vol. 6, no. 1, hal. 247–258, 2025, doi: 10.47738/jads.v6i1.478.
N. Al-Matar, “Optimizing Supply Chain Coordination through Cross-Functional Integration: A Dynamic Model Using Optimal Control Theory,” Int. J. Appl. Inf. Manag., vol. 3, no. 2, hal. 70–81, 2023, doi: 10.47738/ijaim.v3i2.52.
L. Lenus dan A. R. Hananto, “Identifying Regional Hotspots of Gun Violence in the United States Using DBSCAN Clustering,” J. Cyber Law, vol. 1, no. 1, hal. 23–40, 2025.
Politi, E., Stefanidou, A., Chronis, C., Dimitrakopoulos, G., & Varlamis, A., Adaptive Deep Reinforcement Learning for Efficient 3D Navigation of Autonomous Underwater Vehicles. IEEE Access 2024. 12: p. 178209-178221.
Anderlini, E., G.G. Parker, and G. Thomas Docking Control of an Autonomous Underwater Vehicle Using Reinforcement Learning. Applied Sciences, 2019. 9, DOI: 10.3390/app9173456.
D. Mashao dan C. Harley, “Cyber Attack Pattern Analysis Based on Geo-location and Time : A Case Study of Firewall and IDS / IPS Logs,” J. Curr. Res. Blockchain, vol. 2, no. 1, hal. 28–40, 2025, doi: 10.47738/jcrb.v2i1.26.
S. Govindaraju, M. Indirani, S. S. Maidin, dan J. Wei, “Intelligent Transportation System’s Machine Learning-Based Traffic Prediction,” J. Appl. Data Sci., vol. 5, no. 4, hal. 1826–1837, 2024, doi: 10.47738/jads.v5i4.364.
Y. Durachman, “Clustering Student Behavioral Patterns: A Data Mining Approach Using K-Means for Analyzing Study Hours, Attendance, and Tutoring Sessions in Educational Achievement,” Artif. Intell. Learn., vol. 1, no. 1, hal. 35–53, 2025, doi: 10.63913/ail.v1i1.5.
Albarella, N., et al. A Hybrid Deep Reinforcement Learning and Optimal Control Architecture for Autonomous Highway Driving. Energies, 2023. 16, DOI: 10.3390/en16083490.
M.-T. Lai, “Analyzing Company Hiring Patterns Using K-Means Clustering and Association Rule Mining: A Data-Driven Approach to Understanding Recruitment Trends in the Digital Economy,” J. Digit. Soc., vol. 1, no. 1, hal. 20–43, 2025, doi: 10.63913/jds.v1i1.2.
A. D. Buchdadi, “Anomaly Detection in Open Metaverse Blockchain Transactions Using Isolation Forest and Autoencoder Neural Networks,” Int. J. Res. Metaverse, vol. 2, no. 1, hal. 24–51, 2025, doi: 10.47738/ijrm.v2i1.20.
H. T. Sukmana, “Using K-Means Clustering to Enhance Digital Marketing with Flight Ticket Search Patterns,” J. Digit. Mark. Digit. Curr., vol. 1, no. 3, hal. 286–304, 2024, doi: 10.47738/jdmdc.v1i3.22.
Z. Tian, “Investigation into Data Mining for Analysis and Optimization of Direct Maintenance Costs in Civil Aircraft Operations,” IJIIS Int. J. Informatics Inf. Syst., vol. 7, no. 1, hal. 35–43, 2024, doi: 10.47738/ijiis.v7i1.190.
A. S. Samson, N. Sumathi, S. S. Maidin, dan Q. Yang, “Cellular Traffic Prediction Models Using Convolutional Long Short-Term Memory,” J. Appl. Data Sci., vol. 6, no. 1, hal. 20–33, 2025, doi: 10.47738/jads.v6i1.472.
DOI: https://doi.org/10.52088/ijesty.v5i4.1209
Refbacks
- There are currently no refbacks.
Copyright (c) 2025 Sura Sabah, Refat Taleb Hussain, Ismail Abdulaziz Mohammed, Haider Mahmood Jawad, Intesar Abbas, Taqwa Hariguna



























