the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
Review of Deep Reinforcement Learning for Offshore Wind Farm Maintenance Planning
Abstract. Offshore wind farms face unique challenges in maintenance due to harsh weather, remote locations, and complex logistics. Traditional maintenance strategies often fail to optimize operations, leading to unplanned failures or unnecessary servicing. In recent years, Deep Reinforcement Learning (DRL) has shown clear potential to tackle these challenges through a data-driven approach. This paper provides a critical review of representative DRL models for offshore wind farm maintenance planning, elaborating on both single- and multi-agent frameworks, diverse training algorithms, various problem formulations, and the integration of domain-specific knowledge. The review compares the benefits and limitations of these methods, identifying a significant gap in the widely adopted use of simplistic binary maintenance decisions, rather than including multi-level or imperfect repairs in the action space. This work finally suggests directions for future research, to overcome current limitations and enhance the applicability of DRL methods in offshore wind maintenance.
- Preprint
(1172 KB) - Metadata XML
- BibTeX
- EndNote
Status: final response (author comments only)
- RC1: 'Comment on wes-2025-222', Anonymous Referee #1, 13 Nov 2025
-
RC2: 'Comment on wes-2025-222', Anonymous Referee #2, 05 Dec 2025
This manuscript reviews studies on the application of Deep Reinforcement Learning (DRL) for offshore wind-farm maintenance planning and highlights the potential benefits of DRL compared to traditional maintenance strategies. The topic of implementing data-driven approaches for operation and maintenance is highly relevant for the cost-effective deployment of offshore wind farms. Also, the overall structure of the manuscript is organized, with the abstract providing an informative summary of the content.
As stated in Line 474, “real-world applications of DRL in offshore wind O&M are still in early stages”, most of the studies reviewed rely solely on simulations. To provide a more balanced perspective, the manuscript should not only highlight the potential of DRL, but also acknowledges the limitations, assumptions, and simplifications inherent in these simulation-based approaches. Also, some discussions lack sufficient evidence and requires further clarification. The following comments are offered with the aim of improving the quality of the manuscript.
Viewed
| HTML | XML | Total | BibTeX | EndNote | |
|---|---|---|---|---|---|
| 193 | 45 | 13 | 251 | 14 | 14 |
- HTML: 193
- PDF: 45
- XML: 13
- Total: 251
- BibTeX: 14
- EndNote: 14
Viewed (geographical distribution)
| Country | # | Views | % |
|---|
| Total: | 0 |
| HTML: | 0 |
| PDF: | 0 |
| XML: | 0 |
- 1
The paper
“Review of Deep Reinforcement Learning for Offshore Wind Farm Maintenance Planning”,
By
Borsotti et al.,
provides a structured and timely overview of how DRL methods can optimise offshore wind operations and maintenance.
The survey spans single-agent, multi-agent, and hybrid formulations, and argues—persuasively—that binary “maintain vs not” actions limit realism, advocating multi-level repairs. It synthesises algorithmic families, problem formulations, and domain knowledge.
Overall, there are all the components for a good document and an effective contribution to the field. Nevertheless, while the paper’s clarity is commendable, its method is less so. The review work should methodically follow a rigorous process, e.g. thr PRISMA guidelines. Thus, the following remarks should be fully assessed before being reconsidered for acceptance