<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing with OASIS Tables v3.0 20080202//EN" "https://jats.nlm.nih.gov/nlm-dtd/publishing/3.0/journalpub-oasis3.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:oasis="http://docs.oasis-open.org/ns/oasis-exchange/table" xml:lang="en" dtd-version="3.0" article-type="research-article">
  <front>
    <journal-meta><journal-id journal-id-type="publisher">WES</journal-id><journal-title-group>
    <journal-title>Wind Energy Science</journal-title>
    <abbrev-journal-title abbrev-type="publisher">WES</abbrev-journal-title><abbrev-journal-title abbrev-type="nlm-ta">Wind Energ. Sci.</abbrev-journal-title>
  </journal-title-group><issn pub-type="epub">2366-7451</issn><publisher>
    <publisher-name>Copernicus Publications</publisher-name>
    <publisher-loc>Göttingen, Germany</publisher-loc>
  </publisher></journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.5194/wes-11-1251-2026</article-id><title-group><article-title>Biases in preconstruction estimates of wind plant annual energy production</article-title><alt-title>Biases in preconstruction estimates of wind plant annual energy production</alt-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author" corresp="yes" rid="aff1">
          <name><surname>Hammond</surname><given-names>Rob</given-names></name>
          <email>rob.hammond@nlr.gov</email>
        <ext-link>https://orcid.org/0000-0003-4476-6406</ext-link></contrib>
        <contrib contrib-type="author" corresp="no" rid="aff1">
          <name><surname>Simley</surname><given-names>Eric</given-names></name>
          
        <ext-link>https://orcid.org/0000-0002-1027-9848</ext-link></contrib>
        <aff id="aff1"><label>1</label><institution>National Wind Technology Center, National Laboratory of the Rockies, Golden, CO 80401, USA</institution>
        </aff>
      </contrib-group>
      <author-notes><corresp id="corr1">Rob Hammond (rob.hammond@nlr.gov)</corresp></author-notes><pub-date><day>17</day><month>April</month><year>2026</year></pub-date>
      
      <volume>11</volume>
      <issue>4</issue>
      <fpage>1251</fpage><lpage>1265</lpage>
      <history>
        <date date-type="received"><day>15</day><month>July</month><year>2025</year></date>
           <date date-type="rev-request"><day>8</day><month>September</month><year>2025</year></date>
           <date date-type="rev-recd"><day>14</day><month>January</month><year>2026</year></date>
           <date date-type="accepted"><day>16</day><month>February</month><year>2026</year></date>
      </history>
      <permissions>
        <copyright-statement>Copyright: © 2026 Rob Hammond</copyright-statement>
        <copyright-year>2026</copyright-year>
      <license license-type="open-access"><license-p>This work is licensed under the Creative Commons Attribution 4.0 International License. To view a copy of this licence, visit <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link></license-p></license></permissions><self-uri xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026.html">This article is available from https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026.html</self-uri><self-uri xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026.pdf">The full text article is available as a PDF file from https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026.pdf</self-uri>
      <abstract><title>Abstract</title>

      <p id="d2e88">Estimating the energy yield of a wind plant during the preconstruction phase is a historically difficult task, even with industry improvements in these estimations. We build on prior research comparing the realized energy production of wind plants and their estimated annual energy production P50 values (median energy production), using owner-provided energy production and losses. We produced similar results to prior studies but with a slightly increasing bias of overestimating median energy production (a bias between realized and estimated energy production of <inline-formula><mml:math id="M1" display="inline"><mml:mo>-</mml:mo></mml:math></inline-formula>7.4 % to <inline-formula><mml:math id="M2" display="inline"><mml:mo>-</mml:mo></mml:math></inline-formula>6.6 %, depending on the scenario, as opposed to <inline-formula><mml:math id="M3" display="inline"><mml:mo>-</mml:mo></mml:math></inline-formula>6.7 % to <inline-formula><mml:math id="M4" display="inline"><mml:mo>-</mml:mo></mml:math></inline-formula>5.5 % from earlier studies). In addition to assessing annual energy production P50 bias, we compared both the 1-year and the long-term annual energy production P90 and uncertainty energy yield assessment estimates to the observed long-term-corrected energy production. We found that neither the energy yield assessment uncertainty nor the P90 is conservative enough compared to the observed distribution of prediction errors, suggesting significant room for improvement in the energy yield assessment process.</p>
  </abstract>
    
<funding-group>
<award-group id="gs1">
<funding-source>U.S. Department of Energy</funding-source>
<award-id>DE-AC36-08GO28308</award-id>
</award-group>
</funding-group>
</article-meta>
  <notes notes-type="copyrightstatement">
  
      <p id="d2e126">The U.S. Government retains and the publisher, by accepting the article for publication, acknowledges that the U.S. Government retains a nonexclusive, paid-up, irrevocable, worldwide license to publish or reproduce the published form of this work, or allow others to do so, for U.S. Government purposes.</p>
</notes></front>
<body>
      


<sec id="Ch1.S1" sec-type="intro">
  <label>1</label><title>Introduction</title>
      <p id="d2e137">The probability of exceeding varying thresholds of annual energy production (AEP) forms the basis of investment risk for wind power plants. AEP overestimation increases the risk of financial losses for both investors and owners <xref ref-type="bibr" rid="bib1.bibx2" id="paren.1"/>. As such, the net energy estimate includes many sources of potential uncertainty, including but not limited to estimations of gross energy, wakes, electrical losses, turbine performance, environmental factors, and curtailment <xref ref-type="bibr" rid="bib1.bibx2" id="paren.2"/>. Every input into the net AEP estimation from gross energy to curtailment is modeled as a distribution with its own uncertainties caused by factors such as measurement device calibration and model uncertainty to create a Gaussian distribution of expected energy production. The AEP P50 is then the mean energy production, and the P90 is the <inline-formula><mml:math id="M5" display="inline"><mml:mrow><mml:mi>P</mml:mi><mml:mn mathvariant="normal">50</mml:mn><mml:mo>-</mml:mo><mml:mn mathvariant="normal">2.32</mml:mn><mml:mi mathvariant="italic">σ</mml:mi></mml:mrow></mml:math></inline-formula> <xref ref-type="bibr" rid="bib1.bibx2" id="paren.3"/>, where <inline-formula><mml:math id="M6" display="inline"><mml:mi mathvariant="italic">σ</mml:mi></mml:math></inline-formula> is the standard deviation and overall uncertainty. The 20-year (long-term) P90 is typically the basis for determining the project's financial viability and therefore determines if the project's sponsor will maintain an interest in owning that project <xref ref-type="bibr" rid="bib1.bibx2" id="paren.4"/>. Additional metrics, such as the 10-year P90 and the 1-year P99, help inform other classes of investors about the likelihood of default on debt obligations or the applicability of tax credits <xref ref-type="bibr" rid="bib1.bibx2" id="paren.5"/>; therefore, it is essential to properly quantify the major sources of risk and uncertainty.</p>
      <p id="d2e179">In an operational analysis, AEP is similarly estimated in terms of its P50 and P90 values; however, these values correspond to their inverse percentiles of the data. Accordingly, the P90 represents the 10th percentile of energy production, so there is a 90 % chance of exceeding that level of energy production. The P50 corresponds to the 50th percentile or median energy production. The uncertainty in the AEP is also considered from a short-term perspective (1-year variability) and long-term perspective (variability in average AEP over 10<inline-formula><mml:math id="M7" display="inline"><mml:mo>+</mml:mo></mml:math></inline-formula> or 20 years). In this study, we consider the P50, short- and long-term AEP P90, and short- and long-term AEP uncertainty.</p>
      <p id="d2e189">In an initial study, it was suggested that wind power plant preconstruction energy yield assessments (EYAs) overestimate actual wind production by 5.5 % to 6.7 %, with an assumed 1 %–2 % reduction in overestimation if the authors were able to account for unreported curtailment and availability losses <xref ref-type="bibr" rid="bib1.bibx8" id="paren.6"/>. In <xref ref-type="bibr" rid="bib1.bibx8" id="text.7"/>, the authors relied on consultant-provided EYA data combined with publicly available net energy production from the United States Energy Information Administration (EIA). This rendered an analysis of 62 United States-based wind plants, primarily located in the southern and central United States, with commercial operation dates (CODs) between 2008 and 2016 (23 of which are post-2010).</p>
      <p id="d2e198">In a more recent study, it was suggested that  bias has been decreasing over time in the wind industry, with preconstruction estimates overpredicting energy yield by only 1 %–2 % <xref ref-type="bibr" rid="bib1.bibx7" id="paren.8"/>. However, the uncertainty in EYA prediction accuracy remains high, with a 6.6 % standard deviation of the mean bias <xref ref-type="bibr" rid="bib1.bibx7" id="paren.9"/>. This study is a meta-analysis, aggregating the results from presentations in industry and academic conferences, technical reports, white papers, and peer-reviewed articles primarily in North America and Europe. As such, the assumptions for each data point are difficult to track, making it hard to directly compare to <xref ref-type="bibr" rid="bib1.bibx8" id="text.10"/>. To better understand EYA biases and characterize uncertainty in the EYA process, we build on the methods and findings of <xref ref-type="bibr" rid="bib1.bibx8" id="text.11"/>.</p>
      <p id="d2e214">In this study, we compared the monthly gross energy production and the curtailment and availability losses obtained from wind plant owners and operators with the consultant-provided preconstruction EYA estimates for the long-term AEP. Similar to <xref ref-type="bibr" rid="bib1.bibx8" id="text.12"/>, we also focus on projects primarily in the United States and those with a COD later than 2010, when the industry improved existing methods and data quality control for wind energy resource assessments used for EYA reporting <xref ref-type="bibr" rid="bib1.bibx3" id="paren.13"/>. There are several contributions this work makes to the literature. We improve on the data in <xref ref-type="bibr" rid="bib1.bibx8" id="text.14"/> by using owner-reported energy production, curtailment loss, and availability loss data and by updating the analysis for an additional 7 years of data. Further, we improve on the methodology in <xref ref-type="bibr" rid="bib1.bibx8" id="text.15"/> by accounting for the availability and curtailment losses in the long-term AEP estimation. Finally, we benchmark the preconstruction AEP uncertainty estimates against the spread of AEP prediction errors.</p>
      <p id="d2e229">This paper is organized as follows. The next section covers the data collection and analysis methodology. The following section provides a breakdown of the results, including a comparison of the data from <xref ref-type="bibr" rid="bib1.bibx8" id="text.16"/> to this study and comparisons of the operational AEP to EYA P50, P90, and uncertainty estimates. Finally, we conclude with a discussion of the results and future directions for this research.</p>
</sec>
<sec id="Ch1.S2">
  <label>2</label><title>Methodology</title>
      <p id="d2e243">The operational analysis P50 and P90 estimates represent the AEP that the project is expected to exceed 50 % or 90 % of the time, respectively, over the lifetime of the wind plant, and the uncertainty is the standard deviation divided by the mean (<inline-formula><mml:math id="M8" display="inline"><mml:mrow><mml:mi mathvariant="italic">σ</mml:mi><mml:mo>/</mml:mo><mml:mi mathvariant="italic">μ</mml:mi></mml:mrow></mml:math></inline-formula>). The P50 is then the 50th percentile (median) of energy production, and P90 is the 10th percentile of energy production. For an EYA, the AEP is modeled as a Gaussian distribution, where the P50 is the mean, and uncertainty is the standard deviation (<inline-formula><mml:math id="M9" display="inline"><mml:mi mathvariant="italic">σ</mml:mi></mml:math></inline-formula>) with the P90 calculated as <inline-formula><mml:math id="M10" display="inline"><mml:mrow><mml:mi>P</mml:mi><mml:mn mathvariant="normal">50</mml:mn><mml:mo>-</mml:mo><mml:mn mathvariant="normal">2.32</mml:mn><mml:mi mathvariant="italic">σ</mml:mi></mml:mrow></mml:math></inline-formula> <xref ref-type="bibr" rid="bib1.bibx2" id="paren.17"/>. For all comparisons between EYA and operational analyses, we compared the relative uncertainty (<inline-formula><mml:math id="M11" display="inline"><mml:mrow><mml:mi mathvariant="italic">σ</mml:mi><mml:mo>/</mml:mo><mml:mi mathvariant="italic">μ</mml:mi></mml:mrow></mml:math></inline-formula>) or standard deviation (<inline-formula><mml:math id="M12" display="inline"><mml:mi mathvariant="italic">σ</mml:mi></mml:math></inline-formula>) where appropriate. For the P90 and uncertainty estimations, we compared both the short-term and the long-term values. The short-term estimates were considered to have a 1-year variability. The long-term estimates were provided as the variability in average production over a 10-year, 10-or-more-year, or 20-year period, so our long-term comparisons are a mixture of these values. For EYAs with multiple values, we chose the 20-year estimate to align with the analysis period. We also compared the net energy production from the monthly operating reports (MORs) to the results in <xref ref-type="bibr" rid="bib1.bibx8" id="text.18"/> using EIA monthly energy production data for wind plants in the United States.</p>
<sec id="Ch1.S2.SS1">
  <label>2.1</label><title>Wind plants</title>
      <p id="d2e314">We collected preconstruction EYAs or key data from them over the course of multiple years. We built on the work of <xref ref-type="bibr" rid="bib1.bibx8" id="text.19"/> by working with many of the same wind plant owners and operators who provided EYA estimates to collect plant-level MOR containing monthly gross energy production, curtailment losses, and availability losses. Our data collection spanned MORs for 94 wind plants in six countries, covering 115 EYAs for six wind plant owners and 19 specified consultants. Of those, 76 were in the United States (72 with corresponding EIA data), and 70 had a COD of 2011 or later – a reversed COD bias from <xref ref-type="bibr" rid="bib1.bibx8" id="text.20"/> where only 23 of the 56 plants had a COD of 2011 or later. In total, we collected 707 complete wind plant years of data to conduct this study.</p>
      <p id="d2e323">Figure <xref ref-type="fig" rid="F1"/> shows a variety of key project characteristics collected from both the EYAs and the MORs and their distributions to highlight the breadth of  plants represented in this study. The collected EYA data were provided as either the full document or as a data sheet of the essential information, so the project metadata were not universally available, meaning there are fewer results for some variables than others. There are some projects with CODs prior to 2010, but most plants came online between 2012 and 2017. Additionally, the provided project MORs tended to contain 3.5 years or between 8 and 10 years of operational data as opposed to 2 years or less, providing sufficient data for modeling.</p>

      <fig id="F1" specific-use="star"><label>Figure 1</label><caption><p id="d2e330">Summary of the key characteristics of the wind plants that have both  EYA and monthly operating data. From top to bottom: the boxplots show the distribution of the project's nameplate capacity, COD, number of turbines, and turbine capacity, where that information is available in the EYA data.</p></caption>
          <graphic xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026-f01.png"/>

        </fig>

      <fig id="F2" specific-use="star"><label>Figure 2</label><caption><p id="d2e342">Histogram of the correlation between the monthly  gross energy production and each of the monthly MERRA-2 50 m wind speed (top) and ERA5 100 m wind speed (bottom) for each wind plant for the nearest available coordinate.</p></caption>
          <graphic xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026-f02.png"/>

        </fig>

      <p id="d2e351">For each wind plant, we also collected the ERA5 <xref ref-type="bibr" rid="bib1.bibx5" id="paren.21"/> and MERRA-2 <xref ref-type="bibr" rid="bib1.bibx4" id="paren.22"/> reanalysis datasets for  long-term wind resource corrections. The MERRA-2 data are made available at 0.5° <inline-formula><mml:math id="M13" display="inline"><mml:mo>×</mml:mo></mml:math></inline-formula> 0.625° resolution, and the ERA5 data are available at 0.25° resolution, so we identified the closest latitude and longitude to each plant's owner-provided centroid and used the atmospheric data corresponding to that point. For each point, we downloaded the hourly wind speed, wind direction, temperature, humidity, and surface air pressure and aggregated these readings to the monthly mean reading for each variable. Figure <xref ref-type="fig" rid="F2"/> demonstrates the high degree of correlation between gross energy production (discussed further in Sect. <xref ref-type="sec" rid="Ch1.S2.SS2"/>) and the MERRA-2 and ERA5 wind speed data as measured by the Pearson correlation coefficient. While MERRA-2 demonstrates a higher average Pearson correlation coefficient (0.83) than the ERA5 data (0.80), their median values show a reverse trend, with ERA5 having a 0.90 median correlation and MERRA-2 having a 0.88 median correlation. The close alignment of reanalysis wind speed and gross energy suggests that for most projects, the reanalysis wind conditions are suitable for analysis.</p>
      <p id="d2e372">We found that the five projects with a correlation coefficient less than 0.4 represented our most extreme outliers in the operational analysis results, so we removed them from the study. We also found that each of these projects either had upstream  plants introduced following their construction, causing external wake losses, or is located in complex terrain, such as downwind from a mountainous region or among rolling hills. Further, reanalysis wind conditions were not able to fully account for complex terrain near  ground level, as they are modeled from atmospheric conditions, nor could they account for every location within their coarse-grid structure. While a coarse grid means the reanalysis conditions are more or less representative of the data for that plant, there is no correlation between distance from the modeled grid point and the plant's centroid for either the poorly matched plants or the data set writ large.</p>
</sec>
<sec id="Ch1.S2.SS2">
  <label>2.2</label><title>Operational analysis</title>
      <p id="d2e383">The first set of analyses assessed the similarities between the P50 biases using the annual net energy provided through the EIA and MORs. To annualize the energy production, we summed each consecutive 12-month period starting from the COD, excluding any partial years at the end of the time series. We then calculated the percentage difference between each annualized reading and the EYA-estimated P50.</p>

      <fig id="F3" specific-use="star"><label>Figure 3</label><caption><p id="d2e388">Flowchart of the OpenOA Monte Carlo AEP methodology.</p></caption>
          <graphic xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026-f03.png"/>

        </fig>

      <p id="d2e397">To compute the long-term-corrected P50, P90, and uncertainty estimates, we utilized the Monte Carlo AEP analysis method in the National Laboratory of the Rockies' OpenOA software <xref ref-type="bibr" rid="bib1.bibx9" id="paren.23"/>, as described in <xref ref-type="bibr" rid="bib1.bibx1" id="text.24"/>, outlined below, and shown as a flowchart in Fig. <xref ref-type="fig" rid="F3"/>. The method includes the following steps:</p>
      <p id="d2e409"><list list-type="order">
            <list-item>

      <p id="d2e414">Compute the density-corrected wind speeds for each reanalysis product, and average the hourly wind speed, wind direction, and temperature to a monthly level to align with the operating data.</p>
            </list-item>
            <list-item>

      <p id="d2e420">Flag months where the losses were unreported, and remove months where no energy data were available.</p>
            </list-item>
            <list-item>

      <p id="d2e426">Calculate gross energy by adding the revenue meter energy, availability loss, and curtailment loss, and normalize each month to a 30 d period.</p>
            </list-item>
            <list-item>

      <p id="d2e432">Run a linear regression between the gross energy production and density-corrected wind speeds, the sine and cosine components of the wind direction, and temperature.</p>
            </list-item>
            <list-item>

      <p id="d2e438">Apply the linear regression to the long-term, density-corrected wind speed, wind direction, and temperature data.</p>
            </list-item>
            <list-item>

      <p id="d2e445">Correct the long-term gross energy estimates, which are based on 30 d months, for the correct number of days in each month.</p>
            </list-item>
            <list-item>

      <p id="d2e451">Subtract the monthly average availability losses to estimate monthly net energy.</p>
            </list-item>
            <list-item>

      <p id="d2e457">For analyses comparing the 1-year uncertainty, include the interannual variability (IAV) uncertainty factor, and for analyses compared to the long-term uncertainty, exclude the IAV. The IAV uncertainty component is sampled from a normally distributed random variable, centered on the 20-year standard deviation of the energy production estimates.</p>
            </list-item>
          </list></p>
      <p id="d2e462">The long-term correction is similar to the approach taken in <xref ref-type="bibr" rid="bib1.bibx8" id="text.25"/>, with a few essential differences in the data used and how they are filtered. The reanalysis datasets were of a different vintage, and only the wind speed data were used in <xref ref-type="bibr" rid="bib1.bibx8" id="text.26"/>. No curtailment or availability data are provided through the EIA, so there was no gross energy calculation, nor was a 30 d normalization applied in <xref ref-type="bibr" rid="bib1.bibx8" id="text.27"/>. The outlier filtering in <xref ref-type="bibr" rid="bib1.bibx8" id="text.28"/> was based on the standard deviation from the mean, different from the more robust approach taken in this study.</p>
      <p id="d2e477">For each iteration in the Monte Carlo analysis, we randomized the usage of the variables listed in the top section of Table <xref ref-type="table" rid="T1"/>. The bottom section represents variables that are varied between analyses but held constant within an analysis. Additionally, we defined what OpenOA uses as a default value, how each simulation of the Monte Carlo analysis applies that value, and what variations of the Monte Carlo analysis we ran. For the simulation's parameter choice, OpenOA randomly chooses one of the input variables, randomly samples from a normal distribution using the input value as the distribution's center or a uniform distribution with the input value acting as the lower and upper bound, or does not vary the parameter's value at all between runs. The choice for the default values was not documented when the model was originally built; however, it is expected that the uncertainty in energy losses will be higher than the net energy, as the losses are derived based on the estimated potential energy production rather than directly measured at the revenue meter. For the combined loss threshold parameter, we found that typically between 4 and 16 months will be removed from an analysis, representing as little as 9 % and as much as 28 % of the available records. Despite removing many records for some plants, this filtering significantly reduced the overall long-term-corrected uncertainty. We used the Huber loss method because it is more robust to outlier data <xref ref-type="bibr" rid="bib1.bibx6" id="paren.29"/>, such as high curtailment or low-availability months, which are not representative of regular plant operations. For each combination of model configurations, we ran 5000 Monte Carlo iterations, including bootstrapping the monthly operational data and randomly sampling different analysis parameters to understand the uncertainty in the operational assessments for each plant. The rightmost column of Table <xref ref-type="table" rid="T1"/> represents the input parameter variations that we ran to identify the analysis settings that yield the most accurate and reliable estimates.</p>

<table-wrap id="T1" specific-use="star"><label>Table 1</label><caption><p id="d2e490">Description of the modeling parameters varied between individual simulations of the Monte Carlo analysis, their corresponding OpenOA default values, and how these parameters are varied between each Monte Carlo analysis.</p></caption><oasis:table frame="topbot"><oasis:tgroup cols="4">
     <oasis:colspec colnum="1" colname="col1" align="justify" colwidth="5.5cm"/>
     <oasis:colspec colnum="2" colname="col2" align="justify" colwidth="1.3cm"/>
     <oasis:colspec colnum="3" colname="col3" align="justify" colwidth="4.3cm"/>
     <oasis:colspec colnum="4" colname="col4" align="justify" colwidth="3.5cm"/>
     <oasis:thead>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1">Variable</oasis:entry>
         <oasis:entry colname="col2">OpenOA default</oasis:entry>
         <oasis:entry colname="col3">Monte Carlo parameter selection</oasis:entry>
         <oasis:entry colname="col4">Analysis variations</oasis:entry>
       </oasis:row>
     </oasis:thead>
     <oasis:tbody>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1">Reanalysis product</oasis:entry>
         <oasis:entry colname="col2">None</oasis:entry>
         <oasis:entry colname="col3">Random choice</oasis:entry>
         <oasis:entry colname="col4">(ERA5, MERRA-2) for all</oasis:entry>
       </oasis:row>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1">Uncertainty percentage applied to the net energy</oasis:entry>
         <oasis:entry colname="col2">0.5 %</oasis:entry>
         <oasis:entry colname="col3">Normal distribution center</oasis:entry>
         <oasis:entry colname="col4">0.5 % for all</oasis:entry>
       </oasis:row>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1">Uncertainty percentage applied to the availability and curtailment losses</oasis:entry>
         <oasis:entry colname="col2">5 %</oasis:entry>
         <oasis:entry colname="col3">Normal distribution center</oasis:entry>
         <oasis:entry colname="col4">5 % for all</oasis:entry>
       </oasis:row>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1">Number of years for the windiness correction</oasis:entry>
         <oasis:entry colname="col2">(10, 20)</oasis:entry>
         <oasis:entry colname="col3">Uniform distribution bounds</oasis:entry>
         <oasis:entry colname="col4">(10, 20) for all</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Total loss percent threshold</oasis:entry>
         <oasis:entry colname="col2">(10, 20)</oasis:entry>
         <oasis:entry colname="col3">Uniform distribution bounds</oasis:entry>
         <oasis:entry colname="col4">(10 %–20 %,</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2"/>
         <oasis:entry colname="col3"/>
         <oasis:entry colname="col4">60–80th percentile<sup>*</sup>,</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2"/>
         <oasis:entry colname="col3"/>
         <oasis:entry colname="col4">60–90th percentile<sup>*</sup>,</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2"/>
         <oasis:entry colname="col3"/>
         <oasis:entry colname="col4">60–100th percentile<sup>*</sup>,</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2"/>
         <oasis:entry colname="col3"/>
         <oasis:entry colname="col4">80–90th percentile<sup>*</sup>,</oasis:entry>
       </oasis:row>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2"/>
         <oasis:entry colname="col3"/>
         <oasis:entry colname="col4">80–100th percentile<sup>*</sup>)</oasis:entry>
       </oasis:row>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1">Huber loss tuning constant</oasis:entry>
         <oasis:entry colname="col2">(1, 3)</oasis:entry>
         <oasis:entry colname="col3">Uniform distribution bounds</oasis:entry>
         <oasis:entry colname="col4">(1, 3) for all</oasis:entry>
       </oasis:row>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1">Use of reanalysis temperature data</oasis:entry>
         <oasis:entry colname="col2">False</oasis:entry>
         <oasis:entry colname="col3">No variation</oasis:entry>
         <oasis:entry colname="col4">(true, false)</oasis:entry>
       </oasis:row>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1">Use of reanalysis wind direction data</oasis:entry>
         <oasis:entry colname="col2">False</oasis:entry>
         <oasis:entry colname="col3">No variation</oasis:entry>
         <oasis:entry colname="col4">(true, false)</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Use of Huber outlier detection</oasis:entry>
         <oasis:entry colname="col2">False</oasis:entry>
         <oasis:entry colname="col3">No variation</oasis:entry>
         <oasis:entry colname="col4">(true, false)</oasis:entry>
       </oasis:row>
     </oasis:tbody>
   </oasis:tgroup></oasis:table><table-wrap-foot><p id="d2e493"><sup>*</sup> denotes that the listed number is a percentile, and the percentage is calculated separately for each project.</p></table-wrap-foot></table-wrap>

      <p id="d2e781">To select a representative model for each of the combinations, we chose the lowest total rank by comparing  the <inline-formula><mml:math id="M20" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> of each bootstrapped model and the total uncertainty (<inline-formula><mml:math id="M21" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">σ</mml:mi><mml:mi mathvariant="normal">AEP</mml:mi></mml:msub><mml:mo>/</mml:mo><mml:msub><mml:mi mathvariant="italic">μ</mml:mi><mml:mi mathvariant="normal">AEP</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula>). Using each of these metrics, we summed the ranking of the lowest total uncertainty and maximum median <inline-formula><mml:math id="M22" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula>. We selected the lowest total rank model as the representative model for each plant, as it would have minimized uncertainty and maximized how well the model fits the test data. Using the 5000 iterations, we computed the P50, P90, and uncertainty of the AEP estimations.</p>
      <p id="d2e825">When computing the results, we also gradually filtered out time periods and projects to arrive at a more accurate assessment of each wind plant and the industry more broadly. We removed the first year of operational data to account for non-representative issues that arise in the first year of operations. Then, we removed projects with a COD earlier than 2011 to account for changes in EYA methodology. Finally, we considered the combination of both filters.</p>
</sec>
</sec>
<sec id="Ch1.S3">
  <label>3</label><title>Results</title>
      <p id="d2e837">Here, we present the results of our validation of the approach taken in <xref ref-type="bibr" rid="bib1.bibx8" id="text.30"/> that formed the basis of our work and our operational analysis performed on owner-provided data. In Sect. <xref ref-type="sec" rid="Ch1.S3.SS1"/>, we compare results using the EIA-reported monthly net energy to results based on the owner-provided MOR data. In Sect. <xref ref-type="sec" rid="Ch1.S3.SS2"/>, we report on operational analysis of the MOR data to assess the bias present in EYA-estimated gross energy production and the uncertainty in the EYA estimates.</p>
<sec id="Ch1.S3.SS1">
  <label>3.1</label><title>Validation of the original approach</title>
      <p id="d2e854">As a first step, we confirmed the validity of the approach to rely on monthly EIA net energy data to determine AEP from the original study <xref ref-type="bibr" rid="bib1.bibx8" id="paren.31"/>. This step ensures that we are able to make a reasonable comparison because there is not a specification for how the EIA data are provided by plant owners or what processing steps (if any) are performed prior to being provided to end users. To do this, we compared the EIA net energy output to the MOR net energy, where the EIA identifier was provided with the EYA data. As shown in Fig. <xref ref-type="fig" rid="F4"/>, the correlation between the MOR energy production received from plant owners and operators and the energy reported by the EIA was strong (<inline-formula><mml:math id="M23" display="inline"><mml:mrow><mml:mi mathvariant="italic">ρ</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">0.98</mml:mn></mml:mrow></mml:math></inline-formula>). The line of best fit for the EIA and MOR data yields <inline-formula><mml:math id="M24" display="inline"><mml:mrow><mml:mi>y</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">0.99</mml:mn><mml:mi>x</mml:mi><mml:mo>+</mml:mo><mml:mn mathvariant="normal">0.30</mml:mn></mml:mrow></mml:math></inline-formula>, indicating that the EIA data are highly consistent with those reported by wind plant owners. This result indicates that the methods used by <xref ref-type="bibr" rid="bib1.bibx8" id="text.32"/> to set a baseline for P50 bias remain valid overall, though significant discrepancies exist for individual months.</p>

      <fig id="F4"><label>Figure 4</label><caption><p id="d2e897">Relationship between EIA-reported monthly net energy and MOR net energy. The line of best fit demonstrates that the EIA net energy closely tracks that of the MORs.</p></caption>
          <graphic xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026-f04.png"/>

        </fig>

      <p id="d2e906">We then compared the EYA AEP P50 bias for both the net EIA energy and the net MOR energy without a long-term correction to the bias resulting from <xref ref-type="bibr" rid="bib1.bibx8" id="text.33"/>, as shown in Fig. <xref ref-type="fig" rid="F5"/>. We took the best-case scenario, where we removed the first operational year of the wind plant and removed plants with a COD prior to 2011. This approach is similar to the approach taken in <xref ref-type="bibr" rid="bib1.bibx8" id="text.34"/>, which reported a 5.5 % overprediction when using a long-term correction. Here, the average P50 biases for the uncorrected EIA and MOR data are <inline-formula><mml:math id="M25" display="inline"><mml:mo>-</mml:mo></mml:math></inline-formula>14.9 % and <inline-formula><mml:math id="M26" display="inline"><mml:mo>-</mml:mo></mml:math></inline-formula>11.5 %, respectively. However, there is a large variation in the bias for individual plants (19.3 % for EIA data and 13.5 % for MORs), with positive biases (underprediction) for many plants. As expected with uncorrected values, there is a significant number of outliers for AEP overprediction, particularly for the EIA data, which has 55 % more wind farm years of data. In general, this finding indicates that there is still an overprediction of AEP P50 in the EYAs, given that it is at least double the long-term-corrected value reported in <xref ref-type="bibr" rid="bib1.bibx8" id="text.35"/>.</p>

      <fig id="F5" specific-use="star"><label>Figure 5</label><caption><p id="d2e938">Comparison of the P50 bias between the EIA net energy and MOR net energy.</p></caption>
          <graphic xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026-f05.png"/>

        </fig>

</sec>
<sec id="Ch1.S3.SS2">
  <label>3.2</label><title>Long-term-corrected results</title>
      <p id="d2e955">The next phase of the analysis was to estimate the long-term-corrected AEP and uncertainty. As described in Sect. <xref ref-type="sec" rid="Ch1.S2.SS2"/>, we found each project's ideal analysis settings; we report on the long-term-corrected results in this section. We found that 84 % of the best combinations included the use of the Huber loss algorithm outlier detection and filtering, 74 % used the reanalysis temperature data, and 33 % used the reanalysis wind direction. For the uncertainty loss thresholds, we found that 21 % of the best combinations used the OpenOA default value, and another 46 % used thresholds with a minimum of the 80th percentile. Combined, 47 % of the best model combinations also used the outlier detection and temperature data but no wind direction data.</p>
      <p id="d2e960">In Fig. <xref ref-type="fig" rid="F6"/>, we demonstrate the overall long-term-corrected P50 bias by COD where there is modest yet inconsistent improvement in the P50 bias between 2011 and 2019 relative to projects prior to 2011. However, in 2019 and later, the P50 bias worsens, though it is worth noting that there are fewer projects and therefore less operational data in our dataset in this COD grouping, so the results may not be representative. Even with later COD P50 bias increases, the post-2010 COD wind plants exhibit an improving P50 bias, <inline-formula><mml:math id="M27" display="inline"><mml:mo>-</mml:mo></mml:math></inline-formula>7.4 % vs. <inline-formula><mml:math id="M28" display="inline"><mml:mo>-</mml:mo></mml:math></inline-formula>6.6 %, as shown in the text box of Fig. <xref ref-type="fig" rid="F6"/>, which is in line with the baseline result trends from <xref ref-type="bibr" rid="bib1.bibx8" id="text.36"/>. To further test whether or not there is a relationship between CODs prior to 2011, we ran a two-sided Welch's <inline-formula><mml:math id="M29" display="inline"><mml:mi>t</mml:mi></mml:math></inline-formula> test <xref ref-type="bibr" rid="bib1.bibx11" id="paren.37"/> for years prior to 2011, 2011 through 2022, and 2011 through 2019 to account for the effect of small sample sizes in projects coming online later against the P50 bias of all samples and those dropping the first year of operational data. For all combinations, a <inline-formula><mml:math id="M30" display="inline"><mml:mi>p</mml:mi></mml:math></inline-formula> value greater than 0.2 was yielded, demonstrating that there is not a strong relationship between the EYA estimation methods prior to 2011 and those used starting in 2011.</p>

      <fig id="F6" specific-use="star"><label>Figure 6</label><caption><p id="d2e1004">Histogram of the P50 bias by COD for all projects and for all projects when removing their first year of operations.</p></caption>
          <graphic xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026-f06.png"/>

        </fig>

      <p id="d2e1014">We similarly found that there was a possibility of a relationship between states, provinces, or small countries with at least four plants and their long-term-corrected P50 bias. In Fig. <xref ref-type="fig" rid="F7"/>, it can be seen that states such as Oklahoma, New Mexico, Nebraska, and New York tend to have above-average P50 biases compared to some states like Texas and California that have many below-average P50 biases. When running a one-way ANOVA <xref ref-type="bibr" rid="bib1.bibx12" id="paren.38"/>, we found a <inline-formula><mml:math id="M31" display="inline"><mml:mi>p</mml:mi></mml:math></inline-formula> value of 0.003, so we ran Tukey's test <xref ref-type="bibr" rid="bib1.bibx10" id="paren.39"/> against the pairs, only finding a significant difference between California and Oklahoma (<inline-formula><mml:math id="M32" display="inline"><mml:mi>p</mml:mi></mml:math></inline-formula> value 0.03) and California and New Mexico (<inline-formula><mml:math id="M33" display="inline"><mml:mi>p</mml:mi></mml:math></inline-formula> value 0.019). However, when we ran a one-way ANOVA <xref ref-type="bibr" rid="bib1.bibx12" id="paren.40"/> to further inspect the relationship between original equipment manufacturers (OEMs) and long-term-corrected P50 bias, we found no relationship, with a <inline-formula><mml:math id="M34" display="inline"><mml:mi>p</mml:mi></mml:math></inline-formula> value of 0.043 for OEMs with at least four data points. For other key distinguishing variables, there were not enough data available relative to the total plants analyzed to warrant further testing.</p>

      <fig id="F7" specific-use="star"><label>Figure 7</label><caption><p id="d2e1059">Scatter plot of each state, province, or country's individual P50 estimate biases and P50 estimate biases for regions with small sample sizes.</p></caption>
          <graphic xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026-f07.png"/>

        </fig>

      <p id="d2e1068">We also demonstrate the P50 bias for each consultant in Fig. <xref ref-type="fig" rid="F8"/> to determine if there were any significant variations across the 19 consultants. Most deviation from the average P50 bias occurred in consultants with a small sample size or where the spread was large. In particular, consultants 1, 3, 6, and 8 had a large spread in their individual estimates; only consultant 5 demonstrated a high level of precision in their estimates. The unspecified grouping, corresponding to no consultant name being provided, yielded the largest sample size with some of the smallest spread in EYA P50 bias. Nevertheless, when performing a one-way ANOVA test for all consultants with at least three data points, we found a <inline-formula><mml:math id="M35" display="inline"><mml:mi>p</mml:mi></mml:math></inline-formula> value of 0.7263, confirming there is no meaningful difference between consultant predictions of P50 energy production.</p>

      <fig id="F8" specific-use="star"><label>Figure 8</label><caption><p id="d2e1082">Scatter plot of each consultant's individual P50 estimate biases and P50 estimate biases without a specified consultant.</p></caption>
          <graphic xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026-f08.png"/>

        </fig>

      <p id="d2e1091">Using the 5000 simulations, we then computed the P50 and P90 as their respective 50th and 10th percentiles and the uncertainty as <inline-formula><mml:math id="M36" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">σ</mml:mi><mml:mi mathvariant="normal">AEP</mml:mi></mml:msub><mml:mo>/</mml:mo><mml:msub><mml:mi mathvariant="italic">μ</mml:mi><mml:mi mathvariant="normal">AEP</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula>. The boxplots in Figs. <xref ref-type="fig" rid="F9"/>, <xref ref-type="fig" rid="F10"/>, and  <xref ref-type="fig" rid="F11"/> demonstrate the distribution of the per-project spread of the P50, long-term P90, and long-term uncertainty bias, respectively, for each of our four down-selection criteria. Compared to the <xref ref-type="bibr" rid="bib1.bibx8" id="text.41"/> results, we can see a slightly worsening P50 bias. We observe that the short-term P90 bias is positive, indicating that the P90 of the estimated operational AEP exceeds the EYA AEP P90 value on average. However, this result does not necessarily suggest that the EYAs use overly conservative means for estimating high-probability energy production levels; to reach this conclusion, we would need to demonstrate that the estimated operational AEP exceeds the EYA P90 predictions 90 % of the time.</p>

      <fig id="F9" specific-use="star"><label>Figure 9</label><caption><p id="d2e1124">Boxplot showing the project-level P50 bias for each of the project down-selection criteria.</p></caption>
          <graphic xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026-f09.png"/>

        </fig>

      <fig id="F10" specific-use="star"><label>Figure 10</label><caption><p id="d2e1135">Boxplot showing the project-level long-term P90 bias for each of the project down-selection criteria.</p></caption>
          <graphic xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026-f10.png"/>

        </fig>

      <fig id="F11" specific-use="star"><label>Figure 11</label><caption><p id="d2e1146">Boxplot showing the project-level long-term-corrected uncertainty biases relative to their 1-year EYA uncertainty estimate for each of the project down-selection criteria.</p></caption>
          <graphic xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026-f11.png"/>

        </fig>

      <p id="d2e1155">As indicated, the uncertainty bias for the long-term-corrected energy estimates relative to the 1-year EYA uncertainties appears to suggest that the EYAs are routinely conservative estimates of uncertainty, with the average bias in the <inline-formula><mml:math id="M37" display="inline"><mml:mo>-</mml:mo></mml:math></inline-formula>52 % to <inline-formula><mml:math id="M38" display="inline"><mml:mo>-</mml:mo></mml:math></inline-formula>48 % range. It should be noted, though, that the uncertainties for the operational AEP estimates and the EYA AEP predictions are not defined equivalently. The operational AEP uncertainties represent the uncertainty in the long-term AEP based on realized energy production during an initial period of operation. EYA estimates, on the other hand, represent the uncertainty in the predicted AEP before the plant is built. The EYA estimates are expected to have a higher degree of uncertainty because there are a significant number of individual model uncertainties and potential sources of loss to consider when predicting energy production prior to building a wind plant.</p>
      <p id="d2e1172">To better understand the uncertainty and prediction errors, we also computed the prediction error of the EYA P50 and P90 estimates for each plant relative to each Monte Carlo iteration's AEP estimation for the corresponding plant (percentage difference between the EYA reading and each iteration's AEP estimate), as shown in Figs. <xref ref-type="fig" rid="F12"/>–<xref ref-type="fig" rid="F14"/>. In Fig. <xref ref-type="fig" rid="F12"/>, which shows the distribution of P50 prediction errors including IAV, we smooth the error histogram to highlight the centering around the mean values. Importantly, we found that the standard deviation of the sample errors ranged from 11.9 % to 13.7 %, which closely aligns with the average 1-year uncertainty in the corresponding EYAs of <inline-formula><mml:math id="M39" display="inline"><mml:mrow><mml:mn mathvariant="normal">10.5</mml:mn><mml:mspace width="0.125em" linebreak="nobreak"/><mml:mi mathvariant="italic">%</mml:mi><mml:mo>±</mml:mo><mml:mn mathvariant="normal">2.6</mml:mn><mml:mspace linebreak="nobreak" width="0.125em"/><mml:mi mathvariant="italic">%</mml:mi></mml:mrow></mml:math></inline-formula>, shown on the right side of Fig. <xref ref-type="fig" rid="F12"/>. However, when comparing the EYA long-term uncertainty to the P50 prediction errors that exclude IAV, as shown in Fig. <xref ref-type="fig" rid="F13"/>, our standard deviation ranges from 11.4 % to 13.1 %, which is much larger than the corresponding EYA long-term uncertainties of <inline-formula><mml:math id="M40" display="inline"><mml:mrow><mml:mn mathvariant="normal">7.5</mml:mn><mml:mspace width="0.125em" linebreak="nobreak"/><mml:mi mathvariant="italic">%</mml:mi><mml:mo>±</mml:mo><mml:mn mathvariant="normal">2.1</mml:mn><mml:mspace width="0.125em" linebreak="nobreak"/><mml:mi mathvariant="italic">%</mml:mi></mml:mrow></mml:math></inline-formula>. As opposed to the direct comparison in Fig. <xref ref-type="fig" rid="F11"/>, these results indicate that the EYA uncertainty, especially the long-term uncertainty, is being underestimated.</p>

      <fig id="F12" specific-use="star"><label>Figure 12</label><caption><p id="d2e1227">(left) Ridge plot showing the percentage error of each sample's AEP value calculated with IAV relative to EYA P50. (right) Histogram of the EYA 1-year uncertainty values.</p></caption>
          <graphic xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026-f12.png"/>

        </fig>

      <fig id="F13" specific-use="star"><label>Figure 13</label><caption><p id="d2e1238">(left) Ridge plot showing the percentage error of each sample's AEP value calculated without IAV relative to EYA P50. (right) Histogram of the EYA long-term uncertainty values.</p></caption>
          <graphic xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026-f13.png"/>

        </fig>

      <fig id="F14" specific-use="star"><label>Figure 14</label><caption><p id="d2e1249">Ridge plots showing the percentage error of each sample's AEP value relative to the EYA P90. (left) The comparison is made between the analyses run with IAV and the EYA short-term P90. (right) The comparison is made between the analyses run without IAV and the EYA long-term P90.</p></caption>
          <graphic xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026-f14.png"/>

        </fig>

      <p id="d2e1258">In line with our project-level P90 estimate comparisons (though more positive), the individual P90 prediction errors for each Monte Carlo iteration center around positive values (Fig. <xref ref-type="fig" rid="F14"/>); however, the 10th percentile of the prediction errors is roughly <inline-formula><mml:math id="M41" display="inline"><mml:mo>-</mml:mo></mml:math></inline-formula>10 % for all cases, demonstrating that EYA P90 estimates are not conservative enough for both the 1-year and the long-term P90 predictions. Ideally, the 10th percentile should be 0 %, meaning the realized operational AEP exceeds the EYA P90 predictions at least 90 % of the time. Note that we can still see the impact of outlier data in the spread of the distribution and the presence of additional peaks.</p>
      <p id="d2e1270">Considering the continued existence of EYA P50 bias, we then calculated the true probability of exceedance value (PXX value) of the EYA-reported P50 based on the long-term-corrected distribution and the PXX value of the long-term-corrected P50 with respect to the EYA distribution statistics. We derived the respective PXX values by first finding the number of standard deviations by which the estimated P50 value was from the comparative mean, <inline-formula><mml:math id="M42" display="inline"><mml:mrow><mml:mi>x</mml:mi><mml:mo>=</mml:mo><mml:mo>(</mml:mo><mml:mi mathvariant="normal">P</mml:mi><mml:msub><mml:mn mathvariant="normal">50</mml:mn><mml:mi mathvariant="normal">ltc</mml:mi></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mi mathvariant="italic">μ</mml:mi><mml:mi mathvariant="normal">eya</mml:mi></mml:msub><mml:mo>)</mml:mo><mml:mo>/</mml:mo><mml:msub><mml:mi mathvariant="italic">σ</mml:mi><mml:mi mathvariant="normal">eya</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula>. Then, we calculated the probability based on the cumulative distribution function of the number of standard deviations and subtracted it from 1 to ensure the overestimates yield a lower PXX value than the P50. We followed the same process for the long-term-corrected P50 with respect to the EYA AEP distribution by computing the number of standard deviations (<inline-formula><mml:math id="M43" display="inline"><mml:mrow><mml:mi>x</mml:mi><mml:mo>=</mml:mo><mml:mo>(</mml:mo><mml:mi mathvariant="normal">P</mml:mi><mml:msub><mml:mn mathvariant="normal">50</mml:mn><mml:mi mathvariant="normal">eya</mml:mi></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mi mathvariant="italic">μ</mml:mi><mml:mi mathvariant="normal">ltc</mml:mi></mml:msub><mml:mo>)</mml:mo><mml:mo>/</mml:mo><mml:msub><mml:mi mathvariant="italic">σ</mml:mi><mml:mi mathvariant="normal">ltc</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula>), calculating the probability from the cumulative distribution function, and subtracting it from 1 to account for the overprediction bias. The results of these calculations for all plants are shown in Fig. <xref ref-type="fig" rid="F15"/>. We found that the EYA P50 represented a P22 on average, with a standard deviation of 30 percentage points, and that the long-term-corrected P50 represented a P72 with respect to the EYA AEP distribution, with a standard deviation of 26 percentage points. The two distributions roughly mirror each other and form clusters, where the EYA P50 is significantly overestimated. For the long-term-corrected distribution, the EYA P50 is <inline-formula><mml:math id="M44" display="inline"><mml:mo>-</mml:mo></mml:math></inline-formula>0.95 <inline-formula><mml:math id="M45" display="inline"><mml:mo>±</mml:mo></mml:math></inline-formula> 1.61 standard deviations from the long-term-corrected mean, with the bottom 25 % of the distribution being at most <inline-formula><mml:math id="M46" display="inline"><mml:mo>-</mml:mo></mml:math></inline-formula>1.64 standard deviations. Similarly, the long-term-corrected P50 is on average 2.16 <inline-formula><mml:math id="M47" display="inline"><mml:mo>±</mml:mo></mml:math></inline-formula> 2.84 standard deviations from the mean EYA-estimated AEP, with the top 75 % of the data being at least 3.99 standard deviations from the mean. Table <xref ref-type="table" rid="T2"/> shows that when comparing the subset of plants with CODs after 2010 and when removing the first year of operational data, the results improve slightly. For example, the EYA P50 PXX value relative to the long-term-corrected distribution corresponds to a P24 compared to a P22 when viewing all plants with all available operational years. Similarly, the long-term-corrected P50 based on the EYA distribution shifts from a P72 to a P70 when filtering out plants with CODs prior to 2011 and removing the first year of data.</p>

      <fig id="F15" specific-use="star"><label>Figure 15</label><caption><p id="d2e1379">Box and swarm plots for the estimation of the true PXX value of the EYA P50 based on the long-term-corrected distributions of energy generation (top) and the derived PXX value of the long-term-corrected P50 from the EYA-estimated energy production distribution (bottom).</p></caption>
          <graphic xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026-f15.png"/>

        </fig>

<table-wrap id="T2" specific-use="star"><label>Table 2</label><caption><p id="d2e1391">Table of the EYA P50 PXX value based on the long-term-corrected distribution and the long-term-corrected P50 PXX value for each of the data subsets.</p></caption><oasis:table frame="topbot"><oasis:tgroup cols="3">
     <oasis:colspec colnum="1" colname="col1" align="left"/>
     <oasis:colspec colnum="2" colname="col2" align="center"/>
     <oasis:colspec colnum="3" colname="col3" align="center"/>
     <oasis:thead>
       <oasis:row>
         <oasis:entry colname="col1">Data subset</oasis:entry>
         <oasis:entry colname="col2">EYA P50 w.r.t. long-term</oasis:entry>
         <oasis:entry colname="col3">Long-term-corrected P50</oasis:entry>
       </oasis:row>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2">correction</oasis:entry>
         <oasis:entry colname="col3">w.r.t. EYA</oasis:entry>
       </oasis:row>
     </oasis:thead>
     <oasis:tbody>
       <oasis:row>
         <oasis:entry colname="col1">All plants</oasis:entry>
         <oasis:entry colname="col2">22</oasis:entry>
         <oasis:entry colname="col3">72</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Drop first year</oasis:entry>
         <oasis:entry colname="col2">22</oasis:entry>
         <oasis:entry colname="col3">72</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">COD <inline-formula><mml:math id="M48" display="inline"><mml:mo>≥</mml:mo></mml:math></inline-formula> 2011</oasis:entry>
         <oasis:entry colname="col2">23</oasis:entry>
         <oasis:entry colname="col3">70</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Drop first year &amp; COD <inline-formula><mml:math id="M49" display="inline"><mml:mo>≥</mml:mo></mml:math></inline-formula> 2011</oasis:entry>
         <oasis:entry colname="col2">24</oasis:entry>
         <oasis:entry colname="col3">70</oasis:entry>
       </oasis:row>
     </oasis:tbody>
   </oasis:tgroup></oasis:table></table-wrap>

<table-wrap id="T3" specific-use="star"><label>Table 3</label><caption><p id="d2e1499">Table of the EYA P50 and EYA P90 PXX values based on the long-term-corrected prediction error for both the short-term (with IAV) and the long-term (without IAV) distributions.</p></caption><oasis:table frame="topbot"><oasis:tgroup cols="5">
     <oasis:colspec colnum="1" colname="col1" align="left"/>
     <oasis:colspec colnum="2" colname="col2" align="center"/>
     <oasis:colspec colnum="3" colname="col3" align="center" colsep="1"/>
     <oasis:colspec colnum="4" colname="col4" align="center"/>
     <oasis:colspec colnum="5" colname="col5" align="center"/>
     <oasis:thead>
       <oasis:row>
         <oasis:entry colname="col1">Data subset</oasis:entry>
         <oasis:entry rowsep="1" namest="col2" nameend="col3" colsep="1">EYA P50 w.r.t. long-term correction </oasis:entry>
         <oasis:entry rowsep="1" namest="col4" nameend="col5">EYA P90 w.r.t. long-term correction </oasis:entry>
       </oasis:row>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2">Short-term</oasis:entry>
         <oasis:entry colname="col3">Long-term</oasis:entry>
         <oasis:entry colname="col4">Short-term</oasis:entry>
         <oasis:entry colname="col5">Long-term</oasis:entry>
       </oasis:row>
     </oasis:thead>
     <oasis:tbody>
       <oasis:row>
         <oasis:entry colname="col1">All plants</oasis:entry>
         <oasis:entry colname="col2">22</oasis:entry>
         <oasis:entry colname="col3">19</oasis:entry>
         <oasis:entry colname="col4">76</oasis:entry>
         <oasis:entry colname="col5">66</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Drop first year</oasis:entry>
         <oasis:entry colname="col2">22</oasis:entry>
         <oasis:entry colname="col3">19</oasis:entry>
         <oasis:entry colname="col4">75</oasis:entry>
         <oasis:entry colname="col5">67</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">COD <inline-formula><mml:math id="M50" display="inline"><mml:mo>≥</mml:mo></mml:math></inline-formula> 2011</oasis:entry>
         <oasis:entry colname="col2">23</oasis:entry>
         <oasis:entry colname="col3">20</oasis:entry>
         <oasis:entry colname="col4">77</oasis:entry>
         <oasis:entry colname="col5">65</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Drop first year &amp; COD <inline-formula><mml:math id="M51" display="inline"><mml:mo>≥</mml:mo></mml:math></inline-formula> 2011</oasis:entry>
         <oasis:entry colname="col2">24</oasis:entry>
         <oasis:entry colname="col3">20</oasis:entry>
         <oasis:entry colname="col4">77</oasis:entry>
         <oasis:entry colname="col5">68</oasis:entry>
       </oasis:row>
     </oasis:tbody>
   </oasis:tgroup></oasis:table></table-wrap>

      <fig id="F16" specific-use="star"><label>Figure 16</label><caption><p id="d2e1641">Summary of the EYA-estimated availability losses (mixture of time and energy based), long-term-corrected P50 of operational availability and curtailment losses, and the percentage point difference between the operational availability P50 and EYA availability losses. From top to bottom: the boxplots show the distribution of EYA availability loss, the long-term-corrected operational availability loss P50,  the difference between the long-term-corrected operational P50 and EYA-estimated availability losses, the long-term-corrected operational availability loss P50 for all projects, and the long-term-corrected operational curtailment loss P50 for all projects.</p></caption>
          <graphic xlink:href="https://wes.copernicus.org/articles/11/1251/2026/wes-11-1251-2026-f16.png"/>

        </fig>

      <p id="d2e1651">Another way to understand what the P50 and P90 values represent based on the long-term-corrected distribution is to compute what percentage of the P50 and P90 prediction errors (as shown in Figs. <xref ref-type="fig" rid="F12"/>–<xref ref-type="fig" rid="F14"/>) is greater than 0 %. In Table <xref ref-type="table" rid="T3"/> we demonstrate the close alignment these prediction error results have with those presented in Fig. <xref ref-type="fig" rid="F15"/> and Table <xref ref-type="table" rid="T2"/>. The EYA P50 corresponds to a P19 to P20 based on the long-term-corrected AEP distributions without IAV, depending on the scenario, and the long-term EYA P90 corresponds to a P65 to P69, depending on the scenario. For both the short-term P50 and the P90, the PXX values improved with the long-term-corrected AEP distributions, with IAV yielding a realized P22 to P24 for the EYA P50 value and a P75 to P77 for the EYA 1-year P90 value, depending on the scenario.</p>
      <p id="d2e1664">Finally, because EYA availability and curtailment loss estimates were provided for 36 projects, we also computed the P50 value for each of the long-term-estimated operational availability and curtailment losses in the Monte Carlo simulations by project, as shown in Fig. <xref ref-type="fig" rid="F16"/>. We have excluded a detailed comparison between the EYA-estimated curtailment losses and operational curtailment losses, as 31 of the EYAs provided an EYA-estimated 0 % curtailment loss, and only two EYAs estimated 0 % availability losses, which were removed from the comparison. As such, the results of the availability comparison are shown in Fig. <xref ref-type="fig" rid="F16"/>. We found that the average difference between the operational P50 values and EYA estimates for availability losses was 1 percentage point, indicating a slight underestimation of availability loss; however, the median value is 0.22 percentage points, and the standard deviation is 3.63 percentage points. Taken together, this suggests an inconsistent understanding of availability across the board, even if it contributes a minimal amount to the AEP P50 biases. Further, it should be noted that availability data were provided in mixed formats – energy lost and time-based availability – so the comparisons are not wholly on the same basis.</p>
      <p id="d2e1671">The curtailment P50 bias only consists of seven comparisons, so the underestimation of losses by 13 <inline-formula><mml:math id="M52" display="inline"><mml:mo>±</mml:mo></mml:math></inline-formula> 23 percentage points may not be representative of reality. In fact, many of the EYAs with curtailment loss information state outright that situations where the energy grid may force the power plant to stop sending power are not considered, so the comparison of realized curtailment and EYA-estimated curtailment likely provides little insight without further details or delineations.</p>
</sec>
</sec>
<sec id="Ch1.S4" sec-type="conclusions">
  <label>4</label><title>Conclusions</title>
      <p id="d2e1691">In this paper, we built on the methodology of the work performed in <xref ref-type="bibr" rid="bib1.bibx8" id="text.42"/> by gathering operational data from wind plant owners and operators that include the curtailment and availability losses that could not be accurately accounted for in the original work. We observed many of the same relationships as the original study, such as an overprediction of energy yield from consultants and a shrinking P50 bias following the change in methodology in 2011. However, we found that P50 bias has remained in line with but is slightly worse than (0.8 % to 1.1 % higher) <xref ref-type="bibr" rid="bib1.bibx8" id="text.43"/>, now that we include more recent projects and are able to account for curtailment and availability losses. Additionally, we observed that the EYA uncertainty and P90 estimates are not conservative enough when compared to the observed distributions of AEP prediction errors relative to the operational AEP estimates, although the EYA 1-year AEP uncertainties are a reasonable approximation of the observed uncertainty.</p>
      <p id="d2e1700">This work motivates additional research to understand how the EYA process can be improved to reduce the AEP P50 overprediction bias and better characterize the uncertainty in the predictions, especially because P50 bias appears to be increasing compared to the findings in <xref ref-type="bibr" rid="bib1.bibx8" id="text.44"/>. Our results ultimately raise further questions that were out of the scope of this study. One example is to measure the impacts of external wakes from neighboring farms that came online after a plant's EYA was completed, and the extent to which they contribute to P50 overprediction bias. Many projects represented in this dataset are older, and it is likely that neighboring projects have been built since their inception, creating unaccounted-for losses in the EYAs that would be present in the MORs.</p>
      <p id="d2e1706">A more fundamental question would be whether working with higher-resolution operational data, turbine operating status, and consistently delineated loss data combined with more advanced long-term operational AEP modeling could provide a more accurate or less uncertain long-term correction. As in <xref ref-type="bibr" rid="bib1.bibx8" id="text.45"/>, it remains an open question of how to curate a broader selection of wind plant MORs, EYAs, and plant characteristics to understand the effects that plant characteristics such as geography or turbine technology can have on energy production.</p>
</sec>

      
      </body>
    <back><notes notes-type="codeavailability"><title>Code availability</title>

      <p id="d2e1716">The OpenOA software is publicly available on GitHub at <uri>https://github.com/NatLabRockies/OpenOA</uri> (last access: 14 January 2026; <ext-link xlink:href="https://doi.org/10.21105/joss.02171" ext-link-type="DOI">10.21105/joss.02171</ext-link>, <xref ref-type="bibr" rid="bib1.bibx9" id="altparen.46"/>). However, the analysis code is tied to confidential data references and is not made public.</p>
  </notes><notes notes-type="dataavailability"><title>Data availability</title>

      <p id="d2e1731">The monthly operating reports and energy yield assessment data are not publicly available and are under a strict confidentiality agreement with NLR to maintain anonymity of the wind plants and their owners and operators.</p>
  </notes><notes notes-type="authorcontribution"><title>Author contributions</title>

      <p id="d2e1737">RH and ES met with wind plant owners to obtain EYA and operational data. RH performed the analyses and drafted the paper. ES advised on the work, and reviewed and edited the paper.</p>
  </notes><notes notes-type="competinginterests"><title>Competing interests</title>

      <p id="d2e1743">The contact author has declared that neither of the authors has any competing interests.</p>
  </notes><notes notes-type="disclaimer"><title>Disclaimer</title>

      <p id="d2e1749">The views expressed in the article do not necessarily represent the views of the DOE or the U.S. Government.  Publisher's note: Copernicus Publications remains neutral with regard to jurisdictional claims made in the text, published maps, institutional affiliations, or any other geographical representation in this paper. The authors bear the ultimate responsibility for providing appropriate place names. Views expressed in the text are those of the authors and do not necessarily reflect the views of the publisher.</p>
  </notes><ack><title>Acknowledgements</title><p id="d2e1758">The authors would like to thank the anonymous wind farm operators, who provided the wind farm data that were used in this analysis.  <xref ref-type="bibr" rid="bib1.bibx5" id="text.47"/> was downloaded from the Copernicus Climate Change Service (2023).</p></ack><notes notes-type="financialsupport"><title>Financial support</title>

      <p id="d2e1766">This work was authored by the National Laboratory of the Rockies for the US Department of Energy (DOE) under contract no. DE-AC36-08GO28308. Funding was provided by the US Department of Energy Office of Critical Minerals and Energy Innovation Integrated Energy Systems Office.</p>
  </notes><notes notes-type="reviewstatement"><title>Review statement</title>

      <p id="d2e1772">This paper was edited by Andrea Hahmann and reviewed by two anonymous referees.</p>
  </notes><ref-list>
    <title>References</title>

      <ref id="bib1.bibx1"><label>Bodini and Optis(2020)</label><mixed-citation>Bodini, N. and Optis, M.: Operational-based annual energy production uncertainty: are its components actually uncorrelated?, Wind Energ. Sci., 5, 1435–1448, <ext-link xlink:href="https://doi.org/10.5194/wes-5-1435-2020" ext-link-type="DOI">10.5194/wes-5-1435-2020</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bibx2"><label>Clifton et al.(2016)Clifton, Smith, and Fields</label><mixed-citation>Clifton, A., Smith, A., and Fields, M.: Wind Plant Preconstruction Energy Estimates, Current Practice and Opportunities, <ext-link xlink:href="https://doi.org/10.2172/1248798" ext-link-type="DOI">10.2172/1248798</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bibx3"><label>Dickinson et al.(2014)Dickinson, Monache, McCormack, and Magontier</label><mixed-citation>Dickinson, K., Monache, L. D., McCormack, K., and Magontier, P.: Wind Energy Resource Assessment: Information Production, Uses, and Value – Survey Report, <uri>https://sciencepolicy.colorado.edu/admin/publication_files/2014.60.pdf</uri> (last access: 8 January 2026), 2014.</mixed-citation></ref>
      <ref id="bib1.bibx4"><label>Global Modeling and Assimilation Office (GMAO)(2015)</label><mixed-citation>Global Modeling and Assimilation Office (GMAO): MERRA-2 tavg1_2d_slv_Nx: 2d,1-Hourly,Time-Averaged,Single-Level,Assimilation,Single-Level Diagnostics V5.12.4, Goddard Earth Sciences Data and Information Services Center (GES DISC) [data set]<ext-link xlink:href="https://doi.org/10.5067/VJAFPLI1CSIV" ext-link-type="DOI">10.5067/VJAFPLI1CSIV</ext-link>, 2015.</mixed-citation></ref>
      <ref id="bib1.bibx5"><label>Hersbach et al.(2018)Hersbach, Bell, Berrisford, Biavati, Horányi, Muñoz Sabater, Nicolas, Peubey, Radu, Rozum, Schepers, Simmons, Soci, Dee, and Thépaut</label><mixed-citation>Hersbach, H., Bell, B., Berrisford, P., Biavati, G., Horányi, A., Muñoz Sabater, J., Nicolas, J., Peubey, C., Radu, R., Rozum, I., Schepers, D., Simmons, A., Soci, C., Dee, D., and Thépaut, J.-N.: ERA5 hourly data on single levels from 1940 to present, Copernicus Climate Change Service (C3S) Climate Data Store (CDS) [data set], <ext-link xlink:href="https://doi.org/10.24381/cds.adbb2d47" ext-link-type="DOI">10.24381/cds.adbb2d47</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bibx6"><label>Huber(1964)</label><mixed-citation>Huber, P. J.: Robust Estimation of a Location Parameter,  Ann. Math. Stat., 35, 73–101, <ext-link xlink:href="https://doi.org/10.1214/aoms/1177703732" ext-link-type="DOI">10.1214/aoms/1177703732</ext-link>, 1964.</mixed-citation></ref>
      <ref id="bib1.bibx7"><label>Lee and Fields(2021)</label><mixed-citation>Lee, J. C. Y. and Fields, M. J.: An overview of wind-energy-production prediction bias, losses, and uncertainties, Wind Energ. Sci., 6, 311–365, <ext-link xlink:href="https://doi.org/10.5194/wes-6-311-2021" ext-link-type="DOI">10.5194/wes-6-311-2021</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx8"><label>Lunacek et al.(2018)Lunacek, Jason Fields, Craig, Lee, Meissner, Philips, Sheng, and King</label><mixed-citation>Lunacek, M., Jason Fields, M., Craig, A., Lee, J., Meissner, J., Philips, C., Sheng, S., and King, R.: Understanding Biases in Pre-Construction Estimates, J. Phys. Conf. Ser., 1037, <ext-link xlink:href="https://doi.org/10.1088/1742-6596/1037/6/062009" ext-link-type="DOI">10.1088/1742-6596/1037/6/062009</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bibx9"><label>Perr-Sauer et al.(2021)Perr-Sauer, Optis, Fields, Bodini, Lee, Todd, Simley, Hammond, Phillips, Lunacek, Kemper, Williams, Craig, Agarwal, Sheng, and Meissner</label><mixed-citation>Perr-Sauer, J., Optis, M., Fields, J. M., Bodini, N., Lee, J. C., Todd, A., Simley, E., Hammond, R., Phillips, C., Lunacek, M., Kemper, T., Williams, L., Craig, A., Agarwal, N., Sheng, S., and Meissner, J.: OpenOA: An Open-Source Codebase For Operational Analysis of Wind Farms, Journal of Open Source Software, 6, 2171, <ext-link xlink:href="https://doi.org/10.21105/joss.02171" ext-link-type="DOI">10.21105/joss.02171</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx10"><label>Tukey(1949)</label><mixed-citation>Tukey, J. W.: Comparing Individual Means in the Analysis of Variance, Biometrics, 5, 99–114, <ext-link xlink:href="https://doi.org/10.2307/3001913" ext-link-type="DOI">10.2307/3001913</ext-link>,   1949. </mixed-citation></ref>
      <ref id="bib1.bibx11"><label>Welch(1947)</label><mixed-citation>Welch, B. L.: THE GENERALIZATION OF `STUDENT'S' PROBLEM WHEN SEVERAL DIFFERENT POPULATION VARLANCES ARE INVOLVED, Biometrika, 34, 28–35, <ext-link xlink:href="https://doi.org/10.1093/biomet/34.1-2.28" ext-link-type="DOI">10.1093/biomet/34.1-2.28</ext-link>, 1947.</mixed-citation></ref>
      <ref id="bib1.bibx12"><label>Welch(1951)</label><mixed-citation>Welch, B. L.: On the Comparison of Several Mean Values: An Alternative Approach, Biometrika, 38, 330–336, <uri>http://www.jstor.org/stable/2332579</uri> (last access: 8 January 2026), 1951.</mixed-citation></ref>

  </ref-list></back>
    <!--<article-title-html>Biases in preconstruction estimates of wind plant annual energy production</article-title-html>
<abstract-html/>
<ref-html id="bib1.bib1"><label>Bodini and Optis(2020)</label><mixed-citation>
      
Bodini, N. and Optis, M.: Operational-based annual energy production uncertainty: are its components actually uncorrelated?, Wind Energ. Sci., 5, 1435–1448, <a href="https://doi.org/10.5194/wes-5-1435-2020" target="_blank">https://doi.org/10.5194/wes-5-1435-2020</a>, 2020.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib2"><label>Clifton et al.(2016)Clifton, Smith, and Fields</label><mixed-citation>
      
Clifton, A., Smith, A., and Fields, M.: Wind Plant Preconstruction Energy
Estimates, Current Practice and Opportunities, <a href="https://doi.org/10.2172/1248798" target="_blank">https://doi.org/10.2172/1248798</a>, 2016.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib3"><label>Dickinson et al.(2014)Dickinson, Monache, McCormack, and
Magontier</label><mixed-citation>
      
Dickinson, K., Monache, L. D., McCormack, K., and Magontier, P.: Wind Energy
Resource Assessment: Information Production, Uses, and Value – Survey
Report,
<a href="https://sciencepolicy.colorado.edu/admin/publication_files/2014.60.pdf" target="_blank"/> (last access: 8 January 2026),
2014.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib4"><label>Global Modeling and Assimilation Office (GMAO)(2015)</label><mixed-citation>
      
Global Modeling and Assimilation Office (GMAO): MERRA-2 tavg1_2d_slv_Nx:
2d,1-Hourly,Time-Averaged,Single-Level,Assimilation,Single-Level Diagnostics
V5.12.4, Goddard Earth Sciences Data and Information Services Center (GES DISC) [data set]<a href="https://doi.org/10.5067/VJAFPLI1CSIV" target="_blank">https://doi.org/10.5067/VJAFPLI1CSIV</a>, 2015.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib5"><label>Hersbach et al.(2018)Hersbach, Bell, Berrisford, Biavati, Horányi,
Muñoz Sabater, Nicolas, Peubey, Radu, Rozum, Schepers, Simmons, Soci, Dee,
and Thépaut</label><mixed-citation>
      
Hersbach, H., Bell, B., Berrisford, P., Biavati, G., Horányi, A.,
Muñoz Sabater, J., Nicolas, J., Peubey, C., Radu, R., Rozum, I., Schepers,
D., Simmons, A., Soci, C., Dee, D., and Thépaut, J.-N.: ERA5 hourly data on
single levels from 1940 to present, Copernicus Climate Change Service (C3S) Climate Data Store (CDS) [data set], <a href="https://doi.org/10.24381/cds.adbb2d47" target="_blank">https://doi.org/10.24381/cds.adbb2d47</a>, 2018.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib6"><label>Huber(1964)</label><mixed-citation>
      
Huber, P. J.: Robust Estimation of a Location Parameter,  Ann.
Math. Stat., 35, 73–101, <a href="https://doi.org/10.1214/aoms/1177703732" target="_blank">https://doi.org/10.1214/aoms/1177703732</a>, 1964.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib7"><label>Lee and Fields(2021)</label><mixed-citation>
      
Lee, J. C. Y. and Fields, M. J.: An overview of wind-energy-production prediction bias, losses, and uncertainties, Wind Energ. Sci., 6, 311–365, <a href="https://doi.org/10.5194/wes-6-311-2021" target="_blank">https://doi.org/10.5194/wes-6-311-2021</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib8"><label>Lunacek et al.(2018)Lunacek, Jason Fields, Craig, Lee, Meissner,
Philips, Sheng, and King</label><mixed-citation>
      
Lunacek, M., Jason Fields, M., Craig, A., Lee, J., Meissner, J., Philips, C.,
Sheng, S., and King, R.: Understanding Biases in Pre-Construction Estimates,
J. Phys. Conf. Ser., 1037,
<a href="https://doi.org/10.1088/1742-6596/1037/6/062009" target="_blank">https://doi.org/10.1088/1742-6596/1037/6/062009</a>,
2018.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib9"><label>Perr-Sauer et al.(2021)Perr-Sauer, Optis, Fields, Bodini, Lee, Todd,
Simley, Hammond, Phillips, Lunacek, Kemper, Williams, Craig, Agarwal, Sheng,
and Meissner</label><mixed-citation>
      
Perr-Sauer, J., Optis, M., Fields, J. M., Bodini, N., Lee, J. C., Todd, A.,
Simley, E., Hammond, R., Phillips, C., Lunacek, M., Kemper, T., Williams, L.,
Craig, A., Agarwal, N., Sheng, S., and Meissner, J.: OpenOA: An Open-Source
Codebase For Operational Analysis of Wind Farms, Journal of Open Source
Software, 6, 2171, <a href="https://doi.org/10.21105/joss.02171" target="_blank">https://doi.org/10.21105/joss.02171</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib10"><label>Tukey(1949)</label><mixed-citation>
      
Tukey, J. W.: Comparing Individual Means in the Analysis of Variance,
Biometrics, 5, 99–114, <a href="https://doi.org/10.2307/3001913" target="_blank">https://doi.org/10.2307/3001913</a>,   1949.


    </mixed-citation></ref-html>
<ref-html id="bib1.bib11"><label>Welch(1947)</label><mixed-citation>
      
Welch, B. L.: THE GENERALIZATION OF `STUDENT'S' PROBLEM WHEN SEVERAL
DIFFERENT POPULATION VARLANCES ARE INVOLVED, Biometrika, 34, 28–35,
<a href="https://doi.org/10.1093/biomet/34.1-2.28" target="_blank">https://doi.org/10.1093/biomet/34.1-2.28</a>, 1947.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib12"><label>Welch(1951)</label><mixed-citation>
      
Welch, B. L.: On the Comparison of Several Mean Values: An Alternative
Approach, Biometrika, 38, 330–336,
<a href="http://www.jstor.org/stable/2332579" target="_blank"/> (last access: 8 January 2026), 1951.

    </mixed-citation></ref-html>--></article>
