the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
Accelerating compound flood risk assessments through active learning: A case study of Charleston County (USA)
Abstract. Flooding is the most likely natural hazard that affects individuals and can be driven by rainfall, river discharge, storm surge, tides, and waves. Compound floods result from their co-occurrence and can generate a larger flood hazard when compared to the sum of the individual drivers. Current state-of-the-art stochastic compound flood risk assessments are based on statistical, hydrodynamic, and impact simulations. However, the stochastic nature of some key variables in the flooding process is often not accounted for as adding stochastic variables exponentially increases the computational costs (i.e., the curse of dimensionality). These simplifications (e.g., a constant flood driver duration or a constant time lag between flood drivers) may lead to a mis-quantification of the flood risk. This study develops a conceptual framework that allows for a better representation of compound flood risk while limiting the increase in the overall computational time. After generating synthetic events from a statistical model fitted to the selected flood drivers, the proposed framework applies a Treed Gaussian Process (TGP). A TGP uses active learning to explore the uncertainty associated with the response of damages to synthetic events. Thereby, it informs on the best choice of hydrodynamic and impact simulations to run to reduce uncertainty in the damages. Once the TGP predicts the damage of all synthetic events within a tolerated uncertainty range, the flood risk is calculated. As a proof of concept, the proposed framework was applied to the case study of Charleston County (South Carolina, USA) and compared with a state-of-the-art stochastic compound flood risk model, which used equidistant sampling with linear scatter interpolation. The proposed framework decreased the overall computational time by a factor of four, and decreased the root mean square error in damages by a factor of eight. With a reduction in computational time and errors, additional stochastic variables such as the drivers' duration and time lag were included in the compound flood risk assessment. Not accounting for these resulted in an underestimation of 11.6 % (25.47 million USD) in the expected annual damage. Thus, by accelerating compound flood risk assessments with active learning, the framework presented here allows for more comprehensive assessments as it loosens constraints imposed by the curse of dimensionality.
- Preprint
(2801 KB) - Metadata XML
- BibTeX
- EndNote
Status: open (extended)
-
RC1: 'Comment on nhess-2024-196', Anonymous Referee #1, 05 Dec 2024
reply
Review of the paper “Accelerating compound flood risk assessments through active learning: A case study of Charleston County (USA)” submitted to NHESS by Terlinden-Ruhl
The study explores the application of active learning to improve the accuracy of compound flood risk assessments while minimizing computational demands. The framework leverages the input-output uncertainty related to economic damages to minimize the number of required hydrodynamic and impact simulations. Compared to traditional equidistant sampling, the proposed active learning modeling framework shows lower RMSE with lower computational time.
I would like to compliment the authors for a well-written manuscript. To the best of my knowledge, this is the first study to apply active learning techniques in the context of compound flooding, demonstrating an approach to advancing this area of research. I recommend this manuscript for publication after addressing the minor revisions outlined below. However, I am not qualified to evaluate the detailed technical aspects of active learning.
- The authors used the “skew surge” parameter to represent the storm surge component of water levels. However, one of the key parameters for simulating flooding mentioned in the manuscript is the “duration” of the event. The skew surge is a measure of the storm surge integrated over a tidal cycle, and thus it has no duration associated with it. How are the authors defining the duration (or time series) of the skew surge? In addition, SFINCS needs a water level time series as boundary conditions to simulate coastal flooding, how do the authors generate a storm-tide hydrograph of these events using the skew surge (i.e. a single value over a tidal cycle)? Consider explaining more about these steps.
- The methodology for obtaining the skew surge time series is somewhat unclear (see L 112). Specifically, the paper does not specify how tidal levels were determined. Were NOAA-predicted tides used, or was a harmonic analysis performed on water level data? Additionally, when removing sea-level rise, the phrase “subtracting 1-year moving average from the skew surge time series” is unclear.
- The uncertainty bands presented in Figure 6 indicate that, in some cases, the return value of a 1000-year event has narrower uncertainty compared to a 10-year event. Does this result only reflect the uncertainty associated with the surrogate damage model for each return period? It is given that a direct comparison of confidence intervals may not be valid among different numbers of stochastic variables, as the stopping criteria are based on uncertainty. Could the authors provide an additional explanation regarding the interpretation of these uncertainty bands? What information do they provide?
- L 207: Authors explain that they used a Gaussian distribution to reconstruct the water levels time series. It would be better to cite a previous study or provide a proper justification for this assumption. Same for the rainfall, how well does their simplification of the hydrograph and hyetographs represent real events? Additionally, how well we can represent real events using a constant rainfall (point estimate) over the entire domain or constant hydrographs along the coast? I think this is an important point that needs to be addressed at least as a limitation of the analysis presented.
- The authors use the SFINCS model, which was set up and validated in a previous study. While this provides a solid foundation, I believe it would be beneficial to include some additional key details about the model setup in this paper. For example, providing information on the model resolution, whether constant or spatially varying roughness was used, and how infiltration was handled (if included).
- L 134: The description of applying a threshold to skew surge magnitude when it “co-occurred with a higher high tide” is confusing. Since the definition of skew surge inherently depends on the high tide (or higher high tide) in a tidal cycle, further clarification would benefit the reader.
- L 3-5: The phrase “large flood hazard when compared to the sum of the individual drivers” could be explained better.
- L 138: Is there a specific rationale for selecting a wider (14-day) de-clustering window? Justifying this choice would strengthen the methodology section.
- L 144: The manuscript suggests a minimum duration of 6 days for rainfall and skew surge events. Does it mean that there were no hours with rainfall less than 0.3 mm within 6 days during an event? So all the extreme events were simulated for at least 6 days? Could the authors confirm and clarify?
- L 352: Please specify whether the given time durations refer to the computational time required “per simulation” or the total time for simulating all events.
- In Figure 2, ensure that the colors representing coastal and inland counties are correct.
Citation: https://doi.org/10.5194/nhess-2024-196-RC1
Data sets
Compound_TGP Lucas Terlinden-Ruhl https://doi.org/10.5281/zenodo.13910108
Viewed
HTML | XML | Total | BibTeX | EndNote | |
---|---|---|---|---|---|
153 | 49 | 15 | 217 | 8 | 7 |
- HTML: 153
- PDF: 49
- XML: 15
- Total: 217
- BibTeX: 8
- EndNote: 7
Viewed (geographical distribution)
Country | # | Views | % |
---|
Total: | 0 |
HTML: | 0 |
PDF: | 0 |
XML: | 0 |
- 1