the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
Downscaling the probability of heavy rainfall over the Nordic countries
Abstract. We used empirical-statistical downscaling to derive local statistics for 24-hr and sub-daily precipitation over the Nordic countries, based on large-scale information provided by global climate models. The local statistics included probabilities for heavy precipitation and intensity-duration-frequency curves for sub-daily rainfall. The downscaling was based on estimating key parameters defining the shape of mathematical curves describing probabilities and return-values, namely the annual wet-day frequency fw and the wet-day mean precipitation μ. Both parameters were used as predictands representing local precipitation statistics as well as predictors representing large-scale conditions. We used multi-model ensembles of global climate model (CMIP6) simulations, calibrated on the ERA5 reanalysis, to derive local projections for future outlooks. Our analysis included an evaluation of how well the global climate models reproduced the predictors, in addition to assessing the quality of downscaled precipitation statistics. The evaluation suggested that present global climate models capture essential covariance, and there was a good match between annual wet-day frequency and wet-day mean precipitation derived from ERA5 and local rain gauges in the Nordic region. Furthermore, the ensemble downscaled results for fw and μ were approximately normally distributed which may justify using the ensemble mean and standard deviation to describe the ensemble spread. Hence, our efforts provide a demonstration for how empirical-statistical downscaling can be used to provide practical information on heavy rainfall which subsequently may be used for impact studies. Future projections for the Nordic region indicated little increase in precipitation due to more wet days, but most of the contribution comes from increased mean intensity. The west coast of Norway had the highest probabilities of receiving more than 30 mm/day precipitation, but the strongest relative trend in this probability was projected over northern Finland. Furthermore, the highest estimates for trends in 10-year and 25-year return-values were projected over western Norway where they were high from the outset. Our results also suggested that future precipitation intensity is sensitive to future emissions whereas the wet-day frequency is less sensitive.
- Preprint
(13236 KB) - Metadata XML
-
Supplement
(49524 KB) - BibTeX
- EndNote
Status: closed
-
RC1: 'Comment on egusphere-2024-1463', Anonymous Referee #1, 20 Jun 2024
Review of “Downscaling the probability of heavy rainfall over the Nordic Countries”, by Benestad et al.
This manuscript is addressing an important topic and is well-written. However, I would recommend major revisions to give more details on the main body of the manuscript rather than in the SI. I think that the manuscript should be rewritten for the broad readership of HESS, or alternatively be submitted to an more specialized journal.
The essence of the approach presented is to build a statistical model that relates large-scale information (EA5 or CMIP6) to station-level data. Then, this model is used to estimate the future climate at stations locations, based on CMIP6. I am not sure this approach can be termed downscaling, as in my mind downscaling is the passage of data from a coarse grid to a finer grid. Here the correct term, in my understanding, would be something like change of support. In particular, the data are only produced at the stations locations, meaning that it is less effective in areas with a lower stations density (such as northern Norway/Finland).
Moreover, the details of the methodology are very condensed and not comprehensible for a reader that is not already familiar with it. The method is summarized very shortly in section 2.2, and mostly referred to through citations to other papers by the authors, with the reader is not necessarily familiar. Therefore, it is difficult to grasp it by only reading the manuscript. A schematic of the methodology would be helpful.
Very few details are given on the kriging procedure used. Which variogram parameters were used, and what does this tell us about the spatial dependance - which is more discussed from the angle of the variance carried by the principal components, a very indirect way of looking at spatial correlation. Furthermore, I expect the spatial dependance to be non-stationary, e.g. very different in western Norway than in other parts of the domain. How is this addressed?
For instance, section 2.2 mentions the use of EOFs for the large-scale data and PCA for the station data. It is not clear why a different analysis was done for each data type.
The validation of the results is mostly done visually, with a few assessment metrics given along the text in section 2.3. I would expect results of the cross-validation to be given as detailed tables.
I guess that many of the comments above can be addressed by referring to the extensive supplementary material, but I believe that an article should be self-contained and should not require readers to go through some 160 pages of annexes to understand the main points.
Figure 5: are these statistics calibrated? In other terms, for a Normal distribution I would expect about 64% of the data to fall with the 1-standard deviation confidence interval. This does not seem to be the case here. Can this be commented?
The discussion section is long and windy. Some of it could be moved to the introduction (e.g. ll.290-300), or to the conclusion, or removed.
Typos:
l.83: Russian -> Russia
l.91.92: Specify the cause of that: some runs are not available at a daily resolution
l.257: …the presence *of* various…
l.304 (and other places): references to unpublished work should be removed,
Figures 3 and 4: the projection seems incorrect, resulting in deformed maps
Figure 6: The legend foes not match the figure
Citation: https://doi.org/10.5194/egusphere-2024-1463-RC1 -
AC1: 'Reply on RC1', Rasmus Benestad, 18 Jul 2024
We are grateful for the comments from the first reviewer which we find constructive and useful, and hope that we have managed to revise our paper to meet any critical feedback. More detailed account is provided below. The reviewer comments are repeated here in italics.
This manuscript is addressing an important topic and is well-written. However, I would recommend major revisions to give more details on the main body of the manuscript rather than in the SI. I think that the manuscript should be rewritten for the broad readership of HESS, or alternatively be submitted to a more specialized journal."
Thank you for this comment. The paper is intended for the broad readership of HESS, and we hope the revised manuscript will make it accessible to most readers."The essence of the approach presented is to build a statistical model that relates large-scale information (EA5 or CMIP6) to station-level data. Then, this model is used to estimate the future climate at stations locations, based on CMIP6. I am not sure this approach can be termed downscaling, as in my mind downscaling is the passage of data from a coarse grid to a finer grid. Here the correct term, in my understanding, would be something like change of support. In particular, the data are only produced at the stations locations, meaning that it is less effective in areas with a lower station density (such as northern Norway/Finland)."
Thank you for this comment - there are different ways of thinking of downscaling, and the revised paper will explain these different positions, as the passage to a finer grid is not the only correct one. It’s also beneficial with an increased awareness about what we mean by downscaling, and there are many examples of the use of downscaling in the scientific literature, some which are in line with our definition. We interpret all station data as a quantification of local small-scale conditions and there are some cases where global climate model simulations have been downscaled to a single point."Moreover, the details of the methodology are very condensed and not comprehensible for a reader that is not already familiar with it. The method is summarized very shortly in section 2.2, and mostly referred to through citations to other papers by the authors, with whom the reader is not necessarily familiar. Therefore, it is difficult to grasp it by only reading the manuscript. A schematic of the methodology would be helpful."
We are grateful for this constructive feedback. The revised manuscript will include a new appendix that lays out the method in more details and will present schematics illustrating the method. The reason why we go for an appendix is to keep the preamble to a minimum for readers already familiar with the methodology and still make the detailed information available for the readers without the need to read the supporting material. There will still be some supporting material with additional analysis and plots as well as a complete transcript of the analysis in the shape of an R-markdown script and its output (PDF-document)."Very few details are given on the kriging procedure used. Which variogram parameters were used, and what does this tell us about the spatial dependency - which is more discussed from the angle of the variance carried by the principal components, a very indirect way of looking at spatial correlation. Furthermore, I expect the spatial dependancy to be non-stationary, e.g. very different in western Norway than in other parts of the domain. How is this addressed?"
The revised paper will provide more details about the kriging in a new appendix. The kriging routine was developed at UCAR in the US for gridding climate data and is a multiresolution kriging method based on Markov random fields. It uses a large set of basis functions to estimate spatial estimates that are comparable to standard families of covariance functions (Nychka et al., 2016). “LatticeKrig: Multiresolution Kriging Based on Markov Random Fields.” doi:10.5065/D6HD7T1R <https://doi.org/10.5065/D6HD7T1R>). We used this technique as provided and the gridding aspect was not the main focus of our paper as our main thrust of our work is the downscaling of precipitation statistics for the locations with rain gauge data. Any spatial non-stationarity will be captured by different modes of the PCA - it’s essentially a machine learning method that adapts to the information hidden in the data."For instance, section 2.2 mentions the use of EOFs for the large-scale data and PCA for the station data. It is not clear why a different analysis was done for each data type."
Thanks for this comment, and this will be clarified in the revised paper. We reserve the term EOF for regularly gridded data as most cases in literature, using area-based weights for the grid boxes. PCA on the other hand is reserved for irregular groups of coordinates typical for stations and does not involve area weighting. Hence, PCA tends to give more weight to regions with dense networks of measurements."The validation of the results is mostly done visually, with a few assessment metrics given along the text in section 2.3. I would expect results of the cross-validation to be given as detailed tables."
Good point. We will include tables of cross-validation results in the revised paper, in the appendix explaining the method."I guess that many of the comments above can be addressed by referring to the extensive supplementary material, but I believe that an article should be self-contained and should not require readers to go through some 160 pages of annexes to understand the main points."
We agree, and will provide all necessary information in the appendices. However, we will also make the output of our R-markdown script available from FigShare for the sake of transparency and replicability."Figure 5: are these statistics calibrated? In other terms, for a Normal distribution I would expect about 64% of the data to fall with the 1-standard deviation confidence interval. This does not seem to be the case here. Can this be commented on?"
Thanks for this question. The results were not calibrated but were based on probabilities estimated from downscaled annual wet-day frequency and annual wet-day mean precipitation for Oslo-Blindern. The results are slightly biased, but the approximate match also indicates that the method does provide sensible numbers. This will be explained more carefully in the revised version."The discussion section is long and windy. Some of it could be moved to the introduction (e.g. ll.290-300), or to the conclusion, or removed."
Good idea. Some of the text has been moved to appendices."l.83: Russian -> Russia": Thanks - fixed in revised version
"l.91.92: Specify the cause of that: some runs are not available at a daily resolution": The data stored on ESGF appears to include a large number of runs, but the shear data volume makes it practically impossible to use all of it to compute monthly wet-day frequency and wet-day mean precipitation. So since these monthly statistics are not readily available, we kept the selection of runs to include only one per model type. Also, our processing of the data revealed some problems that appeared during the data processing for some runs and SSPs, and some of the runs had suspect data that we excluded. This is perhaps not surprising with vast data volumes that contain data not often visited.
"l.257: …the presence *of* various…": Thanks - fixed in revised version.
"l.304 (and other places): references to unpublished work should be removed": The two references to unpublished work has been left out in the revised paper.
"Figures 3 and 4: the projection seems incorrect, resulting in deformed maps": We understand this comment to regard the geographical projection rather than the projection of future estimates. The figures are a bit squashed in the present version, but will be fixed in the revised paper.
"Figure 6: The legend does not match the figure": Thanks for noticing this, the figure caption is fixed in the revised version.
Citation: https://doi.org/10.5194/egusphere-2024-1463-AC1
-
AC1: 'Reply on RC1', Rasmus Benestad, 18 Jul 2024
-
RC2: 'Comment on egusphere-2024-1463', Anonymous Referee #2, 04 Jul 2024
The subject is valuable, the data used is pertinent, and the paper is interesting and well-written, but major revisions are necessary. The document is not self-sufficient, and relies too much on the supplementary material.
An example is the difficulty to assess the method's quality due to insufficient description of the methodology and evaluation parts.
These sections should be enhanced by adding more details, including figures.
Key details to add in the method part should include the amount of data used to calibrate the regression model (i.e., train the model), the amount of data used to evaluate the model's performance, the size of the data used as predictors (inputs), and the size of the data used as predictands (outputs).
The evaluation on ERA5 data is crucial and should be comprehensively included in the main document. References to supporting material should be limited to more detailed evaluations. For example, clarifications are needed for the statement about the “close match between the rain gauge data and ERA5” (Line 134). It should specify whether this evaluation was done on data used to calibrate the regression model.
Other comments
L71-77 : It seems the authors are describing a bias/variance tradeoff. What I understand is that it is desired to avoid fitting the predictands too closely with statistical techniques to avoid overfitting (high variance, where models become overly sensitive to data used for training/calibrating). It would be helpful to clarify if the exactitude (precision and accuracy in capturing detailed patterns - in calibration data, akin to high variance ?) versus robustness (the ability to generalise to new data, capturing essential patterns) tradeoff is similar to a bias/variance tradeoff. Additionally, the authors describe their strategy as “obtaining a model with high robustness at the cost of precision” but do not detail what happens after this choice. I would expect some specifics on the chosen model in this introduction.L144: “The evaluation also involved testing the ability of the GCMs in reproducing the predictors in a skillful way” needs clarification. I assumed the GCMs were the predictors used for projections. From the context, it seems the GCMs are compared against ERA5 to verify if the model trained with ERA5 data could be applied using GCM predictors.
Typos
L83 : “the highest point being 2062 above” missing unit
L257: missing “of” between various and meteorological
Citation: https://doi.org/10.5194/egusphere-2024-1463-RC2 -
AC2: 'Reply on RC2', Rasmus Benestad, 18 Jul 2024
We are grateful for the comments from the second reviewer which we find constructive and useful, and hope that we have managed to revise our paper to meet any critical feedback. More detailed account is provided below. Our response will be in bold font and we will repeat the comments in italics.
"The subject is valuable, the data used is pertinent, and the paper is interesting and well-written, but major revisions are necessary. The document is not self-sufficient, and relies too much on the supplementary material."
We understand this objection and a new appendix will provide detailed information in the revised version so that it does not rely on the supplementary material."An example is the difficulty to assess the method's quality due to insufficient description of the methodology and evaluation parts."
Thanks for this feedback. We will provide this information in the new appendix of the revised paper."These sections should be enhanced by adding more details, including figures."
We think this is a good idea and a new appendix will be included in the revised paper with infographics illustrating the method and with more details about the method. We opted for an appendix that is accompanying the main text to reduce the preamble for readers already familiar with the method while providing explanations and necessary information for other readers without the need to look up separate sources."Key details to add in the method part should include the amount of data used to calibrate the regression model (i.e., train the model), the amount of data used to evaluate the model's performance, the size of the data used as predictors (inputs), and the size of the data used as predictands (outputs)."
Good point. This will be provided as part of the detailed method description in a new appendix."The evaluation on ERA5 data is crucial and should be comprehensively included in the main document. References to supporting material should be limited to more detailed evaluations. For example, clarifications are needed for the statement about the “close match between the rain gauge data and ERA5” (Line 134). It should specify whether this evaluation was done on data used to calibrate the regression model."
Good point. A separate appendix will be dedicated the evaluation of ERA5 as well as the GCMs. We think the use of appendices makes the paper easier to read for readers who are not so focused on the ERA5 quality, but readily accessible for all readers nevertheless."L71-77 : It seems the authors are describing a bias/variance tradeoff. What I understand is that it is desired to avoid fitting the predictands too closely with statistical techniques to avoid overfitting (high variance, where models become overly sensitive to data used for training/calibrating). It would be helpful to clarify if the exactitude (precision and accuracy in capturing detailed patterns - in calibration data, akin to high variance ?) versus robustness (the ability to generalise to new data, capturing essential patterns) tradeoff is similar to a bias/variance tradeoff. Additionally, the authors describe their strategy as “obtaining a model with high robustness at the cost of precision” but do not detail what happens after this choice. I would expect some specifics on the chosen model in this introduction."
Thanks for asking this. This is not really about bias/variance tradeoff, but using a simpler expression with less parameters for estimating probabilities. In that sense, it’s less advanced than other expressions with a larger set of parameters to fit, but also there is less room for getting a closer match than more advanced expressions. The simple expression is more approximate but also less sensitive to errors. We have tried to rephrase the text to improve it in the revised version."L144: “The evaluation also involved testing the ability of the GCMs in reproducing the predictors in a skillful way” needs clarification. I assumed the GCMs were the predictors used for projections. From the context, it seems the GCMs are compared against ERA5 to verify if the model trained with ERA5 data could be applied using GCM predictors."
Thanks for this question. Yes, predictors from GCMs were used for projections and the point was to ensure that the GCMs were able to reproduce them in a credible manner. They were indeed evaluated against ERA5, which again was evaluated against the rain gauge data. We will revise the paper to make this point more clear."L83 : “the highest point being 2062 above” missing unit": Thanks, unit is m. This is fixed in the revised version.
"L257: missing “of” between various and meteorological": Thanks, this is fixed in the revised version.
Citation: https://doi.org/10.5194/egusphere-2024-1463-AC2 -
AC3: 'Reply on RC2', Rasmus Benestad, 18 Jul 2024
We are grateful for the comments from the second reviewer which we find constructive and useful, and hope that we have managed to revise our paper to meet any critical feedback. More detailed account is provided below. Our response will be in bold font and we will repeat the comments in italics.
"The subject is valuable, the data used is pertinent, and the paper is interesting and well-written, but major revisions are necessary. The document is not self-sufficient, and relies too much on the supplementary material."
We understand this objection and a new appendix will provide detailed information in the revised version so that it does not rely on the supplementary material."An example is the difficulty to assess the method's quality due to insufficient description of the methodology and evaluation parts."
Thanks for this feedback. We will provide this information in the new appendix of the revised paper."These sections should be enhanced by adding more details, including figures."
We think this is a good idea and a new appendix will be included in the revised paper with infographics illustrating the method and with more details about the method. We opted for an appendix that is accompanying the main text to reduce the preamble for readers already familiar with the method while providing explanations and necessary information for other readers without the need to look up separate sources."Key details to add in the method part should include the amount of data used to calibrate the regression model (i.e., train the model), the amount of data used to evaluate the model's performance, the size of the data used as predictors (inputs), and the size of the data used as predictands (outputs)."
Good point. This will be provided as part of the detailed method description in a new appendix."The evaluation on ERA5 data is crucial and should be comprehensively included in the main document. References to supporting material should be limited to more detailed evaluations. For example, clarifications are needed for the statement about the “close match between the rain gauge data and ERA5” (Line 134). It should specify whether this evaluation was done on data used to calibrate the regression model."
Good point. A separate appendix will be dedicated the evaluation of ERA5 as well as the GCMs. We think the use of appendices makes the paper easier to read for readers who are not so focused on the ERA5 quality, but readily accessible for all readers nevertheless."L71-77 : It seems the authors are describing a bias/variance tradeoff. What I understand is that it is desired to avoid fitting the predictands too closely with statistical techniques to avoid overfitting (high variance, where models become overly sensitive to data used for training/calibrating). It would be helpful to clarify if the exactitude (precision and accuracy in capturing detailed patterns - in calibration data, akin to high variance ?) versus robustness (the ability to generalise to new data, capturing essential patterns) tradeoff is similar to a bias/variance tradeoff. Additionally, the authors describe their strategy as “obtaining a model with high robustness at the cost of precision” but do not detail what happens after this choice. I would expect some specifics on the chosen model in this introduction."
Thanks for asking this. This is not really about bias/variance tradeoff, but using a simpler expression with less parameters for estimating probabilities. In that sense, it’s less advanced than other expressions with a larger set of parameters to fit, but also there is less room for getting a closer match than more advanced expressions. The simple expression is more approximate but also less sensitive to errors. We have tried to rephrase the text to improve it in the revised version."L144: “The evaluation also involved testing the ability of the GCMs in reproducing the predictors in a skillful way” needs clarification. I assumed the GCMs were the predictors used for projections. From the context, it seems the GCMs are compared against ERA5 to verify if the model trained with ERA5 data could be applied using GCM predictors."
Thanks for this question. Yes, predictors from GCMs were used for projections and the point was to ensure that the GCMs were able to reproduce them in a credible manner. They were indeed evaluated against ERA5, which again was evaluated against the rain gauge data. We will revise the paper to make this point more clear."L83 : “the highest point being 2062 above” missing unit": Thanks, unit is m. This is fixed in the revised version.
"L257: missing “of” between various and meteorological": Thanks, this is fixed in the revised version.
Citation: https://doi.org/10.5194/egusphere-2024-1463-AC3
-
AC2: 'Reply on RC2', Rasmus Benestad, 18 Jul 2024
-
RC3: 'Comment on egusphere-2024-1463', Anonymous Referee #3, 08 Jul 2024
The study presents a statistical methodology on parameterizing the average annual probability of rainy days (fw), and the annual average daily precipitation (μ) at a rain gauge scale based on co-variates extracted by larger scale models (a procedure called downscaling in this paper). Subsequently fw and μ were used to parameterize daily rainfall extremes, assuming an exponential distribution of daily depths and a parametric form of IDF curves based on fw and μ. Under present climate the statistical methodology was trained using rain gauge observation and ERA5 reanalysis. The methodology was then used as an extrapolation of the future, where climate covariates were obtained from CMIP6 simulations.
The statistical methodology presented here is definitely of interest to the hydrological community, particularly for applications involving design against extremes. However I do find that the paper needs substantial improvements to be suggested for publication.
Major comments:
- The first major concern about the paper is the presentation of the methodology. While the methodology itself is straightforward, its presentation is not clear. Large parts of the methodological description point to previous publications from the lead author, instead of actually describing the method, making it very hard to follow. The use of R-code references and jargon in the methodology itself is also not particularly helpful. Very important information that needs to be much more clearly presented is: (a) what is the spatial resolution of each of the data sets? (b) which covariates were used from the climate models and at which spatial/temporal scales? (c) Why were those covariates chosen? This information needs to be part of the main manuscript and not the supplementary information.
- GCMs and ERA5 do not have the same spatial/temporal scales. Most important ERA5 is a data assimilation scheme that integrates observations, in contrast GCMs. The study needs to address how information transfer from ERA5 to the rain gauge scale is similar to that from the GCM to rain gauge scale. Particularly as ERA5 carries a lot of information, not only from modelled atmospheric/land-surface/ocean dynamics, but also in-situ and remote sensing observations.
- I agree with the first reviewer that the term downscaling might be confusing in the context of this study. Downscaling would in most cases refer to methodologies refining/disaggregating rainfall from large to smaller scales. What is presented here is the estimation of fine spatial scale statistics based on climate models. The authors need to clarify that early in the manuscript.
- Extreme rainfall statistics were derived under the assumption that rainfall is exponentially distributed. It is known that an exponential distribution with light tails can significantly underestimate extremes. The authors need to better support this decision (beyond mathematical tractability).
- An assessment of the performance of the exponential distribution, as well as the skill of the parametric form of the IDF curve needs to be presented in the main manuscript for the entire study domain.
- I am not particularly confident that the structure of the supporting information is helpful for understanding the study. I would strongly encourage the authors to separate the code development/application part in a technical reference from the remaining of the supporting material.
Citation: https://doi.org/10.5194/egusphere-2024-1463-RC3 -
CC1: 'Reply on RC3', Rasmus Benestad, 18 Jul 2024
We are grateful for the comments from the third reviewer which we find constructive and useful, and hope that we have managed to revise our paper to meet any critical feedback. More detailed account is provided below. Our response is in bold font and we repeat the comments in italics.
"The first major concern about the paper is the presentation of the methodology. While the methodology itself is straightforward, its presentation is not clear. Large parts of the methodological description point to previous publications from the lead author, instead of actually describing the method, making it very hard to follow. The use of R-code references and jargon in the methodology itself is also not particularly helpful. Very important information that needs to be much more clearly presented is: (a) what is the spatial resolution of each of the data sets? (b) which covariates were used from the climate models and at which spatial/temporal scales? (c) Why were those covariates chosen? This information needs to be part of the main manuscript and not the supplementary information."
Point taken, and the revised paper will be accompanied with new appendices, which will provide a schematic illustrating the methods and such details. We opt for an appendix accompanying the main paper rather than in the main part to make this information readily available for the interested readers but not being a barrier with “too much information” for others.
"GCMs and ERA5 do not have the same spatial/temporal scales. Most important ERA5 is a data assimilation scheme that integrates observations, in contrast GCMs. The study needs to address how information transfer from ERA5 to the rain gauge scale is similar to that from the GCM to rain gauge scale. Particularly as ERA5 carries a lot of information, not only from modelled atmospheric/land-surface/ocean dynamics, but also in-situ and remote sensing observations."
Thanks for asking this question. The revised paper will more carefully explain the data processing and downscaling approach, where data is regridded and used together in the shape of “common EOFs”. These common EOFs are the basis for our predictors and the input to multiple regression models that are calibrated with the local rain gauge measurements."I agree with the first reviewer that the term downscaling might be confusing in the context of this study. Downscaling would in most cases refer to methodologies refining/disaggregating rainfall from large to smaller scales. What is presented here is the estimation of fine spatial scale statistics based on climate models. The authors need to clarify that early in the manuscript."
There have been different ways of thinking about downscaling which have been discussed in previous publications cited in our paper, and merely refining to higher resolution can be done through spatial interpolation and not necessarily utilise the information about inter-dependencies between different spatial scales (e.g. physical phenomena with different characteristic scales). Also, regridding does not address the problem is about the models' minimum skillful scale, discussed in earlier publications (e.g. Takayabu et al., 2015). The revised paper will explain more carefully how downscaling can be defined, and we think that raising the awareness of different definitions of downscaling can be useful."Extreme rainfall statistics were derived under the assumption that rainfall is exponentially distributed. It is known that an exponential distribution with light tails can significantly underestimate extremes. The authors need to better support this decision (beyond mathematical tractability)."
We think this comment is a bit hasty as the paper is careful to present the statistics as heavy rainfall - not ‘extreme’ but as ‘moderate extreme’. The cited papers demonstrate the utility of this approach, and the general concerns regarding exponential distribution with light tails can significantly underestimate extremes is addressed by including an empirical scaling factor α. There is no need to repeat the empirical validation here that has already been published in Benestad el al. (2019;DOI:10.1088/1748-9326/ab2bb2 and open access) and involved 1875 rain gauge records from North America nad Europe with more than 50 years of valid data over the period 1961–2018.
"An assessment of the performance of the exponential distribution, as well as the skill of the parametric form of the IDF curve needs to be presented in the main manuscript for the entire study domain."
Such assessments have already been carried out and published in papers cited in our paper (e.g. DOI:10.1088/1748-9326/abd4ab and DOI:10.1088/1748-9326/abd4ab), and the revised paper will explain this more carefully. While the performance of the exponential distribution has been tested for both Europe (including the Nordics) and North America, the expression for estimating IDF curves have only been calibrated and evaluated against Norwegian sub-daily data (e.g Parding et al, DOI:10.5194/hess-27-3719-2023). The revised version will explain this point more carefully, but we don’t think it’s necessary to repeat this assessment in the present paper."I am not particularly confident that the structure of the supporting information is helpful for understanding the study. I would strongly encourage the authors to separate the code development/application part in a technical reference from the remaining of the supporting material."
This is a good idea, and we will reorganise the content in a revised version, using appendices for information cited in the paper, but still provide the output of the R-markdown for those who’d be interested to repeat our study. It can be regarded as a form of “lab notebook” for the sake of traceability and reproducibility. Moreover, now the essential details are presented in an appendix with a more suitable structure.Citation: https://doi.org/10.5194/egusphere-2024-1463-CC1 -
AC4: 'Final respond to all referee comments', Rasmus Benestad, 20 Aug 2024
We are grateful for the review comments from all three reviewers and hope that our revised manuscript addresses all the points made by the reviewers. The revised paper has a different structure with extensive use of appendixes to provide more details and discussion about downscaling. It is now stated that we define downscaling as a process that utilises large-scale information that the global climate models skillfully reproduce to derive local information, a definition that also has been used elsewhere in the scientific literature (we provide a couple of prominent examples of such references in our paper). One suggestion that we didn’t follow was to assess the approximation based on the exponential distribution, as it has already been done for numerous sites in North America and Europe and the results are available in the cited open-access paper. However, we have explained in the revised paper that the method has already been evaluated and the results are published in an open-access paper. The formula for estimating intensity-duration-frequency (IDF) curves has also recently been assessed for sites in Norway, and we cite the open-access paper with those results rather than repeating them in our paper. Apart from these points, we have revised the paper according to the points raised by the reviewers. We think this paper should be of interest to the HESS readership and the hydrology community in general, as we present an approach for downscaling robust information on heavy precipitation statistics.
Citation: https://doi.org/10.5194/egusphere-2024-1463-AC4 -
AC5: 'Final respond to all referee comments', Rasmus Benestad, 20 Aug 2024
We are grateful for the review comments from all three reviewers and hope that our revised manuscript addresses all the points made by the reviewers. The revised paper has a different structure with extensive use of appendixes to provide more details and discussion about downscaling. It is now stated that we define downscaling as a process that utilises large-scale information that the global climate models skillfully reproduce to derive local information, a definition that also has been used elsewhere in the scientific literature (we provide a couple of prominent examples of such references in our paper). One suggestion that we didn’t follow was to assess the approximation based on the exponential distribution, as it has already been done for numerous sites in North America and Europe and the results are available in the cited open-access paper. However, we have explained in the revised paper that the method has already been evaluated and the results are published in an open-access paper. The formula for estimating intensity-duration-frequency (IDF) curves has also recently been assessed for sites in Norway, and we cite the open-access paper with those results rather than repeating them in our paper. Apart from these points, we have revised the paper according to the points raised by the reviewers. We think this paper should be of interest to the HESS readership and the hydrology community in general, as we present an approach for downscaling robust information on heavy precipitation statistics.
Citation: https://doi.org/10.5194/egusphere-2024-1463-AC5
Status: closed
-
RC1: 'Comment on egusphere-2024-1463', Anonymous Referee #1, 20 Jun 2024
Review of “Downscaling the probability of heavy rainfall over the Nordic Countries”, by Benestad et al.
This manuscript is addressing an important topic and is well-written. However, I would recommend major revisions to give more details on the main body of the manuscript rather than in the SI. I think that the manuscript should be rewritten for the broad readership of HESS, or alternatively be submitted to an more specialized journal.
The essence of the approach presented is to build a statistical model that relates large-scale information (EA5 or CMIP6) to station-level data. Then, this model is used to estimate the future climate at stations locations, based on CMIP6. I am not sure this approach can be termed downscaling, as in my mind downscaling is the passage of data from a coarse grid to a finer grid. Here the correct term, in my understanding, would be something like change of support. In particular, the data are only produced at the stations locations, meaning that it is less effective in areas with a lower stations density (such as northern Norway/Finland).
Moreover, the details of the methodology are very condensed and not comprehensible for a reader that is not already familiar with it. The method is summarized very shortly in section 2.2, and mostly referred to through citations to other papers by the authors, with the reader is not necessarily familiar. Therefore, it is difficult to grasp it by only reading the manuscript. A schematic of the methodology would be helpful.
Very few details are given on the kriging procedure used. Which variogram parameters were used, and what does this tell us about the spatial dependance - which is more discussed from the angle of the variance carried by the principal components, a very indirect way of looking at spatial correlation. Furthermore, I expect the spatial dependance to be non-stationary, e.g. very different in western Norway than in other parts of the domain. How is this addressed?
For instance, section 2.2 mentions the use of EOFs for the large-scale data and PCA for the station data. It is not clear why a different analysis was done for each data type.
The validation of the results is mostly done visually, with a few assessment metrics given along the text in section 2.3. I would expect results of the cross-validation to be given as detailed tables.
I guess that many of the comments above can be addressed by referring to the extensive supplementary material, but I believe that an article should be self-contained and should not require readers to go through some 160 pages of annexes to understand the main points.
Figure 5: are these statistics calibrated? In other terms, for a Normal distribution I would expect about 64% of the data to fall with the 1-standard deviation confidence interval. This does not seem to be the case here. Can this be commented?
The discussion section is long and windy. Some of it could be moved to the introduction (e.g. ll.290-300), or to the conclusion, or removed.
Typos:
l.83: Russian -> Russia
l.91.92: Specify the cause of that: some runs are not available at a daily resolution
l.257: …the presence *of* various…
l.304 (and other places): references to unpublished work should be removed,
Figures 3 and 4: the projection seems incorrect, resulting in deformed maps
Figure 6: The legend foes not match the figure
Citation: https://doi.org/10.5194/egusphere-2024-1463-RC1 -
AC1: 'Reply on RC1', Rasmus Benestad, 18 Jul 2024
We are grateful for the comments from the first reviewer which we find constructive and useful, and hope that we have managed to revise our paper to meet any critical feedback. More detailed account is provided below. The reviewer comments are repeated here in italics.
This manuscript is addressing an important topic and is well-written. However, I would recommend major revisions to give more details on the main body of the manuscript rather than in the SI. I think that the manuscript should be rewritten for the broad readership of HESS, or alternatively be submitted to a more specialized journal."
Thank you for this comment. The paper is intended for the broad readership of HESS, and we hope the revised manuscript will make it accessible to most readers."The essence of the approach presented is to build a statistical model that relates large-scale information (EA5 or CMIP6) to station-level data. Then, this model is used to estimate the future climate at stations locations, based on CMIP6. I am not sure this approach can be termed downscaling, as in my mind downscaling is the passage of data from a coarse grid to a finer grid. Here the correct term, in my understanding, would be something like change of support. In particular, the data are only produced at the stations locations, meaning that it is less effective in areas with a lower station density (such as northern Norway/Finland)."
Thank you for this comment - there are different ways of thinking of downscaling, and the revised paper will explain these different positions, as the passage to a finer grid is not the only correct one. It’s also beneficial with an increased awareness about what we mean by downscaling, and there are many examples of the use of downscaling in the scientific literature, some which are in line with our definition. We interpret all station data as a quantification of local small-scale conditions and there are some cases where global climate model simulations have been downscaled to a single point."Moreover, the details of the methodology are very condensed and not comprehensible for a reader that is not already familiar with it. The method is summarized very shortly in section 2.2, and mostly referred to through citations to other papers by the authors, with whom the reader is not necessarily familiar. Therefore, it is difficult to grasp it by only reading the manuscript. A schematic of the methodology would be helpful."
We are grateful for this constructive feedback. The revised manuscript will include a new appendix that lays out the method in more details and will present schematics illustrating the method. The reason why we go for an appendix is to keep the preamble to a minimum for readers already familiar with the methodology and still make the detailed information available for the readers without the need to read the supporting material. There will still be some supporting material with additional analysis and plots as well as a complete transcript of the analysis in the shape of an R-markdown script and its output (PDF-document)."Very few details are given on the kriging procedure used. Which variogram parameters were used, and what does this tell us about the spatial dependency - which is more discussed from the angle of the variance carried by the principal components, a very indirect way of looking at spatial correlation. Furthermore, I expect the spatial dependancy to be non-stationary, e.g. very different in western Norway than in other parts of the domain. How is this addressed?"
The revised paper will provide more details about the kriging in a new appendix. The kriging routine was developed at UCAR in the US for gridding climate data and is a multiresolution kriging method based on Markov random fields. It uses a large set of basis functions to estimate spatial estimates that are comparable to standard families of covariance functions (Nychka et al., 2016). “LatticeKrig: Multiresolution Kriging Based on Markov Random Fields.” doi:10.5065/D6HD7T1R <https://doi.org/10.5065/D6HD7T1R>). We used this technique as provided and the gridding aspect was not the main focus of our paper as our main thrust of our work is the downscaling of precipitation statistics for the locations with rain gauge data. Any spatial non-stationarity will be captured by different modes of the PCA - it’s essentially a machine learning method that adapts to the information hidden in the data."For instance, section 2.2 mentions the use of EOFs for the large-scale data and PCA for the station data. It is not clear why a different analysis was done for each data type."
Thanks for this comment, and this will be clarified in the revised paper. We reserve the term EOF for regularly gridded data as most cases in literature, using area-based weights for the grid boxes. PCA on the other hand is reserved for irregular groups of coordinates typical for stations and does not involve area weighting. Hence, PCA tends to give more weight to regions with dense networks of measurements."The validation of the results is mostly done visually, with a few assessment metrics given along the text in section 2.3. I would expect results of the cross-validation to be given as detailed tables."
Good point. We will include tables of cross-validation results in the revised paper, in the appendix explaining the method."I guess that many of the comments above can be addressed by referring to the extensive supplementary material, but I believe that an article should be self-contained and should not require readers to go through some 160 pages of annexes to understand the main points."
We agree, and will provide all necessary information in the appendices. However, we will also make the output of our R-markdown script available from FigShare for the sake of transparency and replicability."Figure 5: are these statistics calibrated? In other terms, for a Normal distribution I would expect about 64% of the data to fall with the 1-standard deviation confidence interval. This does not seem to be the case here. Can this be commented on?"
Thanks for this question. The results were not calibrated but were based on probabilities estimated from downscaled annual wet-day frequency and annual wet-day mean precipitation for Oslo-Blindern. The results are slightly biased, but the approximate match also indicates that the method does provide sensible numbers. This will be explained more carefully in the revised version."The discussion section is long and windy. Some of it could be moved to the introduction (e.g. ll.290-300), or to the conclusion, or removed."
Good idea. Some of the text has been moved to appendices."l.83: Russian -> Russia": Thanks - fixed in revised version
"l.91.92: Specify the cause of that: some runs are not available at a daily resolution": The data stored on ESGF appears to include a large number of runs, but the shear data volume makes it practically impossible to use all of it to compute monthly wet-day frequency and wet-day mean precipitation. So since these monthly statistics are not readily available, we kept the selection of runs to include only one per model type. Also, our processing of the data revealed some problems that appeared during the data processing for some runs and SSPs, and some of the runs had suspect data that we excluded. This is perhaps not surprising with vast data volumes that contain data not often visited.
"l.257: …the presence *of* various…": Thanks - fixed in revised version.
"l.304 (and other places): references to unpublished work should be removed": The two references to unpublished work has been left out in the revised paper.
"Figures 3 and 4: the projection seems incorrect, resulting in deformed maps": We understand this comment to regard the geographical projection rather than the projection of future estimates. The figures are a bit squashed in the present version, but will be fixed in the revised paper.
"Figure 6: The legend does not match the figure": Thanks for noticing this, the figure caption is fixed in the revised version.
Citation: https://doi.org/10.5194/egusphere-2024-1463-AC1
-
AC1: 'Reply on RC1', Rasmus Benestad, 18 Jul 2024
-
RC2: 'Comment on egusphere-2024-1463', Anonymous Referee #2, 04 Jul 2024
The subject is valuable, the data used is pertinent, and the paper is interesting and well-written, but major revisions are necessary. The document is not self-sufficient, and relies too much on the supplementary material.
An example is the difficulty to assess the method's quality due to insufficient description of the methodology and evaluation parts.
These sections should be enhanced by adding more details, including figures.
Key details to add in the method part should include the amount of data used to calibrate the regression model (i.e., train the model), the amount of data used to evaluate the model's performance, the size of the data used as predictors (inputs), and the size of the data used as predictands (outputs).
The evaluation on ERA5 data is crucial and should be comprehensively included in the main document. References to supporting material should be limited to more detailed evaluations. For example, clarifications are needed for the statement about the “close match between the rain gauge data and ERA5” (Line 134). It should specify whether this evaluation was done on data used to calibrate the regression model.
Other comments
L71-77 : It seems the authors are describing a bias/variance tradeoff. What I understand is that it is desired to avoid fitting the predictands too closely with statistical techniques to avoid overfitting (high variance, where models become overly sensitive to data used for training/calibrating). It would be helpful to clarify if the exactitude (precision and accuracy in capturing detailed patterns - in calibration data, akin to high variance ?) versus robustness (the ability to generalise to new data, capturing essential patterns) tradeoff is similar to a bias/variance tradeoff. Additionally, the authors describe their strategy as “obtaining a model with high robustness at the cost of precision” but do not detail what happens after this choice. I would expect some specifics on the chosen model in this introduction.L144: “The evaluation also involved testing the ability of the GCMs in reproducing the predictors in a skillful way” needs clarification. I assumed the GCMs were the predictors used for projections. From the context, it seems the GCMs are compared against ERA5 to verify if the model trained with ERA5 data could be applied using GCM predictors.
Typos
L83 : “the highest point being 2062 above” missing unit
L257: missing “of” between various and meteorological
Citation: https://doi.org/10.5194/egusphere-2024-1463-RC2 -
AC2: 'Reply on RC2', Rasmus Benestad, 18 Jul 2024
We are grateful for the comments from the second reviewer which we find constructive and useful, and hope that we have managed to revise our paper to meet any critical feedback. More detailed account is provided below. Our response will be in bold font and we will repeat the comments in italics.
"The subject is valuable, the data used is pertinent, and the paper is interesting and well-written, but major revisions are necessary. The document is not self-sufficient, and relies too much on the supplementary material."
We understand this objection and a new appendix will provide detailed information in the revised version so that it does not rely on the supplementary material."An example is the difficulty to assess the method's quality due to insufficient description of the methodology and evaluation parts."
Thanks for this feedback. We will provide this information in the new appendix of the revised paper."These sections should be enhanced by adding more details, including figures."
We think this is a good idea and a new appendix will be included in the revised paper with infographics illustrating the method and with more details about the method. We opted for an appendix that is accompanying the main text to reduce the preamble for readers already familiar with the method while providing explanations and necessary information for other readers without the need to look up separate sources."Key details to add in the method part should include the amount of data used to calibrate the regression model (i.e., train the model), the amount of data used to evaluate the model's performance, the size of the data used as predictors (inputs), and the size of the data used as predictands (outputs)."
Good point. This will be provided as part of the detailed method description in a new appendix."The evaluation on ERA5 data is crucial and should be comprehensively included in the main document. References to supporting material should be limited to more detailed evaluations. For example, clarifications are needed for the statement about the “close match between the rain gauge data and ERA5” (Line 134). It should specify whether this evaluation was done on data used to calibrate the regression model."
Good point. A separate appendix will be dedicated the evaluation of ERA5 as well as the GCMs. We think the use of appendices makes the paper easier to read for readers who are not so focused on the ERA5 quality, but readily accessible for all readers nevertheless."L71-77 : It seems the authors are describing a bias/variance tradeoff. What I understand is that it is desired to avoid fitting the predictands too closely with statistical techniques to avoid overfitting (high variance, where models become overly sensitive to data used for training/calibrating). It would be helpful to clarify if the exactitude (precision and accuracy in capturing detailed patterns - in calibration data, akin to high variance ?) versus robustness (the ability to generalise to new data, capturing essential patterns) tradeoff is similar to a bias/variance tradeoff. Additionally, the authors describe their strategy as “obtaining a model with high robustness at the cost of precision” but do not detail what happens after this choice. I would expect some specifics on the chosen model in this introduction."
Thanks for asking this. This is not really about bias/variance tradeoff, but using a simpler expression with less parameters for estimating probabilities. In that sense, it’s less advanced than other expressions with a larger set of parameters to fit, but also there is less room for getting a closer match than more advanced expressions. The simple expression is more approximate but also less sensitive to errors. We have tried to rephrase the text to improve it in the revised version."L144: “The evaluation also involved testing the ability of the GCMs in reproducing the predictors in a skillful way” needs clarification. I assumed the GCMs were the predictors used for projections. From the context, it seems the GCMs are compared against ERA5 to verify if the model trained with ERA5 data could be applied using GCM predictors."
Thanks for this question. Yes, predictors from GCMs were used for projections and the point was to ensure that the GCMs were able to reproduce them in a credible manner. They were indeed evaluated against ERA5, which again was evaluated against the rain gauge data. We will revise the paper to make this point more clear."L83 : “the highest point being 2062 above” missing unit": Thanks, unit is m. This is fixed in the revised version.
"L257: missing “of” between various and meteorological": Thanks, this is fixed in the revised version.
Citation: https://doi.org/10.5194/egusphere-2024-1463-AC2 -
AC3: 'Reply on RC2', Rasmus Benestad, 18 Jul 2024
We are grateful for the comments from the second reviewer which we find constructive and useful, and hope that we have managed to revise our paper to meet any critical feedback. More detailed account is provided below. Our response will be in bold font and we will repeat the comments in italics.
"The subject is valuable, the data used is pertinent, and the paper is interesting and well-written, but major revisions are necessary. The document is not self-sufficient, and relies too much on the supplementary material."
We understand this objection and a new appendix will provide detailed information in the revised version so that it does not rely on the supplementary material."An example is the difficulty to assess the method's quality due to insufficient description of the methodology and evaluation parts."
Thanks for this feedback. We will provide this information in the new appendix of the revised paper."These sections should be enhanced by adding more details, including figures."
We think this is a good idea and a new appendix will be included in the revised paper with infographics illustrating the method and with more details about the method. We opted for an appendix that is accompanying the main text to reduce the preamble for readers already familiar with the method while providing explanations and necessary information for other readers without the need to look up separate sources."Key details to add in the method part should include the amount of data used to calibrate the regression model (i.e., train the model), the amount of data used to evaluate the model's performance, the size of the data used as predictors (inputs), and the size of the data used as predictands (outputs)."
Good point. This will be provided as part of the detailed method description in a new appendix."The evaluation on ERA5 data is crucial and should be comprehensively included in the main document. References to supporting material should be limited to more detailed evaluations. For example, clarifications are needed for the statement about the “close match between the rain gauge data and ERA5” (Line 134). It should specify whether this evaluation was done on data used to calibrate the regression model."
Good point. A separate appendix will be dedicated the evaluation of ERA5 as well as the GCMs. We think the use of appendices makes the paper easier to read for readers who are not so focused on the ERA5 quality, but readily accessible for all readers nevertheless."L71-77 : It seems the authors are describing a bias/variance tradeoff. What I understand is that it is desired to avoid fitting the predictands too closely with statistical techniques to avoid overfitting (high variance, where models become overly sensitive to data used for training/calibrating). It would be helpful to clarify if the exactitude (precision and accuracy in capturing detailed patterns - in calibration data, akin to high variance ?) versus robustness (the ability to generalise to new data, capturing essential patterns) tradeoff is similar to a bias/variance tradeoff. Additionally, the authors describe their strategy as “obtaining a model with high robustness at the cost of precision” but do not detail what happens after this choice. I would expect some specifics on the chosen model in this introduction."
Thanks for asking this. This is not really about bias/variance tradeoff, but using a simpler expression with less parameters for estimating probabilities. In that sense, it’s less advanced than other expressions with a larger set of parameters to fit, but also there is less room for getting a closer match than more advanced expressions. The simple expression is more approximate but also less sensitive to errors. We have tried to rephrase the text to improve it in the revised version."L144: “The evaluation also involved testing the ability of the GCMs in reproducing the predictors in a skillful way” needs clarification. I assumed the GCMs were the predictors used for projections. From the context, it seems the GCMs are compared against ERA5 to verify if the model trained with ERA5 data could be applied using GCM predictors."
Thanks for this question. Yes, predictors from GCMs were used for projections and the point was to ensure that the GCMs were able to reproduce them in a credible manner. They were indeed evaluated against ERA5, which again was evaluated against the rain gauge data. We will revise the paper to make this point more clear."L83 : “the highest point being 2062 above” missing unit": Thanks, unit is m. This is fixed in the revised version.
"L257: missing “of” between various and meteorological": Thanks, this is fixed in the revised version.
Citation: https://doi.org/10.5194/egusphere-2024-1463-AC3
-
AC2: 'Reply on RC2', Rasmus Benestad, 18 Jul 2024
-
RC3: 'Comment on egusphere-2024-1463', Anonymous Referee #3, 08 Jul 2024
The study presents a statistical methodology on parameterizing the average annual probability of rainy days (fw), and the annual average daily precipitation (μ) at a rain gauge scale based on co-variates extracted by larger scale models (a procedure called downscaling in this paper). Subsequently fw and μ were used to parameterize daily rainfall extremes, assuming an exponential distribution of daily depths and a parametric form of IDF curves based on fw and μ. Under present climate the statistical methodology was trained using rain gauge observation and ERA5 reanalysis. The methodology was then used as an extrapolation of the future, where climate covariates were obtained from CMIP6 simulations.
The statistical methodology presented here is definitely of interest to the hydrological community, particularly for applications involving design against extremes. However I do find that the paper needs substantial improvements to be suggested for publication.
Major comments:
- The first major concern about the paper is the presentation of the methodology. While the methodology itself is straightforward, its presentation is not clear. Large parts of the methodological description point to previous publications from the lead author, instead of actually describing the method, making it very hard to follow. The use of R-code references and jargon in the methodology itself is also not particularly helpful. Very important information that needs to be much more clearly presented is: (a) what is the spatial resolution of each of the data sets? (b) which covariates were used from the climate models and at which spatial/temporal scales? (c) Why were those covariates chosen? This information needs to be part of the main manuscript and not the supplementary information.
- GCMs and ERA5 do not have the same spatial/temporal scales. Most important ERA5 is a data assimilation scheme that integrates observations, in contrast GCMs. The study needs to address how information transfer from ERA5 to the rain gauge scale is similar to that from the GCM to rain gauge scale. Particularly as ERA5 carries a lot of information, not only from modelled atmospheric/land-surface/ocean dynamics, but also in-situ and remote sensing observations.
- I agree with the first reviewer that the term downscaling might be confusing in the context of this study. Downscaling would in most cases refer to methodologies refining/disaggregating rainfall from large to smaller scales. What is presented here is the estimation of fine spatial scale statistics based on climate models. The authors need to clarify that early in the manuscript.
- Extreme rainfall statistics were derived under the assumption that rainfall is exponentially distributed. It is known that an exponential distribution with light tails can significantly underestimate extremes. The authors need to better support this decision (beyond mathematical tractability).
- An assessment of the performance of the exponential distribution, as well as the skill of the parametric form of the IDF curve needs to be presented in the main manuscript for the entire study domain.
- I am not particularly confident that the structure of the supporting information is helpful for understanding the study. I would strongly encourage the authors to separate the code development/application part in a technical reference from the remaining of the supporting material.
Citation: https://doi.org/10.5194/egusphere-2024-1463-RC3 -
CC1: 'Reply on RC3', Rasmus Benestad, 18 Jul 2024
We are grateful for the comments from the third reviewer which we find constructive and useful, and hope that we have managed to revise our paper to meet any critical feedback. More detailed account is provided below. Our response is in bold font and we repeat the comments in italics.
"The first major concern about the paper is the presentation of the methodology. While the methodology itself is straightforward, its presentation is not clear. Large parts of the methodological description point to previous publications from the lead author, instead of actually describing the method, making it very hard to follow. The use of R-code references and jargon in the methodology itself is also not particularly helpful. Very important information that needs to be much more clearly presented is: (a) what is the spatial resolution of each of the data sets? (b) which covariates were used from the climate models and at which spatial/temporal scales? (c) Why were those covariates chosen? This information needs to be part of the main manuscript and not the supplementary information."
Point taken, and the revised paper will be accompanied with new appendices, which will provide a schematic illustrating the methods and such details. We opt for an appendix accompanying the main paper rather than in the main part to make this information readily available for the interested readers but not being a barrier with “too much information” for others.
"GCMs and ERA5 do not have the same spatial/temporal scales. Most important ERA5 is a data assimilation scheme that integrates observations, in contrast GCMs. The study needs to address how information transfer from ERA5 to the rain gauge scale is similar to that from the GCM to rain gauge scale. Particularly as ERA5 carries a lot of information, not only from modelled atmospheric/land-surface/ocean dynamics, but also in-situ and remote sensing observations."
Thanks for asking this question. The revised paper will more carefully explain the data processing and downscaling approach, where data is regridded and used together in the shape of “common EOFs”. These common EOFs are the basis for our predictors and the input to multiple regression models that are calibrated with the local rain gauge measurements."I agree with the first reviewer that the term downscaling might be confusing in the context of this study. Downscaling would in most cases refer to methodologies refining/disaggregating rainfall from large to smaller scales. What is presented here is the estimation of fine spatial scale statistics based on climate models. The authors need to clarify that early in the manuscript."
There have been different ways of thinking about downscaling which have been discussed in previous publications cited in our paper, and merely refining to higher resolution can be done through spatial interpolation and not necessarily utilise the information about inter-dependencies between different spatial scales (e.g. physical phenomena with different characteristic scales). Also, regridding does not address the problem is about the models' minimum skillful scale, discussed in earlier publications (e.g. Takayabu et al., 2015). The revised paper will explain more carefully how downscaling can be defined, and we think that raising the awareness of different definitions of downscaling can be useful."Extreme rainfall statistics were derived under the assumption that rainfall is exponentially distributed. It is known that an exponential distribution with light tails can significantly underestimate extremes. The authors need to better support this decision (beyond mathematical tractability)."
We think this comment is a bit hasty as the paper is careful to present the statistics as heavy rainfall - not ‘extreme’ but as ‘moderate extreme’. The cited papers demonstrate the utility of this approach, and the general concerns regarding exponential distribution with light tails can significantly underestimate extremes is addressed by including an empirical scaling factor α. There is no need to repeat the empirical validation here that has already been published in Benestad el al. (2019;DOI:10.1088/1748-9326/ab2bb2 and open access) and involved 1875 rain gauge records from North America nad Europe with more than 50 years of valid data over the period 1961–2018.
"An assessment of the performance of the exponential distribution, as well as the skill of the parametric form of the IDF curve needs to be presented in the main manuscript for the entire study domain."
Such assessments have already been carried out and published in papers cited in our paper (e.g. DOI:10.1088/1748-9326/abd4ab and DOI:10.1088/1748-9326/abd4ab), and the revised paper will explain this more carefully. While the performance of the exponential distribution has been tested for both Europe (including the Nordics) and North America, the expression for estimating IDF curves have only been calibrated and evaluated against Norwegian sub-daily data (e.g Parding et al, DOI:10.5194/hess-27-3719-2023). The revised version will explain this point more carefully, but we don’t think it’s necessary to repeat this assessment in the present paper."I am not particularly confident that the structure of the supporting information is helpful for understanding the study. I would strongly encourage the authors to separate the code development/application part in a technical reference from the remaining of the supporting material."
This is a good idea, and we will reorganise the content in a revised version, using appendices for information cited in the paper, but still provide the output of the R-markdown for those who’d be interested to repeat our study. It can be regarded as a form of “lab notebook” for the sake of traceability and reproducibility. Moreover, now the essential details are presented in an appendix with a more suitable structure.Citation: https://doi.org/10.5194/egusphere-2024-1463-CC1 -
AC4: 'Final respond to all referee comments', Rasmus Benestad, 20 Aug 2024
We are grateful for the review comments from all three reviewers and hope that our revised manuscript addresses all the points made by the reviewers. The revised paper has a different structure with extensive use of appendixes to provide more details and discussion about downscaling. It is now stated that we define downscaling as a process that utilises large-scale information that the global climate models skillfully reproduce to derive local information, a definition that also has been used elsewhere in the scientific literature (we provide a couple of prominent examples of such references in our paper). One suggestion that we didn’t follow was to assess the approximation based on the exponential distribution, as it has already been done for numerous sites in North America and Europe and the results are available in the cited open-access paper. However, we have explained in the revised paper that the method has already been evaluated and the results are published in an open-access paper. The formula for estimating intensity-duration-frequency (IDF) curves has also recently been assessed for sites in Norway, and we cite the open-access paper with those results rather than repeating them in our paper. Apart from these points, we have revised the paper according to the points raised by the reviewers. We think this paper should be of interest to the HESS readership and the hydrology community in general, as we present an approach for downscaling robust information on heavy precipitation statistics.
Citation: https://doi.org/10.5194/egusphere-2024-1463-AC4 -
AC5: 'Final respond to all referee comments', Rasmus Benestad, 20 Aug 2024
We are grateful for the review comments from all three reviewers and hope that our revised manuscript addresses all the points made by the reviewers. The revised paper has a different structure with extensive use of appendixes to provide more details and discussion about downscaling. It is now stated that we define downscaling as a process that utilises large-scale information that the global climate models skillfully reproduce to derive local information, a definition that also has been used elsewhere in the scientific literature (we provide a couple of prominent examples of such references in our paper). One suggestion that we didn’t follow was to assess the approximation based on the exponential distribution, as it has already been done for numerous sites in North America and Europe and the results are available in the cited open-access paper. However, we have explained in the revised paper that the method has already been evaluated and the results are published in an open-access paper. The formula for estimating intensity-duration-frequency (IDF) curves has also recently been assessed for sites in Norway, and we cite the open-access paper with those results rather than repeating them in our paper. Apart from these points, we have revised the paper according to the points raised by the reviewers. We think this paper should be of interest to the HESS readership and the hydrology community in general, as we present an approach for downscaling robust information on heavy precipitation statistics.
Citation: https://doi.org/10.5194/egusphere-2024-1463-AC5
Viewed
HTML | XML | Total | Supplement | BibTeX | EndNote | |
---|---|---|---|---|---|---|
339 | 100 | 116 | 555 | 50 | 15 | 12 |
- HTML: 339
- PDF: 100
- XML: 116
- Total: 555
- Supplement: 50
- BibTeX: 15
- EndNote: 12
Viewed (geographical distribution)
Country | # | Views | % |
---|
Total: | 0 |
HTML: | 0 |
PDF: | 0 |
XML: | 0 |
- 1