the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
Constraining low-frequency variability in climate projections to predict climate on decadal to multi-decadal time scales – a ‘poor-man’ initialized prediction system
Abstract. Near-term projections of climate change are subject to substantial uncertainty from internal climate variability. Here we present an approach to reduce this uncertainty by sub-selecting those ensemble members that more closely resemble observed patterns of ocean temperature variability immediately prior to a certain start date. This constraint aligns the observed and simulated variability phases and is conceptually similar to initialization in seasonal to decadal climate predictions. We apply this variability constraint to large multi-model projection ensembles from the Coupled Model Intercomparison Project phase 6 (CMIP6), consisting of more than 200 ensemble members, and evaluate the skill of the constrained ensemble in predicting the observed near-surface temperature, sea-level pressure and precipitation on decadal to multi-decadal time scales.
We find that the constrained projections show significant skill in predicting the climate of the following ten to twenty years, and added value over the ensemble of unconstrained projections. For the first decade after applying the constraint, the global patterns of skill are very similar and can even outperform those of the multi-model ensemble mean of initialized decadal hindcasts from the CMIP6 Decadal Climate Prediction Project (DCPP). In particular for temperature, larger areas show added skill in the constrained projections compared to DCPP, mainly in the Pacific and some neighboring land regions. Temperature and sea-level pressure in several regions are predictable multiple decades ahead, and show significant added value over the unconstrained projections for forecasting the first two decades and the 20-year averages. We further demonstrate the suitability of regional constraints to attribute predictability to certain ocean regions. On the example of global average temperature changes, we confirm the role of Pacific variability in modulating the reduced rate of global warming in the early 2000s, and demonstrate the predictability of reduced global warming rates over the following 15 years based on the climate conditions leading up to 1998. Our results illustrate that constraining internal variability can significantly improve the accuracy of near-term climate change estimates for the next few decades.
-
Notice on discussion status
The requested preprint has a corresponding peer-reviewed final revised paper. You are encouraged to refer to the final revised version.
-
Preprint
(1917 KB)
-
Supplement
(2133 KB)
-
The requested preprint has a corresponding peer-reviewed final revised paper. You are encouraged to refer to the final revised version.
- Preprint
(1917 KB) - Metadata XML
-
Supplement
(2133 KB) - BibTeX
- EndNote
- Final revised paper
Journal article(s) based on this preprint
Interactive discussion
Status: closed
-
RC1: 'Comment on egusphere-2022-98', Anonymous Referee #1, 11 May 2022
In my opinion this is a very well written and clear paper which presents an interesting new way of combining a large multi-model dataset in order to constrain future projections. As such I have very few comments to make and I believe that this paper should be published with only a few corrections.
The results are impressive and warrant publication. However the one area I would like to see more explanation on is the cause of the skill found. In the summary (line 300) you claim that “these results indicate that there is significant multi-decadal predictability from internal variability”. I am not convinced though that as they stand your results can justify this claim (although I’m not suggesting this is necessarily incorrect). By selecting the best 30 (or similar) simulations from the CMIP6 archive based on the 9 most recent years, could it not be possible that what you are doing is weighting your results to certain models (which is in itself a rich area of literature). This could be those models with the most realistic response to forcing, especially when outside the anomaly period at the start or end of the record. Or it could potentially be those with the most realistic modes of variability. Both of which, I would have thought could theoretically give an increase in skill over the next decades. This is especially pertinent, given that you state that the method gives a better constraint that a single model ensemble (although as you say this is expected given the increased number of simulations to sample from). I think that this point needs to be discussed, and preferably investigated. One simple test you could do, without the need for further analysis, would be to check if you are selecting simulations preferentially from one climate model or if the 30 members are selected from the full model range. In addition it would also be interesting to see if the skill varies through time.
I am also somewhat surprised that you have not shown results for future projections, since that would seems to be the logical direction for this type of analysis. Although, perhaps this is being left for future work (which would be perfectly justifiable).
More minor points:
L 84. I think that adding the acronym for the Extended Reconstructed Sea Surface Temperature Version 5 dataset would be helpful.
L88. Why was a reference period of 1981-2010 chosen, and are the results sensitive to this?
L117. Was the forced signal that you removed the multi-model mean?
L151. Why global warming and not external forcing in general? I would have thought that anthropogenic aerosols and volcanic eruptions might also have an impact.
L261. It would be useful to cite papers which have suggested that forcing (particularly natural) could contribute to the slow down, (see e.g. box 3.1 IPCC AR6 WG1)Citation: https://doi.org/10.5194/egusphere-2022-98-RC1 -
AC1: 'Reply on RC1', Rashed Mahmood, 07 Jul 2022
The comment was uploaded in the form of a supplement: https://egusphere.copernicus.org/preprints/2022/egusphere-2022-98/egusphere-2022-98-AC1-supplement.pdf
-
AC1: 'Reply on RC1', Rashed Mahmood, 07 Jul 2022
-
RC2: 'Comment on egusphere-2022-98', Anonymous Referee #2, 31 May 2022
The paper presents an interesting approach to providing climate predictions based on constraining non-initialised climate projections with observed climate variability. Only those ensemble members of climate projections that show the larges agreement with observed SST anomalies in years prior to the forecast start date are used to construct climate predictions. Instead of a full initialisation with the observed state as normally done in climate predictions (e.g. for seasonal and decadal prediction), a simplified (‘poor-man’) approach of aligning the phase of theses of simulated and observed SST variability is used. After applying the approach to hindcasts from 1961 onwards, the forecast quality of the predictions is evaluated and compared to both fully initialised decadal predictions and unconstraint climate projections. It was concluded that the constrained ensemble provides skilful predictions of near-surface temperature, lea-level pressure and precipitation in large areas of the globe. During the first decade of predictions the skill of the poor-man predictions is comparable to the initialised decadal predictions. Significant added value from the constrained approach was found in the second decade for which initialised predictions are not available. Sensitivities to certain choices like the past period and geographical regions of the constraint, ensemble size and skill metric, have been discussed.
I think the approach explored in this study is very interesting and certainly deserves to be published. In particular, I agree with the authors that the potential benefits of their approach over both initialised decadal predictions and unconstrained projections for providing seamless climate information could be big and important. However, I cannot recommend publication of the paper in its current form because it lacks several critical aspects that are discussed below.
Major comment
In my view the manuscript suffers substantially from the poor demonstration of the results. While the motivation and the methodological approach are nicely laid out, the analysis of the results and their graphical presentation do not provide enough evidence to the reader to be convinced of the benefits that the new prediction system might bring. With “enough evidence” I don’t mean the quantity of analysis or plots but rather the opposite: the authors have taken the approach to include into the manuscript and the supplementary material almost every possible plot one can think of for the quantities they have analysed. However and unfortunately, the large number of plots does not provide an equally large amount of useful information. I would suggest to critically review all plots and only show those which really help support the claims you are trying to make. It is the responsibility of the authors to make a meaningful selection of the diagnostics that help tell the story you wish to convey and should go into a publication. This critical selection should not be left to the reader alone. I have the following specific recommendations:
Fig 1: I think this could be cut short without loss of information by only showing one start date as a demonstrator and carefully describing the methodology in the text and figure caption, as already done.
Fig 2:
- I don’t find showing means over 10 or 20 years are helpful in the prediction context. The window is too long to provide useful information. It would be better to split the windows into smaller ones to identify those time ranges where the approach can improve either decadal predictions of projections. For example, if the added value over non-initialised projections kicks in after 10 years, it would be most interesting to know when this happens – it is just immediately after the 10 years or more towards the end of the 20-year period? Averages over 10 years smear out the impact, and means over 20 years can potentially even be misleading by implying the skill comes from the later years when most likely it is coming from the earlier years. I would suggest looking at 1-5, 6-10, 11-15 and 16-20 years. Or, if the results reveal interesting insight, even for finer forecast ranges. This recommendation applies to almost all plots in the paper and supplementary material.
- Please also show ACC of the unconstrained projections after 10 years to provide a reference to which to compare to. Fig 3 for SLP shows differences which is helpful but Fig 2 for surface temperature does not.
- It would also be interesting to show how a similarly constrained decadal prediction ensemble would perform, that is sub-sampling those ensemble members from DCPP that most closely resemble the past SST observation after e.g. forecast year 1. That would of course imply that the predictions are only useable after applying the constrain (e.g. after 1 year) but for the longer time scales this could still be useful.
- What is the reference forecast used in the RPSS computation? Please add this information in the figure caption.
Since showing too many global maps is not sustainable, I would recommend to condense the critical information either into 2D plots or bar charts (similar to what has already been done in the Supplement but for finer forecast ranges). These could be good options for the various sensitivity studies. For example, global or key regional scores could be plotted in a 2D plot as a function of forecast year and selection period to replace Fig S2 etc. Such a condensation would make space to show a direct comparison (or differences) with the performance of the unconstrained ensemble or the decadal predictions.
I find some of the results are a bit over-interpreted and should be re-worded slightly more carefully. For example, on line 173 you say that added value is found over similar regions across different forecast times providing confidence in the robustness. However, the plots these lines refer to (Fig 2h-j) indicate for example some inconsistencies in the North Atlantic and the tropical East Pacific. Or for SLP in Fig 3, the highly skilful subtropical North Atlantic for FY11-20 (Fig 3b) is not showing during the first 10 years (Fig 3a). Why is this? Around lines 180, mention the problematic issues over the Indian Ocean.
The result that the constrained projections can outperform the initialised predictions is very interesting. I feel it would require some more discussion as to what the mechanisms are that can lead to this perhaps surprising skill. Discussing potential explanations would make the paper much stronger than simply describing it.
Supplementary Information:
It is not clear which variables have been analysed in Fig S1-S3 and S5-S6.
Minor comments:
Fig 4 caption: unclear what exactly is meant by added skill – please specify.
Switching between the ACC and the residual correlations introduces some inconsistencies in the manuscript. For the purpose of this paper, it might be sufficient to only show ACC. Fig 5 and 6 could go in the Supplement.
Why are the atmospheric fields computed on a 5x5 degree grid and not on a 3x3 degree grid as the SSTs?
Section 3.1: The text could be improved by introducing more paragraphs and reducing the use of brackets ().
Sensitivity to temporal averaging of SST anomalies, around lines 206-209: emphasise this finding more as very interesting.
The cited reference of Menary et al (2021) sounds very relevant – could you expand on your discussion of this paper in Section 4?
Sensitivity to ensemble members: Why have you stopped sub-sampling at 50 members? It would be interesting to see the convergence for the full ensemble. Would it be possible to show a plot showing this convergence for perhaps a global quantity? Is there an optimal ensemble size?
Figure 7: show comparison with global pattern (repeat from Fig 1). Also show comparison with unconstrained and decadal prediction. Makes interpretation of these results easier.
Fig 8: Can you speculate as to why the constraint over the North Atlantic makes the forecasts worse? Fig 8b is not needed.
Supplementary Information:
Figures S8 – S10: instead of repeating maps very similar to Fig 2—4, it would perhaps be more informative to show differences to these other maps.
Citation: https://doi.org/10.5194/egusphere-2022-98-RC2 -
AC2: 'Reply on RC2', Rashed Mahmood, 07 Jul 2022
The comment was uploaded in the form of a supplement: https://egusphere.copernicus.org/preprints/2022/egusphere-2022-98/egusphere-2022-98-AC2-supplement.pdf
-
RC3: 'Comment on egusphere-2022-98', Anonymous Referee #3, 01 Jun 2022
General Comments:
This manuscript by Mahmood et al. (egusphere-2022-98) explores an alternate way of decadal to multidecadal predictions by subsampling individual CMIP6 historical simulations that better matches observed SST at a given time (start dates) and by tracking the trajectories of the subsampled simulations for the next two decades (analogue method hereafter). The authors show that the added value of the analogue method over the uninitialized simulations is comparable to that of initialized decadal prediction simulation (DCPP). The manuscript is overall well organized and written. I have enjoyed reading the manuscript. I also acknowledge the thorough effort to test the sensitivity of the results to subsampling criteria. I think the manuscript has potential to draw attention from climate research community, as the proposed method can leverage existing simulations in the application of the prediction of climate, instead of running computationally expensive decadal prediction simulations. However, before I recommend accepting for publication, I have a couple of specific comments that I wish the authors further demonstrate or explain along with some minor suggestions.
Specific Comments:
1) The authors shows that the analogue method exhibits high skill in the Pacific Ocean, even higher than skill in the subpolar North Atlantic, which even lasts for FY11-20 . Such a long memory in the Pacific is surprising and in stark contrast to the current understanding that predictability in the Pacific Ocean is low on decadal timescales while very high in the subpolar North Atlantic. The low predictability in initialized decadal predictions may be related to initialization shock/drift, as the authors also discuss in the manuscript. However, the low predictability in the Pacific Ocean is also pervasive in “perfect model” experiments (e.g., Collins 2002; Pohlmann et al. 2004), which does not suffer from initialization shock/drift. Why the author’s analogue method shows such superior skill in the Pacific Ocean? Isn’t this high skill possibly related to the forced signal that is not completely removed by the method the authors used (Smith et al. 2019)? One way to verify this is to perform a bootstrapping method for the statistical test, rather than Student’s t-test. If the ACC from Best30 is found outside of the (eg., 2.5 to 97.5 percentile) distribution of the ACCs from randomly sampled 30 members, assuming that the uninitialized ensemble mean used in Smith et al.’s method is the total 212-member ensemble mean, the authors can say more confidently that the high Pacific skill is indeed not from the forced signal.
2) If that is the case, why the skill is so high in the Pacific? Since this would the most important finding of the study, in my opinion, as it is in contrast to the current understanding, I recommend that the authors further demonstrate the reasons for the high Pacific skill.
3) The Atlantic skill is low for FY1-10, but picks up for FY11-20 (Fig. 2d-e). Why is this the case? I think the low skill for FY1-10 is because Best30 is dominated by the correlations in the Pacific and as demonstrated in the regional SST constraints, but it is hard to understand why there is an rebound in ACC skill.
4) The authors introduce several statistical methods in section 2, without a description, just by referring to citations. I recommend adding a brief description for each method.
Technical corrections:
l. 43: Remove “in” after phasing.
l. 88: …anomalies “relative to” the reference climatological period…
Citation: https://doi.org/10.5194/egusphere-2022-98-RC3 -
AC3: 'Reply on RC3', Rashed Mahmood, 07 Jul 2022
The comment was uploaded in the form of a supplement: https://egusphere.copernicus.org/preprints/2022/egusphere-2022-98/egusphere-2022-98-AC3-supplement.pdf
-
AC3: 'Reply on RC3', Rashed Mahmood, 07 Jul 2022
Interactive discussion
Status: closed
-
RC1: 'Comment on egusphere-2022-98', Anonymous Referee #1, 11 May 2022
In my opinion this is a very well written and clear paper which presents an interesting new way of combining a large multi-model dataset in order to constrain future projections. As such I have very few comments to make and I believe that this paper should be published with only a few corrections.
The results are impressive and warrant publication. However the one area I would like to see more explanation on is the cause of the skill found. In the summary (line 300) you claim that “these results indicate that there is significant multi-decadal predictability from internal variability”. I am not convinced though that as they stand your results can justify this claim (although I’m not suggesting this is necessarily incorrect). By selecting the best 30 (or similar) simulations from the CMIP6 archive based on the 9 most recent years, could it not be possible that what you are doing is weighting your results to certain models (which is in itself a rich area of literature). This could be those models with the most realistic response to forcing, especially when outside the anomaly period at the start or end of the record. Or it could potentially be those with the most realistic modes of variability. Both of which, I would have thought could theoretically give an increase in skill over the next decades. This is especially pertinent, given that you state that the method gives a better constraint that a single model ensemble (although as you say this is expected given the increased number of simulations to sample from). I think that this point needs to be discussed, and preferably investigated. One simple test you could do, without the need for further analysis, would be to check if you are selecting simulations preferentially from one climate model or if the 30 members are selected from the full model range. In addition it would also be interesting to see if the skill varies through time.
I am also somewhat surprised that you have not shown results for future projections, since that would seems to be the logical direction for this type of analysis. Although, perhaps this is being left for future work (which would be perfectly justifiable).
More minor points:
L 84. I think that adding the acronym for the Extended Reconstructed Sea Surface Temperature Version 5 dataset would be helpful.
L88. Why was a reference period of 1981-2010 chosen, and are the results sensitive to this?
L117. Was the forced signal that you removed the multi-model mean?
L151. Why global warming and not external forcing in general? I would have thought that anthropogenic aerosols and volcanic eruptions might also have an impact.
L261. It would be useful to cite papers which have suggested that forcing (particularly natural) could contribute to the slow down, (see e.g. box 3.1 IPCC AR6 WG1)Citation: https://doi.org/10.5194/egusphere-2022-98-RC1 -
AC1: 'Reply on RC1', Rashed Mahmood, 07 Jul 2022
The comment was uploaded in the form of a supplement: https://egusphere.copernicus.org/preprints/2022/egusphere-2022-98/egusphere-2022-98-AC1-supplement.pdf
-
AC1: 'Reply on RC1', Rashed Mahmood, 07 Jul 2022
-
RC2: 'Comment on egusphere-2022-98', Anonymous Referee #2, 31 May 2022
The paper presents an interesting approach to providing climate predictions based on constraining non-initialised climate projections with observed climate variability. Only those ensemble members of climate projections that show the larges agreement with observed SST anomalies in years prior to the forecast start date are used to construct climate predictions. Instead of a full initialisation with the observed state as normally done in climate predictions (e.g. for seasonal and decadal prediction), a simplified (‘poor-man’) approach of aligning the phase of theses of simulated and observed SST variability is used. After applying the approach to hindcasts from 1961 onwards, the forecast quality of the predictions is evaluated and compared to both fully initialised decadal predictions and unconstraint climate projections. It was concluded that the constrained ensemble provides skilful predictions of near-surface temperature, lea-level pressure and precipitation in large areas of the globe. During the first decade of predictions the skill of the poor-man predictions is comparable to the initialised decadal predictions. Significant added value from the constrained approach was found in the second decade for which initialised predictions are not available. Sensitivities to certain choices like the past period and geographical regions of the constraint, ensemble size and skill metric, have been discussed.
I think the approach explored in this study is very interesting and certainly deserves to be published. In particular, I agree with the authors that the potential benefits of their approach over both initialised decadal predictions and unconstrained projections for providing seamless climate information could be big and important. However, I cannot recommend publication of the paper in its current form because it lacks several critical aspects that are discussed below.
Major comment
In my view the manuscript suffers substantially from the poor demonstration of the results. While the motivation and the methodological approach are nicely laid out, the analysis of the results and their graphical presentation do not provide enough evidence to the reader to be convinced of the benefits that the new prediction system might bring. With “enough evidence” I don’t mean the quantity of analysis or plots but rather the opposite: the authors have taken the approach to include into the manuscript and the supplementary material almost every possible plot one can think of for the quantities they have analysed. However and unfortunately, the large number of plots does not provide an equally large amount of useful information. I would suggest to critically review all plots and only show those which really help support the claims you are trying to make. It is the responsibility of the authors to make a meaningful selection of the diagnostics that help tell the story you wish to convey and should go into a publication. This critical selection should not be left to the reader alone. I have the following specific recommendations:
Fig 1: I think this could be cut short without loss of information by only showing one start date as a demonstrator and carefully describing the methodology in the text and figure caption, as already done.
Fig 2:
- I don’t find showing means over 10 or 20 years are helpful in the prediction context. The window is too long to provide useful information. It would be better to split the windows into smaller ones to identify those time ranges where the approach can improve either decadal predictions of projections. For example, if the added value over non-initialised projections kicks in after 10 years, it would be most interesting to know when this happens – it is just immediately after the 10 years or more towards the end of the 20-year period? Averages over 10 years smear out the impact, and means over 20 years can potentially even be misleading by implying the skill comes from the later years when most likely it is coming from the earlier years. I would suggest looking at 1-5, 6-10, 11-15 and 16-20 years. Or, if the results reveal interesting insight, even for finer forecast ranges. This recommendation applies to almost all plots in the paper and supplementary material.
- Please also show ACC of the unconstrained projections after 10 years to provide a reference to which to compare to. Fig 3 for SLP shows differences which is helpful but Fig 2 for surface temperature does not.
- It would also be interesting to show how a similarly constrained decadal prediction ensemble would perform, that is sub-sampling those ensemble members from DCPP that most closely resemble the past SST observation after e.g. forecast year 1. That would of course imply that the predictions are only useable after applying the constrain (e.g. after 1 year) but for the longer time scales this could still be useful.
- What is the reference forecast used in the RPSS computation? Please add this information in the figure caption.
Since showing too many global maps is not sustainable, I would recommend to condense the critical information either into 2D plots or bar charts (similar to what has already been done in the Supplement but for finer forecast ranges). These could be good options for the various sensitivity studies. For example, global or key regional scores could be plotted in a 2D plot as a function of forecast year and selection period to replace Fig S2 etc. Such a condensation would make space to show a direct comparison (or differences) with the performance of the unconstrained ensemble or the decadal predictions.
I find some of the results are a bit over-interpreted and should be re-worded slightly more carefully. For example, on line 173 you say that added value is found over similar regions across different forecast times providing confidence in the robustness. However, the plots these lines refer to (Fig 2h-j) indicate for example some inconsistencies in the North Atlantic and the tropical East Pacific. Or for SLP in Fig 3, the highly skilful subtropical North Atlantic for FY11-20 (Fig 3b) is not showing during the first 10 years (Fig 3a). Why is this? Around lines 180, mention the problematic issues over the Indian Ocean.
The result that the constrained projections can outperform the initialised predictions is very interesting. I feel it would require some more discussion as to what the mechanisms are that can lead to this perhaps surprising skill. Discussing potential explanations would make the paper much stronger than simply describing it.
Supplementary Information:
It is not clear which variables have been analysed in Fig S1-S3 and S5-S6.
Minor comments:
Fig 4 caption: unclear what exactly is meant by added skill – please specify.
Switching between the ACC and the residual correlations introduces some inconsistencies in the manuscript. For the purpose of this paper, it might be sufficient to only show ACC. Fig 5 and 6 could go in the Supplement.
Why are the atmospheric fields computed on a 5x5 degree grid and not on a 3x3 degree grid as the SSTs?
Section 3.1: The text could be improved by introducing more paragraphs and reducing the use of brackets ().
Sensitivity to temporal averaging of SST anomalies, around lines 206-209: emphasise this finding more as very interesting.
The cited reference of Menary et al (2021) sounds very relevant – could you expand on your discussion of this paper in Section 4?
Sensitivity to ensemble members: Why have you stopped sub-sampling at 50 members? It would be interesting to see the convergence for the full ensemble. Would it be possible to show a plot showing this convergence for perhaps a global quantity? Is there an optimal ensemble size?
Figure 7: show comparison with global pattern (repeat from Fig 1). Also show comparison with unconstrained and decadal prediction. Makes interpretation of these results easier.
Fig 8: Can you speculate as to why the constraint over the North Atlantic makes the forecasts worse? Fig 8b is not needed.
Supplementary Information:
Figures S8 – S10: instead of repeating maps very similar to Fig 2—4, it would perhaps be more informative to show differences to these other maps.
Citation: https://doi.org/10.5194/egusphere-2022-98-RC2 -
AC2: 'Reply on RC2', Rashed Mahmood, 07 Jul 2022
The comment was uploaded in the form of a supplement: https://egusphere.copernicus.org/preprints/2022/egusphere-2022-98/egusphere-2022-98-AC2-supplement.pdf
-
RC3: 'Comment on egusphere-2022-98', Anonymous Referee #3, 01 Jun 2022
General Comments:
This manuscript by Mahmood et al. (egusphere-2022-98) explores an alternate way of decadal to multidecadal predictions by subsampling individual CMIP6 historical simulations that better matches observed SST at a given time (start dates) and by tracking the trajectories of the subsampled simulations for the next two decades (analogue method hereafter). The authors show that the added value of the analogue method over the uninitialized simulations is comparable to that of initialized decadal prediction simulation (DCPP). The manuscript is overall well organized and written. I have enjoyed reading the manuscript. I also acknowledge the thorough effort to test the sensitivity of the results to subsampling criteria. I think the manuscript has potential to draw attention from climate research community, as the proposed method can leverage existing simulations in the application of the prediction of climate, instead of running computationally expensive decadal prediction simulations. However, before I recommend accepting for publication, I have a couple of specific comments that I wish the authors further demonstrate or explain along with some minor suggestions.
Specific Comments:
1) The authors shows that the analogue method exhibits high skill in the Pacific Ocean, even higher than skill in the subpolar North Atlantic, which even lasts for FY11-20 . Such a long memory in the Pacific is surprising and in stark contrast to the current understanding that predictability in the Pacific Ocean is low on decadal timescales while very high in the subpolar North Atlantic. The low predictability in initialized decadal predictions may be related to initialization shock/drift, as the authors also discuss in the manuscript. However, the low predictability in the Pacific Ocean is also pervasive in “perfect model” experiments (e.g., Collins 2002; Pohlmann et al. 2004), which does not suffer from initialization shock/drift. Why the author’s analogue method shows such superior skill in the Pacific Ocean? Isn’t this high skill possibly related to the forced signal that is not completely removed by the method the authors used (Smith et al. 2019)? One way to verify this is to perform a bootstrapping method for the statistical test, rather than Student’s t-test. If the ACC from Best30 is found outside of the (eg., 2.5 to 97.5 percentile) distribution of the ACCs from randomly sampled 30 members, assuming that the uninitialized ensemble mean used in Smith et al.’s method is the total 212-member ensemble mean, the authors can say more confidently that the high Pacific skill is indeed not from the forced signal.
2) If that is the case, why the skill is so high in the Pacific? Since this would the most important finding of the study, in my opinion, as it is in contrast to the current understanding, I recommend that the authors further demonstrate the reasons for the high Pacific skill.
3) The Atlantic skill is low for FY1-10, but picks up for FY11-20 (Fig. 2d-e). Why is this the case? I think the low skill for FY1-10 is because Best30 is dominated by the correlations in the Pacific and as demonstrated in the regional SST constraints, but it is hard to understand why there is an rebound in ACC skill.
4) The authors introduce several statistical methods in section 2, without a description, just by referring to citations. I recommend adding a brief description for each method.
Technical corrections:
l. 43: Remove “in” after phasing.
l. 88: …anomalies “relative to” the reference climatological period…
Citation: https://doi.org/10.5194/egusphere-2022-98-RC3 -
AC3: 'Reply on RC3', Rashed Mahmood, 07 Jul 2022
The comment was uploaded in the form of a supplement: https://egusphere.copernicus.org/preprints/2022/egusphere-2022-98/egusphere-2022-98-AC3-supplement.pdf
-
AC3: 'Reply on RC3', Rashed Mahmood, 07 Jul 2022
Peer review completion
Journal article(s) based on this preprint
Viewed
HTML | XML | Total | Supplement | BibTeX | EndNote | |
---|---|---|---|---|---|---|
362 | 130 | 7 | 499 | 40 | 6 | 3 |
- HTML: 362
- PDF: 130
- XML: 7
- Total: 499
- Supplement: 40
- BibTeX: 6
- EndNote: 3
Viewed (geographical distribution)
Country | # | Views | % |
---|
Total: | 0 |
HTML: | 0 |
PDF: | 0 |
XML: | 0 |
- 1
Markus G. Donat
Pablo Ortega
Francisco J. Doblas-Reyes
Carlos Delgado-Torres
Margarida Samsó
Pierre-Antoine Bretonnière
The requested preprint has a corresponding peer-reviewed final revised paper. You are encouraged to refer to the final revised version.
- Preprint
(1917 KB) - Metadata XML
-
Supplement
(2133 KB) - BibTeX
- EndNote
- Final revised paper