the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
Disappearing day-of-week ozone patterns in US nonattainment areas
Abstract. Past work has shown that traffic patterns in the US and resulting NOX emissions vary by day of week, with NOX emissions typically higher on weekdays than weekends. This pattern of emissions leads to different levels of ozone on weekends versus weekdays and can be leveraged to understand how local ozone formation changes in response to NOX emissions perturbations in different urban areas. Specifically, areas with lower NOX but higher ozone on the weekends (the weekend effect) can be characterized as NOX -saturated and areas with both lower NOX and ozone on weekends (the weekday effect) can be characterized as NOX-limited. In this analysis we assess ozone weekend-weekday differences across US nonattainment areas using 18 years of observed and modeled data from 2002–2019 using two metrics: mean ozone and percentage of days > 70 ppb. In addition, we quantify the modeled and observed trends in these weekend-weekday differences across this period of substantial NOX emissions reductions in the US. The model assessment is carried out using EPA’s Air QUAlity TimE Series Project (EQUATES) CMAQ dataset. We identify 3 types of ozone trends occuring across the US: disappearing weekend effect, disappearing weekday effect, and no trend. The disappearing weekend effect occurs in a subset of large urban areas that were NOX -saturated (i.e., VOC-limited) at the beginning of the analysis period but transitioned to mixed chemical regimes or NOX-limited conditions by the end of the analysis period. Nine areas have disappearing weekend effect trends in both datasets and with both metrics indicating strong agreement that they are shifting to more NOX-limited conditions: Milwaukee, Houston, Phoenix, Denver, Northern Wasatch Front, Southern Wasatch Front, Las Vegas, Los Angeles – San Bernardino County, Los Angeles – South Coast, and San Diego. The disappearing weekday effect was identified for multiple rural and agricultural areas of California which were NOX -limited for the entire analysis period but appear to become less influenced by local day of week emission patterns in more recent years. Finally, we discuss a variety of reasons why there are no statistically significant trends in certain areas including complex impacts of heterogeneous source mixes and stochastic impacts of meteorology. Overall, this assessment finds that the EQUATES modeling simulations indicate more NOX-saturated conditions than the observations but do a good job of capturing year-to-year changes in weekend-weekday ozone patterns.
-
Notice on discussion status
The requested preprint has a corresponding peer-reviewed final revised paper. You are encouraged to refer to the final revised version.
-
Preprint
(2101 KB)
-
Supplement
(2397 KB)
-
The requested preprint has a corresponding peer-reviewed final revised paper. You are encouraged to refer to the final revised version.
- Preprint
(2101 KB) - Metadata XML
-
Supplement
(2397 KB) - BibTeX
- EndNote
- Final revised paper
Journal article(s) based on this preprint
Interactive discussion
Status: closed
-
RC1: 'Comment on egusphere-2023-1974', Anonymous Referee #1, 25 Sep 2023
This is a very good study, written by experts at EPA and it will be a welcome addition to the literature. I find the research question to be important and the conclusions are clearly supported by the analysis. I don’t have any concerns regarding the analysis or conclusions, my only recommendation is that the authors update their terminology regarding the reporting of statistical findings, to be more consistent with the Tropospheric Ozone Assessment Report and current thinking regarding the limitations of the expression “statistically significant”, as described below.
Regarding the use of the Theil-Sen/Mann-Kendall method for calculating trends, the authors state that they chose this method because of the small sample sizes and because it does not require assumptions about the distribution of the residuals. Another reason that is often given for the choice of this method is that it is resistant to outliers. The problem is that in order to remove the impact from outliers, this method automatically ignores up to 29% of the data points in a sample (see Section 2 of Chang et al., 2021). This would be fine if the analyst believed that the outliers are due to instrument errors, but there is no reason to throw out data if they are believed to be reliable. In your case there is no reason to believe that your samples contain erroneous data points that should be ignored. For this reason, the Tropospheric Ozone Assessment Report has abandoned the Theil-Sen/Mann-Kendall method that was used in the first phase of TOAR (2014-2019). A further problem with the Theil-Sen method is that it produces unrealistically narrow 95% confidence intervals. This is shown in Figure 1 of the TOAR-II Recommendations for Statistical Analyses (available at https://igacproject.org/activities/TOAR/TOAR-II). Figure 1 compares the trend and 95% confidence interval calculated by 10 different methods for the ozone time series at Mace Head, Ireland. The Theil-Sen method has the narrowest 95% confidence interval by far, and the reason is that this method ignores 29% of the data; by throwing out all of the extreme values the sample has very little variability and therefore a straight line can be fit through the remaining data within a very narrow range. The second phase of TOAR-II is now recommending the use of quantile regression, as described in the TOAR-II Recommendations for Statistical Analyses. Quantile regression was used to good effect in the very nice paper by co-author B. Wells (Wells et al., 2021), and it could easily be applied to your current analysis.
Throughout the paper the authors use the expression “statistically significant”, however this expression is now recognized as being problematic and it should be abandoned and replaced by the more useful method of reporting all trends (with uncertainty) and all p-values, followed by a discussion of the trends and the author’s opinion regarding their confidence in the trend values. This advice comes from a highly influential paper by Wasserstein et al. (2019), published in the journal, The American Statistician, that has already been cited over 1300 times (according to Web of Science). This advice was adopted by the first phase of TOAR (Tarasick et al., 2019) and will also be used by TOAR-II. Some other recent papers on ozone trends that have taken this advice are: Chang et al., 2020; Cooper et al., 2020; Gaudel et al., 2020; Chang et al., 2022; Wang et al., 2022; Mousavinezhad et al., 2023. Because these papers report all trend values, uncertainties, and all p-values, and also discuss the trend results, there is no confusion regarding the findings, and one does not even notice that the term “statistically significant” is not used at all.
The authors describe a trend as “no trend” when the p-value is greater than 0.05. There are two problems with this approach:
1) as described above the expression “statistically significant” which is tied to the p-value of 0.05 should be abandoned. Just because a trend has a p-value of 0.06, it does not mean that there is absolutely no trend, it just means that there is a gray area and the trend is not as robust as one that has a p-value of 0.02. Chang et al. (2017) provide a nice demonstration of the useful information that can be gleaned from a trend with a p-value greater than 0.05 (see their Figure 13). They calculated a regional ozone trend for the eastern USA using all available ozone monitors (in summer the trend was strongly negative for the period 2000-2014). They then conducted an exercise to see what would happen to the regional trend if they threw out all time series with a p-value less than 0.05. The result was almost the same because the time series with p-values greater than 0.05 still reflected the overall regional decrease of ozone.
2) The authors are using the Theil-Sen method to calculate trends and p-values. As described above the 95% confidence intervals are unrealistically narrow using this method, and therefore the p-values are also too low. This means that too many sites are classified as having a real trend, according to the 0.05 p-value threshold. If the authors use another method for calculating trends (like quantile regression) the p-values will increase and they would then have to classify more sites as having “no trend”. Given the gray area around p-values, and given that trends with p-values greater than 0.05 can still be reliable, there is no justification for dichotomizing ozone time series as “trend” or “no trend” based on a p-value.
When I look at the maps in Figure 4 and 5 I am left wondering about the non-attainment regions labeled as “no trend”. Is there really no trend here, i.e. a flat line, or is there still a decrease, but it just doesn’t reach the arbitrary threshold of p<0.05? A good example is Tuscan Buttes. Table S-1 shows the observed and modelled trend is the same (0.14) but because the model has a p-value of 0.02 this trend is considered to be real, while the observations have a p-value of 0.06 and are classified as “no trend”. The TOAR papers report all trends and all p-values and the trend values in their map plots are colored according to p-value (Fleming et al., 2018). This allow the reader to see if a trend is still notable (e.g. a p-value between 0.05 and 0.10) or if there really and truly is no trend (e.g. a p-value > 0.33). It would be very helpful to the reader if the authors can color their maps according to p-value, in a manner similar to TOAR.References:
Chang, K-L, et al 2017 Regional trend analysis of surface ozone observations from monitoring networks in eastern North America, Europe and East Asia. Elem Sci Anth, 5: 50, DOI: https://doi.org/10.1525/elementa.243
Chang, K.-L., et al. (2020), Statistical regularization for trend detection: An integrated approach for detecting long-term trends from sparse tropospheric ozone profiles, Atmos. Chem. Phys., 20, 9915–9938, https://doi.org/10.5194/acp-20-9915-2020
Chang, K-L, et al. 2021. Trend detection of atmospheric time series: Incorporating appropriate uncertainty estimates and handling extreme events. Elem Sci Anth, 9: 1. DOI: https://doi.org/10.1525/elementa.2021.00035
Chang, K.-L., (2022), Impact of the COVID-19 economic downturn on tropospheric ozone trends: an uncertainty weighted data synthesis for quantifying regional anomalies above western North America and Europe, AGU Advances, 3, e2021AV000542. https://doi.org/10.1029/2021AV000542
Cooper, et al. 2020. Multi-decadal surface ozone trends at globally distributed remote locations. Elem Sci Anth, 8: 23. DOI: https://doi.org/10.1525/elementa.420
Fleming, Z. L., R. M. Doherty, E. von Schneidemesser, C. S. Malley, O. R. Cooper et al. (2018), Tropospheric Ozone Assessment Report: Present-day ozone distribution and trends relevant to human health, Elem Sci Anth, 6(1):12, DOI: https://doi.org/10.1525/elementa.273
Gaudel, A., et al. (2020), Aircraft observations since the 1990s reveal increases of tropospheric ozone at multiple locations across the Northern Hemisphere. Sci. Adv. 6, eaba8272, DOI: 10.1126/sciadv.aba8272
Mousavinezhad, S., Ghahremanloo, M., Choi, Y., Pouyaei, A., Khorshidian, N. and Sadeghi, B., 2023. Surface ozone trends and related mortality across the climate regions of the contiguous United States during the most recent climate period, 1991–2020. Atmospheric Environment, p.119693.
Tarasick, D. W. et al. (2019), Tropospheric Ozone Assessment Report: Tropospheric ozone from 1877 to 2016, observed levels, trends and uncertainties. Elem Sci Anth, 7(1), DOI: http://doi.org/10.1525/elementa.376
Wang, H., et al. (2022), Global tropospheric ozone trends, attributions, and radiative impacts in 1995–2017: an integrated analysis using aircraft (IAGOS) observations, ozonesonde, and multi-decadal chemical model simulations, Atmos. Chem. Phys., 22, 13753–13782, https://doi.org/10.5194/acp-22-13753-2022
Wasserstein, R. L., Schirm, A. L., and Lazar, N. A.: Moving to a world beyond p < 0:05, Am. Stat., 73, 1–29,
https://doi.org/10.1080/00031305.2019.1583913, 2019.Citation: https://doi.org/10.5194/egusphere-2023-1974-RC1 -
AC2: 'Reply on RC1', Heather Simon, 01 Dec 2023
The comment was uploaded in the form of a supplement: https://egusphere.copernicus.org/preprints/2023/egusphere-2023-1974/egusphere-2023-1974-AC2-supplement.pdf
-
AC2: 'Reply on RC1', Heather Simon, 01 Dec 2023
-
CC1: 'Comment on egusphere-2023-1974', David Parrish, 15 Oct 2023
Comment on: “Disappearing day-of-week ozone patterns in US nonattainment areas” by H. Simon et al.
David D. Parrish1 and Richard G. Derwent2
1 David.D.Parrish, LLC, 4630 MacArthur Ln, Boulder, Colorado, USA
2 rdscientific, Newbury, Berkshire, UK
The submitted manuscript undertakes an interesting analysis – an examination of day-of-week patterns in ozone concentrations across many US regions. The goal is to illuminate possible changes in regional photochemical environments due to transitions between NOX versus VOC sensitive ozone formation. The authors present a great deal of analysis of both observational data and model results, and reach important conclusions, e.g., “Nine (large US urban) areas have disappearing weekend effect … indicating .. that they are shifting to more NOX-limited conditions: ….” However, there are shortcomings in the authors’ analysis; below we discuss two of these that compromise the robustness of the author’s conclusions if they cannot be adequately addressed.
First, a significant fractional “disappearance” in observed day-of-week ozone patterns must be expected simply due to the decrease in anthropogenic ozone formation driven by effective precursor emission controls implemented over the past decades. Parrish et al. (2016) show that between 1980 and 2015 in the Los Angeles urban area (i.e., California’s South Coast Air Basin) the temporal dependence of the anthropogenic ozone contribution to the observed distribution of MDA8 ozone concentrations is well-described by an exponential decay with an e-folding time of 22 years. Decreases in anthropogenic ozone formation of similar magnitudes have been documented in other US regions (Parrish et al., 2022, and references therein). In the submitted manuscript, the authors consider 18 years (2002 to 2019) during which this decrease would have amounted to a factor of 2.3. Since the authors use an absolute measure of the day-of-week ozone pattern (i.e., mean difference in ozone in ppb between weekends and weekdays, their Equation 1), that measure would be expected to decrease by that same factor, or by 56% over their selected analysis period, even in the absence of any change in photochemical environment. Unless the authors can demonstrate that the disappearance of day-of-week ozone pattern is significantly greater (or lesser) than 56% of the anthropogenic contribution, those disappearances cannot be taken as evidence for a change in the photochemical environment of the ambient atmosphere.
Second, the details of the trend analysis require further discussion. The authors analyze 18 years (2002 to 2019) of observed MDA8 ozone concentrations. However, rather than analyze the 18 individual years, the authors choose to base their analysis on 5-year rolling periods (i.e., 14 different periods covering the 18-year time series). Importantly, there is a great deal of autocorrelation between the 14 different 5-year rolling means. In fact, the 18 years of observations gives fewer than 4 (i.e., 3.6) independent 5-year means. Attempts to derive trends with reliable confidence limits from such a limited number of independent data is quite uncertain. It must be fully appreciated that the use of 5-year rolling means does not improve the confidence limits of derived trends over those of trends derived from the individual years. In particular, use of the Mann-Kendall test to determine the statistical significance of the derived trends in WE-WD O3 differences will give overly optimistic results if the 14 different 5-year rolling periods are considered to be independent data. The rolling means can give plots that apparently illustrate well-defined trends (e.g., upper right graphs in Figures 1-3), but these illustrations are misleading if the autocorrelation of the 5-year means are not fully considered. A complete discussion of the trend analysis and derivation of confidence limits based upon less than 4 independent observational data points is required; if the autocorrelation of the 5-year rolling means has not been adequately considered, then revisions are required.
References
Parrish, D.D., Xu, J., Croes, B., and Shao, M.: Air Quality Improvement in Los Angeles - Perspectives for Developing Cities, Frontiers of Environ Sci. & Engineering, 2016, 10(5): 11 DOI 10.1007/s11783-016-0859-5, 2016.
Parrish, D.D., Faloona, I.C., and Derwent R.G.: Observational-based Assessment of Contributions to Maximum Ozone Concentrations in the western United States, Journal of the Air & Waste Management Association, 72:5, 434–454, https://doi.org/10.1080/10962247.2022.2050962, 2022.
Citation: https://doi.org/10.5194/egusphere-2023-1974-CC1 -
AC3: 'Reply on CC1', Heather Simon, 01 Dec 2023
The comment was uploaded in the form of a supplement: https://egusphere.copernicus.org/preprints/2023/egusphere-2023-1974/egusphere-2023-1974-AC3-supplement.pdf
-
AC3: 'Reply on CC1', Heather Simon, 01 Dec 2023
-
RC2: 'Comment on egusphere-2023-1974', D.A.J. Jaffe, 17 Oct 2023
Review of “Disappearing day-of-week ozone patterns in US nonattainment areas” by Simon et al.
In this analysis the authors evaluate day of week patterns in average O3 and O3 exceedance days for the 2002-2019 timeframe at ~50 different high O3 sites in the U.S. The authors identify several different patterns including disappearing weekend-weekday differences and others. The analysis mixes observations and model results, which I find to be somewhat problematic. In addition I have some statistical concerns that would need to be corrected or clarified before this could be published.
Major points:
- The authors mix observed patterns with model patterns in a way that I believe is misleading. I think its incumbent on the authors to first clearly document what the observations show. Then we can ask how well the model reproduces the observations and what we can learn from the model where it is consistent with the observations, or if not consistent, then why. Certainly there are plenty of NOx observations that could have been used for this work (see Jaffe et al 2020), so I am not sure what is gained by showing and using only the modeled NOx. But NOx is not that big of a concern. Its formaldehyde that I find much more problematic. For formaldehyde, we have much poorer understanding of emissions and chemistry, both of which are essential to understanding the concentrations. Without any evaluation of the modeled formaldehyde, these results should be removed. In other places the authors quote both modeled trends and observed trends and appear to put equal weight on these. That is incorrect, in my opinion, for the reasons stated above.
- I believe the authors may over-state some of the statistical significance due to auto-correlation. This could be true for both the t-tests for individual years and the trends, which use 5-year running means.
- Finally I note that this has a lot of overlap with our earlier analysis (Jaffe et al 2020). We used data for 1995-2020. This analysis uses data for 2002-2019. There are some modest differences, but overall the results are quite similar. I think its essential that the authors clearly describe what is new and/or whether these results are consistent with the earlier analysis. One area that is different is use of probability of exceedance vs mean concentration. The authors seem to want to discount any differences as being due to random variability, but I am not sure that is true. One focuses on the highest days and the other approach focuses on all days in the O3 Do these days have the same VOC-NOx sensitivity?
Abstract: It is important in abstract to describe the scope: All US O3, all US urban areas or all US non-attainment areas. What years? How many regions considered? In addition, I am unclear what it means if you have a “disappearing weekday” effect. The information here is contained in the relative O3 and NOx behavior between weekday and weekend. So the terms “disappearing weekday” is confusing.
Line 25-26: “both datasets” ?
Line 27: The abstract uses area names that are consistent (I think) with EPA designations, but are often rather non-intuitive. For example, Los Angeles – San Bernardino County vs Los Angeles – South Coast. The San Bernardino monitors are in Riverside CBSA, so aren’t these two locations essentially same region. It would be more interesting to include a site closer to downtown LA like Azusa, where we might expect a different pattern.
Line 33: It is not clear what model evaluation for this work. As near as I can tell, nothing was shown about the models ability to capture year-year variations. The model does seem to capture the trend in weekend-weekday differences.
132-135: While I understand why you excluded 3 out of 7 days, does this change the results?
144: Not clear how t-tests were done. I think you took every weekday and weekend day in one year and compared the means and treated each day as an independent observation. If this is right, then I don’t think autocorrelation was taken into account. In any case please clarify how the t-tests were performed.
169: Given the 5 year running means, these values will have sig autocorrelation. Was this taken into account in the results?
173-174: Unclear meaning.
193-194: Unclear meaning.
Figure 1: Please clarify meaning of P values in top right plot. I think these are for each individual year, correct? Given that the NOx and CH2O plots are for all years, not sure what is the value in showing these. There are major differences (for NOx) between the early and later part of the data record.
234: As noted above, the terminology “disappearing weekend effect” is very misleading. Its really about the difference between weekend and weekday values.
258-259: So how do we interpret these model obs differences? You may say the random variations impact the obs more than the model, but aren’t these variations important?
283: I don’t think the probability approach is inherently noisier, especially when averaged over several years as you have done. I think this is an interesting spot to do a deeper dive.
Conclusions: As noted above it would be good to understand what is new here. Please add some discussion to clarify, perhaps focusing on the differences between the prob of an exceedance approach and mean O3 approach.
Finally, I note that the regression information in the right plots of figures 1-9 (not 5) is almost impossible to read.
Citation: https://doi.org/10.5194/egusphere-2023-1974-RC2 -
RC3: 'Reply on RC2', D.A.J. Jaffe, 18 Oct 2023
Sorry. The earlier reference is jaffe et al 2022, (not 2020).
https://doi.org/10.1029/2021JD036385
Citation: https://doi.org/10.5194/egusphere-2023-1974-RC3 -
AC1: 'Reply on RC2', Heather Simon, 01 Dec 2023
The comment was uploaded in the form of a supplement: https://egusphere.copernicus.org/preprints/2023/egusphere-2023-1974/egusphere-2023-1974-AC1-supplement.pdf
Interactive discussion
Status: closed
-
RC1: 'Comment on egusphere-2023-1974', Anonymous Referee #1, 25 Sep 2023
This is a very good study, written by experts at EPA and it will be a welcome addition to the literature. I find the research question to be important and the conclusions are clearly supported by the analysis. I don’t have any concerns regarding the analysis or conclusions, my only recommendation is that the authors update their terminology regarding the reporting of statistical findings, to be more consistent with the Tropospheric Ozone Assessment Report and current thinking regarding the limitations of the expression “statistically significant”, as described below.
Regarding the use of the Theil-Sen/Mann-Kendall method for calculating trends, the authors state that they chose this method because of the small sample sizes and because it does not require assumptions about the distribution of the residuals. Another reason that is often given for the choice of this method is that it is resistant to outliers. The problem is that in order to remove the impact from outliers, this method automatically ignores up to 29% of the data points in a sample (see Section 2 of Chang et al., 2021). This would be fine if the analyst believed that the outliers are due to instrument errors, but there is no reason to throw out data if they are believed to be reliable. In your case there is no reason to believe that your samples contain erroneous data points that should be ignored. For this reason, the Tropospheric Ozone Assessment Report has abandoned the Theil-Sen/Mann-Kendall method that was used in the first phase of TOAR (2014-2019). A further problem with the Theil-Sen method is that it produces unrealistically narrow 95% confidence intervals. This is shown in Figure 1 of the TOAR-II Recommendations for Statistical Analyses (available at https://igacproject.org/activities/TOAR/TOAR-II). Figure 1 compares the trend and 95% confidence interval calculated by 10 different methods for the ozone time series at Mace Head, Ireland. The Theil-Sen method has the narrowest 95% confidence interval by far, and the reason is that this method ignores 29% of the data; by throwing out all of the extreme values the sample has very little variability and therefore a straight line can be fit through the remaining data within a very narrow range. The second phase of TOAR-II is now recommending the use of quantile regression, as described in the TOAR-II Recommendations for Statistical Analyses. Quantile regression was used to good effect in the very nice paper by co-author B. Wells (Wells et al., 2021), and it could easily be applied to your current analysis.
Throughout the paper the authors use the expression “statistically significant”, however this expression is now recognized as being problematic and it should be abandoned and replaced by the more useful method of reporting all trends (with uncertainty) and all p-values, followed by a discussion of the trends and the author’s opinion regarding their confidence in the trend values. This advice comes from a highly influential paper by Wasserstein et al. (2019), published in the journal, The American Statistician, that has already been cited over 1300 times (according to Web of Science). This advice was adopted by the first phase of TOAR (Tarasick et al., 2019) and will also be used by TOAR-II. Some other recent papers on ozone trends that have taken this advice are: Chang et al., 2020; Cooper et al., 2020; Gaudel et al., 2020; Chang et al., 2022; Wang et al., 2022; Mousavinezhad et al., 2023. Because these papers report all trend values, uncertainties, and all p-values, and also discuss the trend results, there is no confusion regarding the findings, and one does not even notice that the term “statistically significant” is not used at all.
The authors describe a trend as “no trend” when the p-value is greater than 0.05. There are two problems with this approach:
1) as described above the expression “statistically significant” which is tied to the p-value of 0.05 should be abandoned. Just because a trend has a p-value of 0.06, it does not mean that there is absolutely no trend, it just means that there is a gray area and the trend is not as robust as one that has a p-value of 0.02. Chang et al. (2017) provide a nice demonstration of the useful information that can be gleaned from a trend with a p-value greater than 0.05 (see their Figure 13). They calculated a regional ozone trend for the eastern USA using all available ozone monitors (in summer the trend was strongly negative for the period 2000-2014). They then conducted an exercise to see what would happen to the regional trend if they threw out all time series with a p-value less than 0.05. The result was almost the same because the time series with p-values greater than 0.05 still reflected the overall regional decrease of ozone.
2) The authors are using the Theil-Sen method to calculate trends and p-values. As described above the 95% confidence intervals are unrealistically narrow using this method, and therefore the p-values are also too low. This means that too many sites are classified as having a real trend, according to the 0.05 p-value threshold. If the authors use another method for calculating trends (like quantile regression) the p-values will increase and they would then have to classify more sites as having “no trend”. Given the gray area around p-values, and given that trends with p-values greater than 0.05 can still be reliable, there is no justification for dichotomizing ozone time series as “trend” or “no trend” based on a p-value.
When I look at the maps in Figure 4 and 5 I am left wondering about the non-attainment regions labeled as “no trend”. Is there really no trend here, i.e. a flat line, or is there still a decrease, but it just doesn’t reach the arbitrary threshold of p<0.05? A good example is Tuscan Buttes. Table S-1 shows the observed and modelled trend is the same (0.14) but because the model has a p-value of 0.02 this trend is considered to be real, while the observations have a p-value of 0.06 and are classified as “no trend”. The TOAR papers report all trends and all p-values and the trend values in their map plots are colored according to p-value (Fleming et al., 2018). This allow the reader to see if a trend is still notable (e.g. a p-value between 0.05 and 0.10) or if there really and truly is no trend (e.g. a p-value > 0.33). It would be very helpful to the reader if the authors can color their maps according to p-value, in a manner similar to TOAR.References:
Chang, K-L, et al 2017 Regional trend analysis of surface ozone observations from monitoring networks in eastern North America, Europe and East Asia. Elem Sci Anth, 5: 50, DOI: https://doi.org/10.1525/elementa.243
Chang, K.-L., et al. (2020), Statistical regularization for trend detection: An integrated approach for detecting long-term trends from sparse tropospheric ozone profiles, Atmos. Chem. Phys., 20, 9915–9938, https://doi.org/10.5194/acp-20-9915-2020
Chang, K-L, et al. 2021. Trend detection of atmospheric time series: Incorporating appropriate uncertainty estimates and handling extreme events. Elem Sci Anth, 9: 1. DOI: https://doi.org/10.1525/elementa.2021.00035
Chang, K.-L., (2022), Impact of the COVID-19 economic downturn on tropospheric ozone trends: an uncertainty weighted data synthesis for quantifying regional anomalies above western North America and Europe, AGU Advances, 3, e2021AV000542. https://doi.org/10.1029/2021AV000542
Cooper, et al. 2020. Multi-decadal surface ozone trends at globally distributed remote locations. Elem Sci Anth, 8: 23. DOI: https://doi.org/10.1525/elementa.420
Fleming, Z. L., R. M. Doherty, E. von Schneidemesser, C. S. Malley, O. R. Cooper et al. (2018), Tropospheric Ozone Assessment Report: Present-day ozone distribution and trends relevant to human health, Elem Sci Anth, 6(1):12, DOI: https://doi.org/10.1525/elementa.273
Gaudel, A., et al. (2020), Aircraft observations since the 1990s reveal increases of tropospheric ozone at multiple locations across the Northern Hemisphere. Sci. Adv. 6, eaba8272, DOI: 10.1126/sciadv.aba8272
Mousavinezhad, S., Ghahremanloo, M., Choi, Y., Pouyaei, A., Khorshidian, N. and Sadeghi, B., 2023. Surface ozone trends and related mortality across the climate regions of the contiguous United States during the most recent climate period, 1991–2020. Atmospheric Environment, p.119693.
Tarasick, D. W. et al. (2019), Tropospheric Ozone Assessment Report: Tropospheric ozone from 1877 to 2016, observed levels, trends and uncertainties. Elem Sci Anth, 7(1), DOI: http://doi.org/10.1525/elementa.376
Wang, H., et al. (2022), Global tropospheric ozone trends, attributions, and radiative impacts in 1995–2017: an integrated analysis using aircraft (IAGOS) observations, ozonesonde, and multi-decadal chemical model simulations, Atmos. Chem. Phys., 22, 13753–13782, https://doi.org/10.5194/acp-22-13753-2022
Wasserstein, R. L., Schirm, A. L., and Lazar, N. A.: Moving to a world beyond p < 0:05, Am. Stat., 73, 1–29,
https://doi.org/10.1080/00031305.2019.1583913, 2019.Citation: https://doi.org/10.5194/egusphere-2023-1974-RC1 -
AC2: 'Reply on RC1', Heather Simon, 01 Dec 2023
The comment was uploaded in the form of a supplement: https://egusphere.copernicus.org/preprints/2023/egusphere-2023-1974/egusphere-2023-1974-AC2-supplement.pdf
-
AC2: 'Reply on RC1', Heather Simon, 01 Dec 2023
-
CC1: 'Comment on egusphere-2023-1974', David Parrish, 15 Oct 2023
Comment on: “Disappearing day-of-week ozone patterns in US nonattainment areas” by H. Simon et al.
David D. Parrish1 and Richard G. Derwent2
1 David.D.Parrish, LLC, 4630 MacArthur Ln, Boulder, Colorado, USA
2 rdscientific, Newbury, Berkshire, UK
The submitted manuscript undertakes an interesting analysis – an examination of day-of-week patterns in ozone concentrations across many US regions. The goal is to illuminate possible changes in regional photochemical environments due to transitions between NOX versus VOC sensitive ozone formation. The authors present a great deal of analysis of both observational data and model results, and reach important conclusions, e.g., “Nine (large US urban) areas have disappearing weekend effect … indicating .. that they are shifting to more NOX-limited conditions: ….” However, there are shortcomings in the authors’ analysis; below we discuss two of these that compromise the robustness of the author’s conclusions if they cannot be adequately addressed.
First, a significant fractional “disappearance” in observed day-of-week ozone patterns must be expected simply due to the decrease in anthropogenic ozone formation driven by effective precursor emission controls implemented over the past decades. Parrish et al. (2016) show that between 1980 and 2015 in the Los Angeles urban area (i.e., California’s South Coast Air Basin) the temporal dependence of the anthropogenic ozone contribution to the observed distribution of MDA8 ozone concentrations is well-described by an exponential decay with an e-folding time of 22 years. Decreases in anthropogenic ozone formation of similar magnitudes have been documented in other US regions (Parrish et al., 2022, and references therein). In the submitted manuscript, the authors consider 18 years (2002 to 2019) during which this decrease would have amounted to a factor of 2.3. Since the authors use an absolute measure of the day-of-week ozone pattern (i.e., mean difference in ozone in ppb between weekends and weekdays, their Equation 1), that measure would be expected to decrease by that same factor, or by 56% over their selected analysis period, even in the absence of any change in photochemical environment. Unless the authors can demonstrate that the disappearance of day-of-week ozone pattern is significantly greater (or lesser) than 56% of the anthropogenic contribution, those disappearances cannot be taken as evidence for a change in the photochemical environment of the ambient atmosphere.
Second, the details of the trend analysis require further discussion. The authors analyze 18 years (2002 to 2019) of observed MDA8 ozone concentrations. However, rather than analyze the 18 individual years, the authors choose to base their analysis on 5-year rolling periods (i.e., 14 different periods covering the 18-year time series). Importantly, there is a great deal of autocorrelation between the 14 different 5-year rolling means. In fact, the 18 years of observations gives fewer than 4 (i.e., 3.6) independent 5-year means. Attempts to derive trends with reliable confidence limits from such a limited number of independent data is quite uncertain. It must be fully appreciated that the use of 5-year rolling means does not improve the confidence limits of derived trends over those of trends derived from the individual years. In particular, use of the Mann-Kendall test to determine the statistical significance of the derived trends in WE-WD O3 differences will give overly optimistic results if the 14 different 5-year rolling periods are considered to be independent data. The rolling means can give plots that apparently illustrate well-defined trends (e.g., upper right graphs in Figures 1-3), but these illustrations are misleading if the autocorrelation of the 5-year means are not fully considered. A complete discussion of the trend analysis and derivation of confidence limits based upon less than 4 independent observational data points is required; if the autocorrelation of the 5-year rolling means has not been adequately considered, then revisions are required.
References
Parrish, D.D., Xu, J., Croes, B., and Shao, M.: Air Quality Improvement in Los Angeles - Perspectives for Developing Cities, Frontiers of Environ Sci. & Engineering, 2016, 10(5): 11 DOI 10.1007/s11783-016-0859-5, 2016.
Parrish, D.D., Faloona, I.C., and Derwent R.G.: Observational-based Assessment of Contributions to Maximum Ozone Concentrations in the western United States, Journal of the Air & Waste Management Association, 72:5, 434–454, https://doi.org/10.1080/10962247.2022.2050962, 2022.
Citation: https://doi.org/10.5194/egusphere-2023-1974-CC1 -
AC3: 'Reply on CC1', Heather Simon, 01 Dec 2023
The comment was uploaded in the form of a supplement: https://egusphere.copernicus.org/preprints/2023/egusphere-2023-1974/egusphere-2023-1974-AC3-supplement.pdf
-
AC3: 'Reply on CC1', Heather Simon, 01 Dec 2023
-
RC2: 'Comment on egusphere-2023-1974', D.A.J. Jaffe, 17 Oct 2023
Review of “Disappearing day-of-week ozone patterns in US nonattainment areas” by Simon et al.
In this analysis the authors evaluate day of week patterns in average O3 and O3 exceedance days for the 2002-2019 timeframe at ~50 different high O3 sites in the U.S. The authors identify several different patterns including disappearing weekend-weekday differences and others. The analysis mixes observations and model results, which I find to be somewhat problematic. In addition I have some statistical concerns that would need to be corrected or clarified before this could be published.
Major points:
- The authors mix observed patterns with model patterns in a way that I believe is misleading. I think its incumbent on the authors to first clearly document what the observations show. Then we can ask how well the model reproduces the observations and what we can learn from the model where it is consistent with the observations, or if not consistent, then why. Certainly there are plenty of NOx observations that could have been used for this work (see Jaffe et al 2020), so I am not sure what is gained by showing and using only the modeled NOx. But NOx is not that big of a concern. Its formaldehyde that I find much more problematic. For formaldehyde, we have much poorer understanding of emissions and chemistry, both of which are essential to understanding the concentrations. Without any evaluation of the modeled formaldehyde, these results should be removed. In other places the authors quote both modeled trends and observed trends and appear to put equal weight on these. That is incorrect, in my opinion, for the reasons stated above.
- I believe the authors may over-state some of the statistical significance due to auto-correlation. This could be true for both the t-tests for individual years and the trends, which use 5-year running means.
- Finally I note that this has a lot of overlap with our earlier analysis (Jaffe et al 2020). We used data for 1995-2020. This analysis uses data for 2002-2019. There are some modest differences, but overall the results are quite similar. I think its essential that the authors clearly describe what is new and/or whether these results are consistent with the earlier analysis. One area that is different is use of probability of exceedance vs mean concentration. The authors seem to want to discount any differences as being due to random variability, but I am not sure that is true. One focuses on the highest days and the other approach focuses on all days in the O3 Do these days have the same VOC-NOx sensitivity?
Abstract: It is important in abstract to describe the scope: All US O3, all US urban areas or all US non-attainment areas. What years? How many regions considered? In addition, I am unclear what it means if you have a “disappearing weekday” effect. The information here is contained in the relative O3 and NOx behavior between weekday and weekend. So the terms “disappearing weekday” is confusing.
Line 25-26: “both datasets” ?
Line 27: The abstract uses area names that are consistent (I think) with EPA designations, but are often rather non-intuitive. For example, Los Angeles – San Bernardino County vs Los Angeles – South Coast. The San Bernardino monitors are in Riverside CBSA, so aren’t these two locations essentially same region. It would be more interesting to include a site closer to downtown LA like Azusa, where we might expect a different pattern.
Line 33: It is not clear what model evaluation for this work. As near as I can tell, nothing was shown about the models ability to capture year-year variations. The model does seem to capture the trend in weekend-weekday differences.
132-135: While I understand why you excluded 3 out of 7 days, does this change the results?
144: Not clear how t-tests were done. I think you took every weekday and weekend day in one year and compared the means and treated each day as an independent observation. If this is right, then I don’t think autocorrelation was taken into account. In any case please clarify how the t-tests were performed.
169: Given the 5 year running means, these values will have sig autocorrelation. Was this taken into account in the results?
173-174: Unclear meaning.
193-194: Unclear meaning.
Figure 1: Please clarify meaning of P values in top right plot. I think these are for each individual year, correct? Given that the NOx and CH2O plots are for all years, not sure what is the value in showing these. There are major differences (for NOx) between the early and later part of the data record.
234: As noted above, the terminology “disappearing weekend effect” is very misleading. Its really about the difference between weekend and weekday values.
258-259: So how do we interpret these model obs differences? You may say the random variations impact the obs more than the model, but aren’t these variations important?
283: I don’t think the probability approach is inherently noisier, especially when averaged over several years as you have done. I think this is an interesting spot to do a deeper dive.
Conclusions: As noted above it would be good to understand what is new here. Please add some discussion to clarify, perhaps focusing on the differences between the prob of an exceedance approach and mean O3 approach.
Finally, I note that the regression information in the right plots of figures 1-9 (not 5) is almost impossible to read.
Citation: https://doi.org/10.5194/egusphere-2023-1974-RC2 -
RC3: 'Reply on RC2', D.A.J. Jaffe, 18 Oct 2023
Sorry. The earlier reference is jaffe et al 2022, (not 2020).
https://doi.org/10.1029/2021JD036385
Citation: https://doi.org/10.5194/egusphere-2023-1974-RC3 -
AC1: 'Reply on RC2', Heather Simon, 01 Dec 2023
The comment was uploaded in the form of a supplement: https://egusphere.copernicus.org/preprints/2023/egusphere-2023-1974/egusphere-2023-1974-AC1-supplement.pdf
Peer review completion
Journal article(s) based on this preprint
Viewed
HTML | XML | Total | Supplement | BibTeX | EndNote | |
---|---|---|---|---|---|---|
527 | 195 | 28 | 750 | 49 | 13 | 14 |
- HTML: 527
- PDF: 195
- XML: 28
- Total: 750
- Supplement: 49
- BibTeX: 13
- EndNote: 14
Viewed (geographical distribution)
Country | # | Views | % |
---|
Total: | 0 |
HTML: | 0 |
PDF: | 0 |
XML: | 0 |
- 1
Christian Hogrefe
Andrew Whitehill
Kristen M. Foley
Jennifer Liljegren
Norm Possiel
Benjamin Wells
Barron H. Henderson
Lukas C. Valin
Gail Tonnesen
K. Wyat Appel
Shannon Koplitz
The requested preprint has a corresponding peer-reviewed final revised paper. You are encouraged to refer to the final revised version.
- Preprint
(2101 KB) - Metadata XML
-
Supplement
(2397 KB) - BibTeX
- EndNote
- Final revised paper