the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
Comparison of the H2O, HDO and δD stratospheric climatologies between the MIPAS-ESA V8, MIPAS-IMK V5 and ACE-FTS V4.1/4.2 satellite datasets
Karen De Los Ríos
Paulina Ordoñez
Gabriele P. Stiller
Piera Raspollini
Marco Gai
Kaley A. Walker
Cristina Peña-Ortiz
Download
- Final revised paper (published on 04 Jun 2024)
- Preprint (discussion started on 20 Jul 2023)
Interactive discussion
Status: closed
-
CC1: 'Comment on egusphere-2023-1348', Geoff Toon, 30 Aug 2023
Publisher’s note: this comment is a copy of RC1 and its content was therefore removed.
Citation: https://doi.org/10.5194/egusphere-2023-1348-CC1 -
RC1: 'Comment on egusphere-2023-1348', Geoff Toon, 31 Aug 2023
General Comments
This paper compares MIPAS and ACE H2O and HDO over the period 2004 to 2012 when MIPAS was working. Similar comparisons have been done previously, such as Lossow et al., 2020, Lossow et al., 2011, Sheese et al., 2017, Ordonez-Perez et al., 2021, Risi et al., 2011, Hogberg 2019. This latest comparison utilizes more recent versions of the data products which are presumably better. Indeed, the authors state " The HDO data version used here differs significantly from the data versions assessed by Lossow et al. (2020) and Högberg et al. (2019) and used by Steinwagner et al. (2007, 2010)". In this latest paper, a new MIPAS product is presented for the first time: MIPAS-ESA. This and MIPAS-IMK are compared with ACE and with each other. It is not clear to me where, in the processing chain, these two MIPAS products diverge. The retrieval methods seem to be different. Not clear whether the spectra are the same.
To be honest, I didn't feel that I learned much reading this paper. There have already been several similar comparisons using earlier versions of the MIPAS and ACE data products. The authors show that the MIPAS-IMK H2O and HDO products agree well with ACE, but the MIPAS-ESA HDO profiles are discrepant around 40 km altitude. Since the error bars on the MIPAS HDO profiles are quite large above 30 km, the discrepancy is not significant.
Line 21 states "Stratospheric H2O and HDO global average coincident profiles reveal good agreement." I disagree. In my opinion a 37.5% bias in HDO at 40 km is not good agreement. Although the MIPAS HDO profiles have large enough uncertainties such that they bridge this 37.5% gap, this doesn't mean that the agreement is good. It just means that the MIPAS HDO measurements are not useful at 40 km and above.
Specific Comments
Lines 24-25 of the abstract state "ACE-FTS agrees better to MIPAS-IMK than MIPAS-ESA, with biases of -4.8% and -37.5%, respectively. The HDO bias between MIPAS-IMK and MIPAS ESA is 28.1 % at this altitude" So ACE is 4.8% lower than MIPAS-IMK and 37.5% lower than MIPAS-ESA. One might naively expect MIPAS-IMK HDO to be 37.5-4.8 = 32.7% lower than MIPAS-ESA. But it is only 28.1% lower. Presumably this is because different data were used for comparing ACE with MIPAS, than comparing MIPAS-IMK and MIPAS-ESA. Perhaps this should be made clearer in the text.
Lines 29-30 of the abstract state: " ...aligns more closely with expected stratospheric behavior for the entire stratosphere". Delete "stratospheric" or " for the entire stratosphere" . This is unnecessary to have both. Also, this sentence states that MIPAS-IMK calculates dD. I consider it more of a measurement. A model would calculate dD.
Line 15: I've not seen the word "isotopological" before. According to Google it is a mathematical term meaning "having the same topology". Perhaps the authors mean "isotopic"? Line 29 of the abstract uses "isotopic" in a similar context. The word "isotopological" occurs later in the paper, e.g. lines 54, 56. So I'm not sure if the authors are trying to make a distinction between "isotopological" and "isotopic", or they consider these terms synonymous. I suggest that "isotopological" NOT be used, because mathematicians have already defined this word for use in topology.
Line 54 states: "isotopological composition of WV" change to "isotopic composition of WV"
Line 56 states: "Among the isotopological species of WV..." Change to "Among the isotopologues of WV...".
Having read the paper, what I would really like to know is why MIPAS-IMK and MIPAS-ESA HDO are different. Presumably these products come from the same raw data. It undermines confidence in MIPAS to see different groups obtain such different results.
Example of duplication; Line 25-26: The meridional cross-sections of H2O and HDO exhibit the expected distribution that has been established in previous studies. Lines 27-28: The meridional cross-sections of δD are in good agreement with the previous version of MIPAS-IMK and ACE-FTS data. The sentence on lines 27-28 seems superfluous. Given that H2O and HDO are in good agreement with previous datasets and studies, readers will assume that dD will also be in good agreement. No need to tell them that it is.
Raspollini et al., 2022 is cited on lines 86 and 148, but doesn't exist in the References. Either these citations are typos (should be 2020, perhaps?), or the Raspollini 2022 reference is missing.
Line 106 states "we focus here on newer data versions that cover the full mission period of ten years.". If the newer data versions cover 10 years, why do all the tables and figures cover only eight years (2004-2012)? Also, this sentence is missing a final ". "
Lines 152 to 154: It seems that for the MIPAS-ESA processing, different retrieval methods were employed for H2O and HDO. The text needs to explain why this was necessary. Also, why is it "opportune" to use an a priori atmospheric HDO profile that is (3.107 × 10-4) of that of H2O. This is the value in VSMOW, not the atmosphere. In the UTLS the HDO/H2O ratio is closer to (1 × 10-4) so using the 3.107 × 10-4 value might adversely bias the HDO retrievals.
Tables 1 and 2 can be put side-by-side and hence merged into a single table.
Figure 1. Why are the ACE/MIPAS-IMK coincidences ~5 deg. to the South of the ACE/MIPAS-ESA coincidences? So there is no overlap in the ACE data used for MIPAS-IMK validation and for MIPAS-ESA validation -- they are at different latitudes and hence dates. I don't understand why the same ACE data can't be used for both.
Figure 2 should have 1 panel with 3 curves in different colors showing the number of coincidences between: (1) MIPAS-IMK and ACE, (2) MIPAS-ESA and ACE, and (3) MIPAS-IMK and MIPAS-ESA. This will provide the reader more information in less space.
Figure 3: I don't understand the rationale for comparing ACE separately to each MIPAS version. This requires 4 panels and repeats the ACE profiles. Why not have two panels; one for H2O and the other for HDO? Each panel contains the 3 profiles (ACE, MIPAS-IMK, MIPAS-ESA) in different colors. I guess the reason is that ACE data compared with MIPAS-IMK is different from that compared with MIPAS-ESA. In which case you need 4 profiles in each panel: MIPAS-IMK, MIPAS-ESA, ACEIMK, ACEESA.
Figure 6 should be appended to the bottom of fig.5, making a single figure with a single caption. This will allow the reader to compare the features in the dD panels with those in the H2O and HDO panels. This won't be possible with the dD panels on a different page. it will also eliminate repetition in the caption.
Similarly, fig. 8 should be appended to the right of fig.7. It has exactly the same x- and y-axes.
Line 465 states: " the MIPAS instrument shown a negative bias at the troposphere" Change to " the MIPAS instrument shows a negative bias at the troposphere"
Line 420 states:" The general distribution of HDO (Figs 5(c) and 5(d)) shows some similarities to that of H2O (Fig. 5(a) and 5(b)), reflecting that both species have a common in situ source in the stratosphere, i.e., oxidation of CH4 and H2." But HDO comes from CH3D and HD, so change sentence to: " The general distribution of HDO (Fig.s 5(c) and 5(d)) shows some similarities to that of H2O (Fig. 5(a) and 5(b)), reflecting that both species have a common in situ source in the stratosphere, i.e., oxidation of methane and hydrogen."
Line 447 states: "...diagrams over 30S and 30N...". This is ambiguous. Perhaps "...diagrams covering 30S to 30N..."
Line 465:"As it was previously shown in the Fig.3, the MIPAS instrument shown a negative bias at the troposphere" Three grammatical errors in this half sentence. Change to: " As was previously shown in Fig.3, the MIPAS instrument shows a negative bias at the troposphere". Also, I don't see anything negative in Fig.3. Perhaps the authors mean Fig.4?
Line 490 states: " The analysis conducted in this study highlights a higher level of agreement in HDO measurements obtained from ACE-FTS in both comparison cases." This seems to imply that ACE agrees better with MIPAS than MIPAS-IMK and MIPAS-ESA agree with each other?
Line 524 states: "the findings from this study suggest that the MIPAS-IMK dataset provides a more realistic signal for the entire stratosphere". More realistic than what? ACE or MIPAS-ESA.
Line 526 states: "it is crucial to exercise caution when interpreting these results, specifically considering the sampling limitations of ACE-FTS in the tropics, during the period of study, especially at lower altitudes." I don't recall much discussion of this in the main part of the paper. It is true that ACE occultations are sparse in the tropics and that high clouds can often limit penetration of the troposphere. But it seems unfair to ACE to call this an conclusion. And it not clear what altitude range this comment is aimed.
Line 530: "The code in MATLAB is available from the authors upon request." The term "the code" is too vague. Add one sentence explaining what "the code" does.
The format of the References is unfriendly. There is no indentation at the start of a new reference, nor a gap between references. So it is hard to tell where one reference ends and the next begins. Perhaps this is the journal style.
Citation: https://doi.org/10.5194/egusphere-2023-1348-RC1 -
AC2: 'Reply on RC1', Paulina Ordóñez, 07 Nov 2023
Publisher’s note: this comment is a copy of AC3 and its content was therefore removed.
Citation: https://doi.org/10.5194/egusphere-2023-1348-AC2 -
AC3: 'Reply on RC1', Paulina Ordóñez, 07 Nov 2023
General Comments
We thank the reviewer for his/her constructive comments, which will result in an improvement of the manuscript.
1. This paper compares MIPAS and ACE H2O and HDO over the period 2004 to 2012 when MIPAS was working. Similar comparisons have been done previously, such as Lossow et al., 2020, Lossow et al., 2011, Sheese et al., 2017, Ordonez-Perez et al., 2021, Risi et al., 2011, Hogberg 2019. This latest comparison utilizes more recent versions of the data products which are presumably better. Indeed, the authors state " The HDO data version used here differs significantly from the data versions assessed by Lossow et al. (2020) and Högberg et al. (2019) and used by Steinwagner et al. (2007, 2010)". In this latest paper, a new MIPAS product is presented for the first time: MIPAS-ESA. This and MIPAS-IMK are compared with ACE and with each other. It is not clear to me where, in the processing chain, these two MIPAS products diverge. The retrieval methods seem to be different. Not clear whether the spectra are the same.
The two products differ already in the version of the spectral data. While the ESA product uses Level-1b version 8.03, the IMK data uses version 5.02/5.06. The differences between the versions result from several factors including improved calibration procedures, a compensation of the detector drift due to aging, and subtle adjustments of the geolocations. We are aware that it is not optimal to use different level-1b data versions, however, it is unavoidable in this case, since there are no earlier HDO data versions of the ESA product, and HDO from version 8 is not yet available from the IMK processing.
The level-2 processing (retrieval of trace gases) has many substantial differences between the processors: in general, MIPAS IMK and MIPAS ESA use different spectral intervals, and there is a rich literature about the differences of the retrieval set ups and the reasons for these choices. In particular, the two papers Laeng et al., 2015, and Raspollini et al., 2013 describe differences between the algorithms performing MIPAS analysis and the differences in the products for O3 and other trace species, but not HDO. For additional papers on the MIPAS IMK data set and retrievals, we refer one to the web site on the IMK products (https://www.imk-asf.kit.edu/english/298.php), while the evolution of MIPAS ESA Level 2 algorithm and its products is described in these papers (Ridolfi et al., 2000; Raspollini et al., 2006, Raspollini et al., 2013, Raspollini et al., 2022, Dinelli et al., 2021) and references therein; it is beyond the scope of this paper to summarize them all.
We will clarify this in the revised version of the manuscript.
2. To be honest, I didn't feel that I learned much reading this paper. There have already been several similar comparisons using earlier versions of the MIPAS and ACE data products.
It is true that several comparisons have already been made. Recently, the SPARC-WAVAS-II activity compared H2O and HDO from all available satellite instruments since the year 2000, also including ACE-FTS version 3.5, MIPAS IMK version 5, and ESA version 5 and 7 (only for H2O) (see ACP/AMT Special Issue “Water vapour in the upper troposphere and middle atmosphere: a WCRP/SPARC satellite data quality assessment including biases, variability, and drifts”, https://amt.copernicus.org/articles/special_issue10_830.html). Newer data versions of H2O and HDO have been used in our paper here for all three data sets. We think that it is useful to assess the quality of new data products from satellite data, in particular here the new MIPAS ESA product, and it is necessary to do so before further scientific work with these data can be done.
Particularly important is the case of using δD data to study the origin of water vapor that enters the stratosphere. Therefore, it is critical to understand the quality of the H2O and HDO that are used in this calculation as is a major focus of this paper.
3. The authors show that the MIPAS-IMK H2O and HDO products agree well with ACE, but the MIPAS-ESA HDO profiles are discrepant around 40 km altitude. Since the error bars on the MIPAS HDO profiles are quite large above 30 km, the discrepancy is not significant.
Line 21 states "Stratospheric H2O and HDO global average coincident profiles reveal good agreement." I disagree. In my opinion a 37.5% bias in HDO at 40 km is not good agreement. Although the MIPAS HDO profiles have large enough uncertainties such that they bridge this 37.5% gap, this doesn't mean that the agreement is good. It just means that the MIPAS HDO measurements are not useful at 40 km and above.
We agree with this comment. In addition, we note that ACE-FTS H2O has a significant deviation from the other two data records. This deviation was also found in the SPARC/WAVAS-II comparisons for earlier data versions (see, for example, Lossow et al., 2019) with respect to many other satellite data records. Therefore, because of these known discrepancies, we will restrict our extended analyses (those after discussion of Fig. 3) to altitudes below 30 km in the revised version of the paper.
Specific Comments
4. Lines 24-25 of the abstract state "ACE-FTS agrees better to MIPAS-IMK than MIPAS-ESA, with biases of -4.8% and -37.5%, respectively. The HDO bias between MIPAS-IMK and MIPAS ESA is 28.1 % at this altitude" So ACE is 4.8% lower than MIPAS-IMK and 37.5% lower than MIPAS-ESA. One might naively expect MIPAS-IMK HDO to be 37.5-4.8 = 32.7% lower than MIPAS-ESA. But it is only 28.1% lower. Presumably this is because different data were used for comparing ACE with MIPAS, than comparing MIPAS-IMK and MIPAS-ESA. Perhaps this should be made clearer in the text.
Indeed, a different number of coincident profiles were used in each of the three comparisons. There were differences in the geolocations because the values were arithmetically averaged if multiple coincidences were found in each of the coincidence regions, as we wrote in the text. However, we concur with the referee in the necessity of compare data with the same geolocation.
In the revised version of the manuscript, when multiple MIPAS profiles are spatially coincident with an ACE-FTS profile, the MIPAS profile closest in time is selected. In addition, there must be both MIPAS IMK and MIPAS ESA processed data available for this coincident profile.
5. Lines 29-30 of the abstract state: " ...aligns more closely with expected stratospheric behavior for the entire stratosphere". Delete "stratospheric" or " for the entire stratosphere". This is unnecessary to have both.
Will be done.
6. Also, this sentence states that MIPAS-IMK calculates dD. I consider it more of a measurement. A model would calculate dD.
Will be adjusted in the paper.
7. Line 15: I've not seen the word "isotopological" before. According to Google it is a mathematical term meaning "having the same topology". Perhaps the authors mean "isotopic"? Line 29 of the abstract uses "isotopic" in a similar context. The word "isotopological" occurs later in the paper, e.g. lines 54, 56. So I'm not sure if the authors are trying to make a distinction between "isotopological" and "isotopic", or they consider these terms synonymous. I suggest that "isotopological" NOT be used, because mathematicians have already defined this word for use in topology.
8. Line 54 states: "isotopological composition of WV" change to "isotopic composition of WV"
9. Line 56 states: "Among the isotopological species of WV..." Change to "Among the isotopologues of WV...".
Thank you very much for the constructive comment. We were making a distinction between "isotopological" and "isotopic" since “isotopic” is the adjective for “isotope” and "isotopological" is the adjective for “isotopologue”.
Certainly, two terms - “isotopologic” (e.g., Herbin et al., 2007; Bahr and Wolff, 2022) and “isotopological” (e.g., Schneider et al., 2020; Israel, 2023) - can be found in the literature to describe characteristics or properties related to “isotopologues”. Since the term “isotopological” is also a mathematical term meaning "having the same topology", we think that the suggestion of the referee in not using the term “isotopological” is right, and term “isotopologic” will be used in the revised version of the manuscript.
10. Having read the paper, what I would really like to know is why MIPAS-IMK and MIPAS-ESA HDO are different. Presumably these products come from the same raw data. It undermines confidence in MIPAS to see different groups obtain such different results.
We respectfully disagree. It is correct that the MIPAS ESA and MIPAS IMK HDO product comes from the same raw data (interferograms), but the spectral data are from different level-1b data versions. The level-2 processing (the retrieval set-up in this case) is a most relevant part in the data generation. Even the same level-2 processor produces different results with different retrieval settings. In cases where several level-2 processors of other satellite data exist, they often result in differing products (e.g., GOMOS, SCIAMACHY, SMILES, OMPS). The level-1b data for the Envisat instruments were made public to encourage different processing techniques to be developed and applied.
Further, we would like to point out that the differences between the two MIPAS products are rather limited (see Fig. 4). H2O differences are close to 5% at max, while HDO differences remain <10% (all below 30 km). Differences as large as this can occur between different versions of the same data product of the same processor.
11. Example of duplication; Line 25-26: The meridional cross-sections of H2O and HDO exhibit the expected distribution that has been established in previous studies. Lines 27-28: The meridional cross-sections of δD are in good agreement with the previous version of MIPAS-IMK and ACE-FTS data. The sentence on lines 27-28 seems superfluous. Given that H2O and HDO are in good agreement with previous datasets and studies, readers will assume that δD will also be in good agreement. No need to tell them that it is.
As the δD here is calculated from individual profiles then averaged rather than mean profiles, it is important to point out this good agreement. From our experience, even subtle differences in H2O and HDO (such as those between versions) are enlarged severely by calculating δD from them.
12. Raspollini et al., 2022 is cited on lines 86 and 148, but doesn't exist in the References. Either these citations are typos (should be 2020, perhaps?), or the Raspollini 2022 reference is missing.
The reviewer is right, the reference Raspollini et al., 2022 is missing in the list of references, while Raspollini et al., 2020 is correctly reported in the list. We also updated the DOI of reference Dinelli et al., 2021.
13. Line 106 states "we focus here on newer data versions that cover the full mission period of ten years.". If the newer data versions cover 10 years, why do all the tables and figures cover only eight years (2004-2012)? Also, this sentence is missing a final ". "
The reviewer is right. The sentence is corrected in the revised version of the manuscript since we are focusing on the overlap period between MIPAS and ACE-FTS which is from 2004 to 2012.
14. Lines 152 to 154: It seems that for the MIPAS-ESA processing, different retrieval methods were employed for H2O and HDO. The text needs to explain why this was necessary. Also, why is it "opportune" to use an a priori atmospheric HDO profile that is (3.107 × 10-4) of that of H2O. This is the value in VSMOW, not the atmosphere. In the UTLS the HDO/H2O ratio is closer to (1 × 10-4) so using the 3.107 × 10-4 value might adversely bias the HDO retrievals.
A different retrieval approach has been used for H2O and HDO because the approach used for H2O (namely a Levenberg-Marquardt regularization approach within the iterations followed by an a posteriori regularization) was not sufficient to constrain the HDO retrieval. For the HDO, an a priori error of 100% is used in order not to introduce a bias, as written in the text (line 161).
15. Tables 1 and 2 can be put side-by-side and hence merged into a single table.
Thank you for the recommendation. The tables will be merged.
16. Figure 1. Why are the ACE/MIPAS-IMK coincidences ~5 deg. to the South of the ACE/MIPAS-ESA coincidences? So, there is no overlap in the ACE data used for MIPAS-IMK validation and for MIPAS-ESA validation -- they are at different latitudes and hence dates. I don't understand why the same ACE data can't be used for both.
Thank you for the comment. The referee is right, there were differences in the geolocations of the same profiles in the two data versions due to the method we used for determining the coincident profiles (see comment #4). Now the same ACE-FTS data have been used for both comparisons.
17. Figure 2 should have 1 panel with 3 curves in different colors showing the number of coincidences between: (1) MIPAS-IMK and ACE, (2) MIPAS-ESA and ACE, and (3) MIPAS-IMK and MIPAS-ESA. This will provide the reader more information in less space.
Done in the revised version of the manuscript.
18. Figure 3: I don't understand the rationale for comparing ACE separately to each MIPAS version. This requires 4 panels and repeats the ACE profiles. Why not have two panels; one for H2O and the other for HDO? Each panel contains the 3 profiles (ACE, MIPAS-IMK, MIPAS-ESA) in different colors. I guess the reason is that ACE data compared with MIPAS-IMK is different from that compared with MIPAS-ESA. In which case you need 4 profiles in each panel: MIPAS-IMK, MIPAS-ESA, ACEIMK, ACEESA.
We agree with the reviewer suggestion, the three curves will be inserted in one panel.
We also improve the figure 3 including the MIPAS-IMK to MIPAS-ESA comparison.
19. Figure 6 should be appended to the bottom of fig.5, making a single figure with a single caption. This will allow the reader to compare the features in the dD panels with those in the H2O and HDO panels. This won't be possible with the dD panels on a different page. it will also eliminate repetition in the caption.
Thank you for the recommendation. This is done in the revised version of the manuscript.
Similarly, fig. 8 should be appended to the right of fig.7. It has exactly the same x- and y-axes.
Yes, done in the revised version of the manuscript. Thank you for the suggestion.
Line 465 states: " the MIPAS instrument shown a negative bias at the troposphere" Change to " the MIPAS instrument shows a negative bias at the troposphere"
Done in the revised version of the manuscript.
20. Line 420 states:" The general distribution of HDO (Figs 5(c) and 5(d)) shows some similarities to that of H2O (Fig. 5(a) and 5(b)), reflecting that both species have a common in situ source in the stratosphere, i.e., oxidation of CH4 and H2." But HDO comes from CH3D and HD, so change sentence to: " The general distribution of HDO (Fig.s 5(c) and 5(d)) shows some similarities to that of H2O (Fig. 5(a) and 5(b)), reflecting that both species have a common in situ source in the stratosphere, i.e., oxidation of methane and hydrogen."
Done in the revised version of the manuscript.
21. Line 447 states: "...diagrams over 30S and 30N...". This is ambiguous. Perhaps "...diagrams covering 30S to 30N..."
Done in the revised version of the manuscript.
22. Line 465:"As it was previously shown in the Fig.3, the MIPAS instrument shown a negative bias at the troposphere" Three grammatical errors in this half sentence. Change to: " As was previously shown in Fig.3, the MIPAS instrument shows a negative bias at the troposphere".
Done in the revised version of the manuscript.
23. Also, I don't see anything negative in Fig.3. Perhaps the authors mean Fig.4?
Thank you for the observation. Yes, it is figure 4, it is changed in the revised version of the manuscript.
24. Line 490 states: " The analysis conducted in this study highlights a higher level of agreement in HDO measurements obtained from ACE-FTS in both comparison cases." This seems to imply that ACE agrees better with MIPAS than MIPAS-IMK and MIPAS-ESA agree with each other?.
It was not our intention to imply this. In fact, the agreement between the two MIPAS products of H2O is at least as good (below 30 km), as that between ACE-FTS and one of the MIPAS products. Regarding HDO, the differences between the two MIPAS data sets are somewhat larger, indeed, especially between 20 and 30 km, but still < 10%.
The text will be modified in the updated version of the manuscript.
25. Line 524 states: "the findings from this study suggest that the MIPAS-IMK dataset provides a more realistic signal for the entire stratosphere". More realistic than what? ACE or MIPAS-ESA.
“More realistic than MIPAS-ESA” was to be meant. However, we will reworde the sentence in the revised version of the manuscript since the affirmation "more realistic" is inaccurate given the existing dD data.
26. Line 526 states: "it is crucial to exercise caution when interpreting these results, specifically considering the sampling limitations of ACE-FTS in the tropics, during the period of study, especially at lower altitudes." I don't recall much discussion of this in the main part of the paper. It is true that ACE occultations are sparse in the tropics and that high clouds can often limit penetration of the troposphere. But it seems unfair to ACE to call this a conclusion. And it not clear what altitude range this comment is aimed.
Thank you for the comment, we concur with the reviewer that this is not a conclusion of this analysis. The last paragraph of the revised version of manuscript will be modified.
27. Line 530: "The code in MATLAB is available from the authors upon request." The term "the code" is too vague. Add one sentence explaining what "the code" does.
Changed in the revised version of the manuscript.
The format of the References is unfriendly. There is no indentation at the start of a new reference, nor a gap between references. So it is hard to tell where one reference ends and the next begins. Perhaps this is the journal style.
Thank you for the observation. Changed.
References:
Israel, F. P. (2023). Central molecular zones in galaxies: Multitransition survey of dense gas tracers HCN, HNC, and HCO+. Astronomy and Astrophysics, 671, A59.
Laeng, A.; Hubert, D.; Verhoelst, T.; von Clarmann, T.; Dinelli, B. M.; Dudhia, A.; Raspollini, P.; Stiller, G.; Grabowski, U.; Keppens, A.; Kiefer, M.; Sofieva, V.; Froidevaux, L.; Walker, K. A.; Lambert, J. -C.; Zehner, C., The ozone climate change initiative: Comparison of four Level-2 processors for the Michelson Interferometer for Passive Atmospheric Sounding (MIPAS), REMOTE SENSING OF ENVIRONMENT, https://doi.org/10.1016/j.rse.2014.12.013, 2015
Lossow, S., Khosrawi, F., Kiefer, M., Walker, K. A., Bertaux, J.-L., Blanot, L., Russell, J. M., Remsberg, E. E., Gille, J. C., Sugita, T., Sioris, C. E., Dinelli, B. M., Papandrea, E., Raspollini, P., García-Comas, M., Stiller, G. P., von Clarmann, T., Dudhia, A., Read, W. G., Nedoluha, G. E., Damadeo, R. P., Zawodny, J. M., Weigel, K., Rozanov, A., Azam, F., Bramstedt, K., Noël, S., Burrows, J. P., Sagawa, H., Kasai, Y., Urban, J., Eriksson, P., Murtagh, D. P., Hervig, M. E., Högberg, C., Hurst, D. F., and Rosenlof, K. H.: The SPARC water vapour assessment II: profile-to-profile comparisons of stratospheric and lower mesospheric water vapour data sets obtained from satellites, Atmos. Meas. Tech., 12, 2693–2732, https://doi.org/10.5194/amt-12-2693-2019, 2019.
Raspollini, Piera; Arnone, Enrico; Barbara, Flavio; Carli, Bruno; Castelli, Elisa; Ceccherini, Simone; Dinelli, Bianca Maria; Dudhia, Anu; Kiefer, Michael; Papandrea, Enzo; Ridolfi, Marco, Comparison of the MIPAS products obtained by four different level 2 processors, ANNALS OF GEOPHYSICS, https://doi.org/10.4401/ag-6338, 2013
Raspollini, P., Belotti, C., Burgess, A., Carli, B., Carlotti, M., Ceccherini, S., Dinelli, B. M., Dudhia, A., Flaud, J.-M., Funke, B., Höpfner, M., López-Puertas, M., Payne, V., Piccolo, C., Remedios, J. J., Ridolfi, M., and Spang, R.: MIPAS level 2 operational analysis, Atmos. Chem. Phys., 6, 5605–5630, https://doi.org/10.5194/acp-6-5605-2006, 2006.
Raspollini, P., Carli, B., Carlotti, M., Ceccherini, S., Dehn, A., Dinelli, B. M., Dudhia, A., Flaud, J.-M., López-Puertas, M., Niro, F., Remedios, J. J., Ridolfi, M., Sembhi, H., Sgheri, L., and von Clarmann, T.: Ten years of MIPAS measurements with ESA Level 2 processor V6 – Part 1: Retrieval algorithm and diagnostics of the products, Atmos. Meas. Tech., 6, 2419–2439, https://doi.org/10.5194/amt-6-2419-2013, 2013
Ridolfi, B. Carli, M. Carlotti, T. von Clarmann, B. M. Dinelli, A. Dudhia, J.-M. Flaud, M. Höpfner, P. E. Morris, P. Raspollini, G. Stiller, and R. J. Wells, “Optimized forward model and retrieval scheme for MIPAS near-real-time data processing”, Applied Optics, 39, 1323-1340 (2000).
Schneider, A., Borsdorff, T., Aemisegger, F., Feist, D. G., Kivi, R., Hase, F., ... & Landgraf, J. (2020). First data set of H 2 O/HDO columns from the Tropospheric Monitoring Instrument (TROPOMI). Atmospheric Measurement Techniques, 13(1), 85-100.
Zeng, Z. C., Addington, O., Pongetti, T., Herman, R. L., Sung, K., Newman, S., ... & Sander, S. P. (2022). Remote sensing of atmospheric HDO/H2O in southern California from CLARS-FTS. Journal of Quantitative Spectroscopy and Radiative Transfer, 288, 108254.
Zeng, Z. C., Addington, O., Pongetti, T., Herman, R. L., Sung, K., Newman, S., ... & Sander, S. P. (2022). Remote sensing of atmospheric HDO/H2O in southern California from CLARS-FTS. Journal of Quantitative Spectroscopy and Radiative Transfer, 288, 108254.
Lossow, S., Steinwagner, J., Urban, J., Dupuy, E., Boone, C. D., Kellmann, S., ... & Stiller, G. P. (2011). Comparison of HDO measurements from Envisat/MIPAS with observations by Odin/SMR and SCISAT/ACE-FTS. Atmospheric Measurement Techniques, 4(9), 1855-1874.
Lossow, S., Högberg, C., Khosrawi, F., Stiller, G. P., Bauer, R., Walker, K. A., ... & Eichinger, R. (2020). A reassessment of the discrepancies in the annual variation of δD-H 2 O in the tropical lower stratosphere between the MIPAS and ACE-FTS satellite data sets. Atmospheric Measurement Techniques, 13(1), 287-308.
Raspollini, P., Arnone, E., Barbara, F., Bianchini, M., Carli, B., Ceccherini, S., Chipperfield, M. P., Dehn, A., Della Fera, S., Dinelli, B. M., Dudhia, A., Flaud, J.-M., Gai, M., Kiefer, M., López-Puertas, M., Moore, D. P., Piro, A., Remedios, J. J., Ridolfi, M., Sembhi, H., Sgheri, L., and Zoppetti, N.: Level 2 processor and auxiliary data for ESA Version 8 final full mission analysis of MIPAS measurements on ENVISAT, Atmos. Meas. Tech., 15, 1871–1901, https://doi.org/10.5194/amt-15-1871-2022, 2022.
Dinelli, B. M., Raspollini, P., Gai, M., Sgheri, L., Ridolfi, M., Ceccherini, S., Barbara, F., Zoppetti, N., Castelli, E., Papandrea, E., Pettinari, P., Dehn, A., Dudhia, A., Kiefer, M., Piro, A., Flaud, J.-M., López-Puertas, M., Moore, D., Remedios, J., and Bianchini, M.: The ESA MIPAS/Envisat level2-v8 dataset: 10 years of measurements retrieved with ORM v8.22, Atmos. Meas. Tech., 14, 7975–7998, https://doi.org/10.5194/amt-14-7975-2021, 2021.
Bahr M-S and Wolff M (2022), PAS-based isotopologic analysis of highly concentrated methane. Front. Environ. Chem. 3:1029708. doi: 10.3389/fenvc.2022.1029708.
Herbin, D. Hurtmans, S. Turquety, C. Wespes, B. Barret, J. Hadji-Lazaro, C. Clerbaux, and P.-F. Coheur (2007). Global distributions of water vapour isotopologues retrieved from IMG/ADEOS data. Atmos. Chem. Phys., 7, 3957 - 3968.
-
AC2: 'Reply on RC1', Paulina Ordóñez, 07 Nov 2023
-
RC2: 'Comment on egusphere-2023-1348', Anonymous Referee #2, 03 Oct 2023
Comparison of the H2O, H2O and \delta D stratospheric climatologies
de los Rios et al.Referee report
Overview
The paper compares the H2O and HDO data retrieved from the ACE-FTS
solar occultation instrument and from two different retrieval
algorithms applied to the MIPAS limb-emission instrument. This is an
update on similar work performed by Hogberg and Lossow in 2019 but
using reprocessed ACE-FTS and MIPAS-ESA data. However, it is
difficult to know what conclusions can be reached, and how or if these
have changed with the new data.This seems a rather 'mechanical' paper - mostly just reproducing
earlier results, the only new aspect being the updated
datasets. Indeed, the sort of paper one can imagine being generated,
in a few years if not already, by some of the more advanced AI
machines.It might have been acceptable if the original authors wanted to update
their paper with new results, in which case I would expect a narrative
focusing on the algorithm changes, and the expected and observed
impacts on the intercomparisons with respect to their earlier results
rather than, as here, analysing the results in absolute terms as if
they were being presented for the first time. But if it's a paper with
new lead authors it also needs some significant new insight or
analysis. I also have some doubts about the methodology.General Comments and Suggestions
1) Colocations
The comparison has been performed on the two versions of MIPAS data as
if these were independent satellites. A more satisfactory approach
would have been, first, to apply the respective filters to these MIPAS
datasets and take just the common profiles, and then compare this with
ACE-FTS. This would not only ensure no time/space mismatch between the
two MIPAS datasets but also ensure exactly the same time/space
mismatch between ACE-FTS and either of the MIPAS datasets.On this topic, Fig 1 looks very odd. It seems most unlikely that in a 3 day
period the best ACE-IMK coincidences are in a different latitude to the best
ACE-ESA coincidences - I would expect them to be mostly the same locations.The averaging of all MIPAS profiles within the coincidence criteria also seems
undesirable. The averaging will reduce the random noise in the MIPAS profiles so
the contribution to the overall SD is no longer straightforward. Better to take
just the closest profile. Also, noting the difference in time/space would allow
subsequent analysis as to whether the chosen margins are adequate or, more
ambitiously, allow the colocation error to be quantified eg when switching to
grid boxes or zonal averages.
---------2) Algorithm descriptions
In 2.1.1/2.1.2 the descriptions of the two retrievals seem to be taken
directly from the source papers using their own terminology (possibly via the
SPARC papers), with little effort to standardise the information let alone
provide some interpretation in terms of possible impact on the comparisons.
For example, 'non-linear least-squares global-fitting technique with
Tikhonov regularisation' (for IMK) and 'least-squares global fitting,
using the Gauss-Newton approach modified with the Levenberg-Marquardt
method' plus 'a posteriori regularisation' (for ESA). The reader has
to work quite hard to understand whether or not these are essentially
the same thing and hence whether significant differences may arise
from these aspects of the algorithm.Even on a more basic level:
- Does ESA-MIPAS retrieve H2O and HDO as
log(VMR) or VMR (which affects how you should evalulate biases)?
- Does the IMK-MIPAS account for horizontal inhomogeneities in the
line-of-sight direction and/or assume LTE?
- Do both use microwindows in the same spectral region?Similarly with ACE-FTS, orbital altitude and inclination are measured
but these are not given for Envisat. Spectral bands are given for
ACE-FTS but not for MIPAS.
-------3) Retrieval diagnostics
There is no reference here to the retrieval characteristics such as expected
random noise, systematic errors or averaging kernels, at least typical values
- these are not likely to change much except (for MIPAS) towards the poles where
the atmospheric temperature is low.The SD of the bias, for example, could be put in the context of the retrieval
random errors, and the bias itself in terms of the overall systematic errors.
Meanwhile the averaging kernel describes the ability of the retrieval to follow
'real' atmospheric variations, which has an impact on the correlations as well
as the SD of the comparisons. The lack of an HDO time signature in the ESA data
might be explained in terms of the averaging kernel.
--------4) CH4 consistency
Given the large discrepancy between ACE and the two MIPAS profiles in
the stratosphere a simply self-consistency check would be to see how
these compare with their equivalent CH4 retrievals, on the basis that
(H2O + 2CH4) should be conserved.
--------5) The authors should be aware of the difference in LaTeX between a minus
sign ($-$) and a dash (--) indicating a range of numbers. Where both
positive and negative values are possible it would also help if '+'
were added in front of the positive numbers to further distinguish
them from the --. Thus, in L24: $-$8.6--+10.6 Incidentally, assuming
this is taken from Table 2, the actual number in the table is
-8.7. Further on this particular point, negative and positive biases
don't have any particular meaning in this sentence, so I would just
have said 'biases of up to 10%'.
-------6) There are numerous references to 'global' averages whereas I would suggest
'dataset' averages, or something similar, unless you really claim that your
intercomparison dataset represents some sort of uniform sampling of the globe.
--------7) It is unclear whether or not the MIPAS-IMK product has been updated
(these authors refer to 'V5H' and 'V5R' whereas Hogberg (2019) referred to
V20 as being the newer version).
------8) 'Standard Deviation' is already defined as the spread around some
mean value, so 'debiased standard deviation' is just 'standard
deviation' since we are already talking about mean differences (or
'the standard deviation of the relative bias' which is how it is
described in the caption for Fig.4). Perhaps you thought SD might be
confused with root-mean-square-difference? Also in Tables 1 and 2 this
has become '1 \sigma Bias' which sounds like a different thing
altogether, but I assume it's the same.
--------9) The time series plots could be enhanced by subtracting out the mean profile and
then also removing the average annual cycle to show interannual variability.
The latter may have some QBO correlation which could be investigated.
-------10)
Details of bias determination (3.1.3, 3.2) are, firstly, confusing because of the
repeated use of the same 4 coordinates for each parameter, secondly difficult to
read because of the small fonts and, finally, quite standard.
Even then, there are a few problems here.Eq 1 presents the bias as 4-coordinate average of b_i which are themselves
4-coordinate quantities. However it seems unlikely that any two measurements are
exactly matching in latitude or time (it's not clear what 'period' means here) so
I assume these coordinates are what is being averaged over in i=1,n so should not
appear on the l.h.s. as well. And b_i is presumably also a function of longitude,
also averaged.sigma_x1 and sigma_x2 in Eq (7) are undefined,
$\sigma_b$ has a bar over the b here but not in Fig 4
Fig 4 shows SD as a percentage (of what?) but Eq 5 defines this in absolute terms.
When considering relative bias, if the two datasets were retrieved as
log(VMR), the geometric rather than arithmetic mean of the two values
would be more appropriate as the reference value.I appreciate that this sort of thing was included in the previous
papers (and had I reviewed those I would have said the same thing) but
that's even less reason to include it again here. Everyone knows what
you mean (and they can look up Pearson Correlation on Wikipedia) so no
need to drag the reader through all the small print. On this (rare)
occasion, it really is simpler to explain in words rather than
equations.
---------11)
You could save a lot of wordage simply referring to these data as 'IMK', 'ESA' and
'ACE' (at least in the text, perhaps not in the figures)
---------Minor/Typographical comments
Abstract
L139: Use a regular reference for the SPARC special issue to avoid the typesetting
difficulties caused by putting the URL (http:...) in the text.L185: 'from 1 to 70 km'
L189: Presumably the two MIPAS datasets are automatically colocated so this just
refers to colocating ACE-FTS with MIPAS.Fig 2: I assume the variation in low altitudes is due to cloud-screening but what
causes the reduction in MIPAS-ESA comparison data at high altitude?L201: It would be useful to have at least an approximate figure as to what
percentage of MIPAS profiles fail the quality control tests.L208: In L187 the grid is from 1-70km. Assuming the IMK grid is at 1km intervals
how do you get 58 levels?L304: I would not refer to these as 'error bars', just 'bars'.
L319: Describing these values as 'global' minima is misleading,
they're actually the minima in the intercomparison dataset.L325: Similarly 'global mean' profiles.
Fig 3. These plots would be more informative (and take less space) if all three
datasets were combined on the same plot, allowing MIPAS-IMK and MIPAS-ESA to be
directly compared. Use eg dashed lines to mark the 1sigma variation for each
(Also, I wouldn't call these 'error bars' as in the current caption).L365: It is perhaps worth mentioning that the low SD (Fig 4(c)) between
ESA-ACE and coupled with the high correlation (Fig 4(d)) are consistent
with the MIPAS-IMK retrievals being less sensitive to actual atmospheric
variations in H2O, and conversely for IMK-ACE for HDO.Fig 4(g) mis-labelled(?) as '\sigma_b Bias' (and similarly Table 1)
Table 1 & 2: I assume these figures are a summary of the plots shown
in Fig 4, in which case make it explicit in the Table caption (and similarly in the
Fig 4 caption refer to these tables). Rather than have rows labelled eg
'MIPAS-IMK vs 'ACE-FTS' I suggest replacing 'vs' with $-$ to make it absolutely
clear which way around you are defining the sign of the bias.L393: Is there such a thing as a 1\sigma standard deviation? I though the SD was,
by definition, 1 \sigma.Fig 5 caption refers to 'meridional cross-sections' which suggests a single slice
through a particular longitude. 'Zonal mean' or 'Zonally averaged' distributions
are the usual terminology for such plots. And rather than use 'summer' and
'winter' - which differ from north to south hemispheres - give the actual
months averaged (as in Fig 6).L405 I'm assuming that the labelling on Fig 5 is correct (I would have expected
ACE-FTS to be missing the high latitude measurements during the local
winter months but, on the other hand, they may also specifically
target as high latitude as possible during polar winter months).
However the very low ACE values for both H2O and HDO in the winter polar vortex
(compared to the two MIPAS datasets) are worthy of comment.Also, why is the IMK cloud filtering any different for HDO than for H2O?
L450: 'interannual variability' implies what's left after removing the annual
cycle. From Figs 7a,b,c alone it is not clear to me that ACE-IMK have the
greatest similarity.L465: 'shows' rather than 'shown'.
L467: 'while the ACE-FTS instrument can measure with higher sensitivity ...'
What is the basis for this statement?L485: 'vertical behavior' (or 'behaviour') presumably just refers to the small
mean profile bias averaged over the whole dataset, shown by the red-lines in
Fig 4a,b. 'Behaviour' suggests that the two track each other well in other
respects such as low SD and high correlation, which they don't (at least not
above 20km).L487: 'according to the uncertainties' Which uncertainties are these? There was
no reference to estimated uncertainties in the datasets in the main part of
the paper. Given that the two MIPAS profiles are 1ppmv higher than ACE
after averaging over several thousand profiles suggests rather a
significant discrepancy to me.L495: 'quantitaty' ?
L525: 'MIPAS-IMK dataset provides a more realistic signal for the entire
stratosphere'
This is a bold statement and needs some qualification otherwise it could get
quoted out of context.
Is this for H2O, HDO, delta D or all three?
Is it in terms of the time-evolution or mean profile?
Is it because it has the best correlation or lowest SD compared to both
other datasets?---------------
Citation: https://doi.org/10.5194/egusphere-2023-1348-RC2 -
AC4: 'Reply on RC2', Paulina Ordóñez, 07 Nov 2023
Overview
We thank the reviewer for his/her comments, which will result in an improvement of the manuscript.
The paper compares the H2O and HDO data retrieved from the ACE-FTS solar occultation instrument and from two different retrieval algorithms applied to the MIPAS limb-emission instrument. This is an update on similar work performed by Hogberg and Lossow in 2019 but using reprocessed ACE-FTS and MIPAS-ESA data. However, it is difficult to know what conclusions can be reached, and how or if these have changed with the new data.
In addition to the new ACE-FTS and MIPAS-ESA versions, the MIPAS-IMK data are a new version and employ a new processor approach, too. We say this explicitly in line 96 and specify which version we use in line 118. In the revised paper, we will make clearer which version was used in previous papers, and which version we use here.
This seems a rather 'mechanical' paper - mostly just reproducing earlier results, the only new aspect being the updated datasets. Indeed, the sort of paper one can imagine being generated, in a few years if not already, by some of the more advanced AI machines.
It might have been acceptable if the original authors wanted to update their paper with new results, in which case I would expect a narrative focusing on the algorithm changes, and the expected and observed impacts on the intercomparisons with respect to their earlier results rather than, as here, analysing the results in absolute terms as if they were being presented for the first time. But if it's a paper with new lead authors it also needs some significant new insight or analysis. I also have some doubts about the methodology.
It is critical to intercompare and validate each new satellite dataset as it is produced. As we have new data versions and methodologies, the data users need to understand changes and updates to these H2O and HDO data products. This validation work should not only fall to the data producers but should be taken up by other members of the community.
In case of MIPAS-IMK (versus ACE-FTS), we refer several times to previous results (e.g., line 315/316; line 335-337; line 373-377; 436-438; 458-461; 492/493; 501-503; 513-515). We will make clearer in the revised version what changes in the retrieval set-up were made and will discuss the expected versus observed changes between the data versions.
Furthermore, we would like to point out that it was by intention that we followed the methodology of Högberg et al. so closely. By using the same methods, we ensure that the new results of this paper are directly comparable to the results presented in Högberg.
General Comments and Suggestions
1) Colocations
The comparison has been performed on the two versions of MIPAS data as if these were independent satellites. A more satisfactory approach would have been, first, to apply the respective filters to these MIPAS datasets and take just the common profiles, and then compare this with ACE-FTS. This would not only ensure no time/space mismatch between the two MIPAS datasets but also ensure exactly the same time/space mismatch between ACE-FTS and either of the MIPAS datasets.
On this topic, Fig 1 looks very odd. It seems most unlikely that in a 3 day period the best ACE-IMK coincidences are in a different latitude to the best ACE-ESA coincidences - I would expect them to be mostly the same locations. The averaging of all MIPAS profiles within the coincidence criteria also seems undesirable. The averaging will reduce the random noise in the MIPAS profiles so the contribution to the overall SD is no longer straightforward. Better to take just the closest profile. Also, noting the difference in time/space would allow subsequent analysis as to whether the chosen margins are adequate or, more ambitiously, allow the colocation error to be quantified eg when switching to grid boxes or zonal averages.
This was addressed in the response to reviewer G. Toon as follows. “In the revised version of the manuscript, when multiple MIPAS profiles are spatially coincident with an ACE-FTS profile, the MIPAS profile closest in time is selected. In addition, there must be both MIPAS IMK and MIPAS ESA processed data available for this coincident profile.”
2) Algorithm descriptions
In 2.1.1/2.1.2 the descriptions of the two retrievals seem to be taken directly from the source papers using their own terminology (possibly via the SPARC papers), with little effort to standardise the information let alone provide some interpretation in terms of possible impact on the comparisons.
For example, 'non-linear least-squares global-fitting technique with Tikhonov regularisation' (for IMK) and 'least-squares global fitting, using the Gauss-Newton approach modified with the Levenberg Marquardt method' plus 'a posteriori regularisation' (for ESA). The reader has to work quite hard to understand whether or not these are essentially the same thing and hence whether significant differences may arise from these aspects of the algorithm.
Even on a more basic level:
- Does ESA-MIPAS retrieve H2O and HDO as log(VMR) or VMR (which affects how you should evalulate biases)?
- Does the IMK-MIPAS account for horizontal inhomogeneities in the line-of-sight direction and/or assume LTE?
- Do both use microwindows in the same spectral region?
Similarly with ACE-FTS, orbital altitude and inclination are measured but these are not given for Envisat. Spectral bands are given for ACE-FTS but not for MIPAS.
We will improve the descriptions of the algorithms and retrieval set-ups and make them more consistent to each other. In each case, we will also describe the changes applied since the previous data version.
3) Retrieval diagnostics
There is no reference here to the retrieval characteristics such as expected random noise, systematic errors or averaging kernels, at least typical values - these are not likely to change much except (for MIPAS) towards the poles where the atmospheric temperature is low.The SD of the bias, for example, could be put in the context of the retrieval random errors, and the bias itself in terms of the overall systematic errors. Meanwhile the averaging kernel describes the ability of the retrieval to follow 'real' atmospheric variations, which has an impact on the correlations as well as the SD of the comparisons. The lack of an HDO time signature in the ESA data might be explained in terms of the averaging kernel.
We will add all available information to the revised version.
4) CH4 consistency
Given the large discrepancy between ACE and the two MIPAS profiles in the stratosphere a simply self-consistency check would be to see how these compare with their equivalent CH4 retrievals, on the basis that (H2O + 2CH4) should be conserved.While this is an interesting idea, it is beyond the scope of this paper and will be considered by the authors for future work.
5) The authors should be aware of the difference in LaTeX between a minus sign ($-$) and a dash (--) indicating a range of numbers. Where both positive and negative values are possible it would also help if '+' were added in front of the positive numbers to further distinguish them from the --. Thus, in L24: $-$8.6--+10.6 Incidentally, assuming this is taken from Table 2, the actual number in the table is -8.7. Further on this particular point, negative and positive biases don't have any particular meaning in this sentence, so I would just have said 'biases of up to 10%'.
Thank you very much for this suggestion. The changes that the referee proposes will be considered. By one hand and according to the AMT style, we will use the word “to” for indicating a range and en dashes (–) for numerical purposes. By the other hand, the biases will be referred in total percentage, when appropriate, along the text in the revised version of the manuscript.
6) There are numerous references to 'global' averages whereas I would suggest 'dataset' averages, or something similar, unless you really claim that your intercomparison dataset represents some sort of uniform sampling of the globe.
The term “global” will be clarified in the revised manuscript.
7) It is unclear whether or not the MIPAS-IMK product has been updated (these authors refer to 'V5H' and 'V5R' whereas Hogberg (2019) referred to V20 as being the newer version.
See our comment above. As a result of the analyses by Hoegberg et al., 2019 and Lossow et al., 2020, a new retrieval approach for HDO has been developed for MIPAS-IMK, which we present here. We will clarify which different versions have been used for the H2O, HDO and derived delta-D results for all instruments.
8) 'Standard Deviation' is already defined as the spread around some mean value, so 'debiased standard deviation' is just 'standard deviation' since we are already talking about mean differences (or 'the standard deviation of the relative bias' which is how it is described in the caption for Fig.4). Perhaps you thought SD might be confused with root-mean-square-difference? Also in Tables 1 and 2 this has become '1 \sigma Bias' which sounds like a different thing altogether, but I assume it's the same.
We will make sure to use the same technical terms throughout the paper. We have used the term “de-biased standard deviation” to make clear that it is the standard deviation around the mean difference (spread around the bias) and not the spread of the data sets themselves. What is called “1-sigma bias” in Tables 1 and 2 is meant to be the de-biased standard deviation. This will be updated, too.
9) The time series plots could be enhanced by subtracting out the mean profile and then also removing the average annual cycle to show interannual variability. The latter may have some QBO correlation which could be investigated.
Again, we are intrigued by this suggestion for additional analyses. However, they are beyond the focus of this current paper.
10) Details of bias determination (3.1.3, 3.2) are, firstly, confusing because of the repeated use of the same 4 coordinates for each parameter, secondly difficult to read because of the small fonts and, finally, quite standard. Even then, there are a few problems here Eq 1 presents the bias as 4-coordinate average of b_i which are themselves 4-coordinate quantities. However, it seems unlikely that any two measurements are exactly matching in latitude or time (it's not clear what 'period' means here) so I assume these coordinates are what is being averaged over in i=1, n so should not appear on the l.h.s. as well. And b_i is presumably also a function of longitude, also averaged.
We agree with the referee in the confusing use of 4 coordinates for each parameter. In the revised version of the manuscript the notation is simplified using the formalism of Dupuy et al ACP, 9, 287–343 (2009).
sigma_x1 and sigma_x2 in Eq (7) are undefined,
Will be fixed.
$\sigma_b$ has a bar over the b here but not in Fig 4.
Will be fixed.
Fig 4 shows SD as a percentage (of what?) but Eq 5 defines this in absolute terms.
The debiased standard deviation is calculated to the mean relative bias, therefore the unit is percentage. Eq (5) is defined in relative terms, it will be clarified in the revised version of the manuscript.
When considering relative bias, if the two datasets were retrieved as log(VMR), the geometric rather than arithmetic mean of the two values would be more appropriate as the reference value.
Only one of the datasets is retrieved in log(VMR) so the arithmetic mean is applied in all cases.
I appreciate that this sort of thing was included in the previous papers (and had I reviewed those I would have said the same thing) but that's even less reason to include it again here. Everyone knows what you mean (and they can look up Pearson Correlation on Wikipedia) so no need to drag the reader through all the small print. On this (rare) occasion, it really is simpler to explain in words rather than equations.
For clarity on how the correlation calculation is performed, we have chosen to include this equation.
11) You could save a lot of wordage simply referring to these data as 'IMK', 'ESA' and 'ACE' (at least in the text, perhaps not in the figures)
Thank you for the idea. Since these abbreviations are not commonly used in the previous literature, we would prefer to maintain the references to the datasets as they are.
Minor/Typographical comments
Abstract
L139: Use a regular reference for the SPARC special issue to avoid the typesetting difficulties caused by putting the URL (http:...) in the text.
This is the standard method for referring to this special issue.
L185: 'from 1 to 70 km'
The referee is right! The word “since” will be changed by “from”.
L189: Presumably the two MIPAS datasets are automatically colocated, so this just refers to colocating ACE-FTS with MIPAS.
See methodology response above.
Fig 2: I assume the variation in low altitudes is due to cloud-screening but what causes the reduction in MIPAS-ESA comparison data at high altitude?
The reduced number of coincidences for the ESA profiles above 50 km is due to the fact that profiles from different observation modes are used, in particular measurements from the UTLS observation mode, which are about 8% of the total MIPAS observations and are performed mainly in the period August 2004/August 2005, are characterised by the altitude range 8.5 - 52 km.
However, as stated in the first comment, in the revised version of the manuscript we compare the same number of coincident profiles for the three databases. The number of coincidences in the stratosphere is 15263 and it decreases going to the lowest altitudes up to 4078 at 10 km.
L201: It would be useful to have at least an approximate figure as to what percentage of MIPAS profiles fail the quality control tests.
For MIPAS-IMK about 0.16% of all started retrievals of HDO did not converge. About 0.05% of all started retrievals failed or encountered corrupted spectra. In total 2,314,011 profiles were processed. This means that 99.79% of all profiles are considered healthy.
The two flags that we provide with the data (visibility flag needs to be 1, and averaging kernel diagonal needs to be > 0.03) are meant to be applied to single points in the vertical profile, i.e., these flags reduce the altitude coverage of one profile, but leave the other parts of the profile valid. Please note that we always provide the profiles from 0 to 120 km. The flags define the valid altitude range.
MIPAS-ESA uses a different approach for filtering out bad profiles.
The quality of the retrieved profiles is judged “good” when three requirements are met: the retrieved profile adequately reproduces the measurements (i.e. the chi-square value at the final step of the iterative procedure is smaller than a pre-defined mode and species-dependent threshold), there are no outliers in the retrieval error (i.e. the maximum value of the retrieval error profile is smaller than a pre-defined mode and species-dependent threshold) and the iterative retrieval procedure successfully converges.
When at least one of the previous requirements is not verified, the whole retrieved profile is flagged as bad in the output file (post_quality_flag=1) and it is not used as either profile of an interfering species or initial guess in subsequent retrievals. Otherwise, if all previous conditions are verified, the post_quality_flag is set to 0 so that the retrieved profile is considered “good”, and it can be used for subsequent retrievals. If the retrieved temperature is flagged as bad, no VMR retrieval is performed, since a proper temperature profile is fundamental for the retrieval of the trace species.
Each retrieved profile is properly and fully characterised on the full retrieval range provided in the output files by the corresponding CM and AKM. Altitude regions with poor information on the retrieved target can be identified by the low values of the diagonal elements of the AKM and/or the large values of the diagonal elements of the CM. Since the AKM and the CM are calculated considering the retrieval on the full vertical range, even if some of the retrieved values are discarded by the user, we recommend to use the full profile along with its full CM and AKM.
2.54 million HDO profiles are available in the products. The percentage of ESA good profiles is reported in Fig. 6 of Dinelli et al., 2021 paper for all retrieved trace species. 8% of HDO retrieval procedures fails.
Part of this information will be provided in the revised version of the paper.
L208: In L187 the grid is from 1-70km. Assuming the IMK grid is at 1km intervals how do you get 58 levels?
The MIPAS-IMK grid is not strictly a 1-km grid. It is a 1-km grid from 0 to 44 km, followed by a 2-km step width from 46 to 70 km. These are 58 levels. We will correct this description in the revised version.
L304: I would not refer to these as 'error bars', just 'bars'.
Will be changed. Thank you.
L319: Describing these values as 'global' minima is misleading, they're actually the minima in the intercomparison dataset.
The term “global” will be clarified in the revised manuscript.
L325: Similarly, ‘global mean’ profiles.
Also, here.
Fig 3. These plots would be more informative (and take less space) if all three datasets were combined on the same plot, allowing MIPAS-IMK and MIPAS-ESA to be directly compared. Use eg dashed lines to mark the 1sigma variation for each (Also, I wouldn't call these 'error bars' as in the current caption).
We agree with the reviewer suggestion, the three curves will be combined in the same plot. We also improve the figure 3 including the MIPAS-IMK to MIPAS-ESA comparison. Regarding the “error bar”, we think that the referee is right, they are “bars”.
L365: It is perhaps worth mentioning that the low SD (Fig 4(c)) between ESA-ACE and coupled with the high correlation (Fig 4(d)) are consistent with the MIPAS-IMK retrievals being less sensitive to actual atmospheric variations in H2O, and conversely for IMK-ACE for HDO.
We agree with this statement for altitudes above ~25 and ~15 km for H2O and HDO, respectively, and will add it to the revised version.
Fig 4(g) mis-labelled(?) as '\sigma_b Bias' (and similarly Table 1)
No, this notation is correct as it is the de-biased standard deviation.
Table 1 & 2: I assume these figures are a summary of the plots shown in Fig 4, in which case make it explicit in the Table caption (and similarly in the Fig 4 caption refer to these tables). Rather than have rows labelled eg 'MIPAS-IMK vs 'ACE-FTS' I suggest replacing 'vs' with $-$ to make it absolutely clear which way around you are defining the sign of the bias.
These changes will be made in the revised paper.
L393: Is there such a thing as a 1\sigma standard deviation? I though the SD was, by definition, 1 \sigma.
We agree with the referee. Our sentence is redundant. Will be changed.
Fig 5 caption refers to 'meridional cross-sections' which suggests a single slice through a particular longitude. 'Zonal mean' or 'Zonally averaged' distributions are the usual terminology for such plots. And rather than use 'summer' and 'winter' - which differ from north to south hemispheres - give the actual months averaged (as in Fig 6).
These updates will be made.
L405: I'm assuming that the labelling on Fig 5 is correct (I would have expected ACE-FTS to be missing the high latitude measurements during the local winter months but, on the other hand, they may also specifically target as high latitude as possible during polar winter months). However, the very low ACE values for both H2O and HDO in the winter polar vortex (compared to the two MIPAS datasets) are worthy of comment.
Yes, this figure is correct. ACE-FTS is in an orbit that targets high latitude measurements, more than 50% are at latitudes higher than 60 degrees, to investigate polar ozone chemistry. Note, that the local winter mean from ACE-FTS does not include data from all of these months because of the requirement for sunlight for its measurements. This requirement leads to the ACE values sampling only the later part of this season (in austral winter, mainly Aug. in JJA at the highest latitudes) than the two MIPAS datasets sampling all months. It is likely this sampling difference that leads to ACE-FTS showing more dehydration than MIPAS in these zonal mean plots.
Also, why is the IMK cloud filtering any different for HDO than for H2O?
Lossow et al., 2020 have made sensitivity tests regarding the retrieval of HDO from MIPAS data. They found that for the HDO retrieval, the upward error propagation was pronounced at the lower end of the profiles in previous data versions. I.e., incorrectly retrieved vmrs due to, e.g. unidentified cloud contamination trigger retrieval errors in the vmrs above this altitude. To be on the safe side, Lossow et al., 2020 recommended to discard the retrieved values in the altitude range of the lowest two (V5H) to three (V5R) tangent altitudes. They found that the propagated error fades out sufficiently above this level so that data from levels above can be used.
L450: 'interannual variability' implies what's left after removing the annual cycle. From Figs 7a,b,c alone it is not clear to me that ACE-IMK have the greatest similarity.
The referee is right, we are not analyzing the “interannual variability” (seasonal cycle subtracted from the data sets) but the time series themselves. This sentence will be changed.
The referee is also right again, while the tape-recording effect is clearly seen in the MIPAS-IMK HDO time series, this is less evident in both MIPAS-ESA and ACE time series.
L465: 'shows' rather than 'shown'.
Will be changed.
L467: 'while the ACE-FTS instrument can measure with higher sensitivity ...' What is the basis for this statement?
ACE-FTS has a higher sensitivity due to the combination of the long-path length through the atmosphere in limb-view and the solar occultation measurement technique used. This makes the ACE-FTS measurements less susceptible to perturbations due to thin cirrus clouds in the UTLS.
L485: 'vertical behavior' (or 'behaviour') presumably just refers to the small mean profile bias averaged over the whole dataset, shown by the red-lines in Fig 4a,b. 'Behaviour' suggests that the two track each other well in other respects such as low SD and high correlation, which they don't (at least not above 20km).
This sentence will be clarified in the revised manuscript.
L487: 'according to the uncertainties' Which uncertainties are these? There was no reference to estimated uncertainties in the datasets in the main part of the paper.
Given that the two MIPAS profiles are 1ppmv higher than ACE after averaging over several thousand profiles suggests rather a significant discrepancy to me.
As mentioned above, we will provide all available information on uncertainties in the revised paper.
L495: 'quantitaty'?
Will be fixed.
L525: 'MIPAS-IMK dataset provides a more realistic signal for the entire stratosphere' This is a bold statement and needs some qualification otherwise it could get quoted out of context. Is this for H2O, HDO, delta D or all three? Is it in terms of the time-evolution or mean profile? Is it because it has the best correlation or lowest SD compared to both other datasets?
“More realistic than MIPAS-ESA” was to be meant and below the 30 km of altitude. However, we agree with the reviewer and even with all the context the statement can be a bold given the existing data. We reword this paragraph in the revised version of the manuscript.
-
AC4: 'Reply on RC2', Paulina Ordóñez, 07 Nov 2023
-
EC1: 'Comment on egusphere-2023-1348', Justus Notholt, 11 Oct 2023
Dear authors/coauthors,
when looking at both reviews I would suggest to withdraw the manuscript. Both reviews argue that the manuscript is not acceptablen for publication, and that it will be very difficult or not be possible to modify it.
Regards
Justus NotholtCitation: https://doi.org/10.5194/egusphere-2023-1348-EC1 -
AC1: 'Reply on EC1', Paulina Ordóñez, 14 Oct 2023
Dear Prof. Justus Notholt,We are writing to you to reconsider the suggestion of withdrawing the manuscript. Reviewers #1 and #2 recommend major revisions. Both reviewers are willing to review the revision of the paper. We are working to satisfactorily address the comments. It is not easy, but neither is it impossible.
Therefore, we attentively request the opportunity to send back the manuscript to the reviewers for a re-review and determine whether the comments have been satisfactorily addressed.
Sincerely yours,
Paulina.Citation: https://doi.org/10.5194/egusphere-2023-1348-AC1 -
EC2: 'Reply on AC1', Justus Notholt, 17 Oct 2023
Dear Paulina,
I contacted both reviewers, you are welcome to submit a modified manuscript.
Regards
JustusCitation: https://doi.org/10.5194/egusphere-2023-1348-EC2
-
EC2: 'Reply on AC1', Justus Notholt, 17 Oct 2023
-
AC1: 'Reply on EC1', Paulina Ordóñez, 14 Oct 2023