Articles | Volume 16, issue 11
https://doi.org/10.5194/amt-16-2733-2023
© Author(s) 2023. This work is distributed under the Creative Commons Attribution 4.0 License.
Applying machine learning to improve the near-real-time products of the Aura Microwave Limb Sounder
Download
- Final revised paper (published on 02 Jun 2023)
- Preprint (discussion started on 27 Jan 2023)
Interactive discussion
Status: closed
Comment types: AC – author | RC – referee | CC – community | EC – editor | CEC – chief editor
| : Report abuse
-
RC1: 'Comment on egusphere-2023-101', Anonymous Referee #1, 16 Feb 2023
- AC1: 'Reply on RC1', Frank Werner, 16 Apr 2023
-
RC2: 'Comment on egusphere-2023-101', Anonymous Referee #2, 20 Feb 2023
- AC2: 'Reply on RC2', Frank Werner, 16 Apr 2023
-
RC3: 'Comment on egusphere-2023-101', Anonymous Referee #3, 28 Feb 2023
- AC3: 'Reply on RC3', Frank Werner, 16 Apr 2023
Peer review completion
AR – Author's response | RR – Referee report | ED – Editor decision | EF – Editorial file upload
AR by Frank Werner on behalf of the Authors (16 Apr 2023)
Author's response
Author's tracked changes
Manuscript
ED: Referee Nomination & Report Request started (20 Apr 2023) by Jian Xu
RR by Anonymous Referee #1 (21 Apr 2023)
RR by Anonymous Referee #2 (23 Apr 2023)
ED: Publish as is (23 Apr 2023) by Jian Xu
AR by Frank Werner on behalf of the Authors (08 May 2023)
Author's response
Manuscript
The authors present a near real-time processor of Aura/MLS observations using a supervised neural network. The manuscript is easy to follow and shows that the processor has very good performance, very close to the operational processor. The new method presents a significant improvement compared to the previous near real-time processor based on a simplified optimal estimation method. I recommend the manuscript for publication, but I have minor comments that could be clarified by the authors.
General comments
1) I am impressed by the results overall, and more particularly with the ability of the model to capture the increase in H2O induced by the volcanic eruption, though the statistical weight of such events in the training dataset should be low. This illustrates the high potential of the model to capture special disturbances that occur over a restricted spatio-temporal range. However, I found that such abnormal conditions are not sufficiently discussed in the manuscript. Indeed, these are scientifically the most interesting cases but have a low impact on the overall statistical evaluation. For example, in Figure 4b, the increase in H2O at 100 hPa over India and part of Southeast Asia is clearly underestimated with ANN-NRT. This should be discussed in the manuscript and the authors should mention if they have found other cases where significant discrepancies were seen.
2) More generally, the authors do not show results for the whole test dataset (5% of 17 years corresponds to almost 1 year), in particular winter time which is strongly disturbed in the northern hemisphere. Is there a seasonal pattern in the results? Authors should clarify why the test data are well suited for describing the capability of the model and the limitations of such a choice (that could further be investigated in future studies). For instance, I would personally have used 2 entire years with very different conditions (e.g., SSW strength or QBO phase) to test the models.
3) Regarding the vertical resolution of profiles predicted with ANN-NRT. This issue is not addressed in the manuscript and could be clarified. If I understand the NN setting correctly, the vertical resolution of the predicted profile is the same as that of the level 2 operational product (here I am referring to the resolution derived from the operational averaging kernels and not the retrieval levels spacing). Am I right? This could be clarified.
For low SNR cases, the authors mentioned that the NN tends to smooth the noise compared to the operational product. Is this effect could be related to a degradation of the vertical resolution similar to the regularization effect in the OE method?
Specific comments
Line 87: “n” is already used to define the number of input features. It would be clearer if another letter is used for the number of neurons per hidden layer.
Line 93: is the levels of the predicted profile the same as the number of levels of the operational product?
Table 1: I understand that the hyperparameters are defined by a set of tests but the differences between the models could be discussed. Why the number of hidden neurons is much smaller for the H2O model than for T and O3? Why is the tanh activation preferred over Relu for some species? (It is considered that Relu make the training more efficient)
Line188/Table 3: Are the scores calculated for the same periods as Figure 3?
Line207: “Here the ANN … , and the results are close to L2 data”: there is a clear underestimation of the H2O vmr over india and East-Asia. This issue could be mentioned and what could be the reason?
Line 219/Line 241: Would it be possible to complete a small training dataset with simulated data?
Line 244: I don’t understand the sentence “The previous version…”. Do the authors mean: The previous version of MLS NRT data products (OE-NRT, Lambert et al., 2022) is replaced with predictions from an artificial neural network (ANN).