the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
Optimization of Aeolus' aerosol optical properties by maximumlikelihood estimation
Frithjof Ehlers
Thomas Flament
Alain Dabas
Dimitri Trapon
Adrien Lacour
Holger Baars
Anne Grete StraumeLindner
Download
 Final revised paper (published on 11 Jan 2022)
 Preprint (discussion started on 06 Aug 2021)
Interactive discussion
Status: closed

RC1: 'Comment on amt2021212', Anonymous Referee #1, 21 Aug 2021
Review of “Optimization of Aeolus Optical Properties Products by MaximumLikelihood Estimation”
This paper outlines the application of maximum likelihood estimation to the analysis of profiles observed by the high spectral resolution lidar mounted on ESA’s Aeolus satellite. To facilitate a bounded optimisation, the lidar equations are cast in terms of layer optical depths and lidar ratios while measurement uncertainties are approximated by the variance of the observations. The method is compared to a more traditional lidar analysis techniques for both real and simulated observations and is shown to produce smoother, more physically consistent fields that are more consistent with other measurements and the truth, respectively. This work, inspired by unexpected limitations in the instrument after launch, hopefully represents the beginning of more routine use of modern retrieval techniques in the analysis of lidar profiles.
I recommend this paper for publication after minor corrections and thank the authors for providing a genuinely enjoyable read on a rainy morning. My only significant comments relate to the discussion of iterations from lines 278 to 292:

In Eqn. 15 you say L_p > 0 but on L281 you say you start iteration at L_p=0, which is then outofbounds. Presumably you meant to say L_p >= 0. Regardless, I am surprised you chose to start at one end of your solution space. I would have started at some climatological mean value to keep the number of iterations down by giving the method the ability to increase or decrease L_p in the first step.

Forty thousand is a preposterous number of iterations! Are you sure L288 shouldn’t say 40? Every optimization routine I’ve worked with tends to converge in 5 to 10 iterations unless the model is extremely nonlinear. If you need more than several dozen iterations to get useable solutions, I would guess you’ve done something wrong – either a gradient is being miscalculated or an inappropriate optimization method is being used.

I don’t agree with you at L291 that artificially extending the number of iterations produces a fairer comparison to the SCA. This sounds like an attempt to avoid discussing the method of identifying convergence because those are contentious. I honestly don’t care how you do it, but I do think the paper should explain how you intend to produce data outside of the context of this validation and, ideally, give some idea of the magnitude of difference. I suspect it makes very little difference, which is why you ran a set number of iterations, but that should be stated.

Something similar comes up again at L357 when you say the first guess “contaminates” your mean. A first guess is not a prior; it should not affect your final solution. Exceptions are where the method fails (and so you throw away the solution) or there are multiple minima (which would require a more advanced minimisation routine). Could you explain what you mean here? Is it just that you aren’t doing any quality control and so failed retrievals are present in the output? If so, you should do quality control! You don’t need a value from every single pixel.
The following comments are mostly matters of personal curiosity rather than issues that need to be addressed:

The text in figures should aim to be approximately the same size as the text in its caption. If it is possible, all the figures in this paper would benefit from being regenerated with a smaller page size, such that the font is larger relative to the image.

L68) I would say backscatter is ‘measured’ rather than ‘known’ with higher precision.

L77) Is there a reason you preferred to constrain L to be positive rather than retrieve its logarithm? I could see this being useful in your proposed future work to produce consistent regularization in variables that span several orders of magnitude. Further, aerosol optical depth (which is what you actually retrieve) is known to be lognormally distributed, such that the log of L is a more natural basis in which to define a regularization.

L103) I feel like this paper demonstrates that such a retrieval is possible rather than it being something you need to assert in advance.

L204) I’ve not encountered this use of a backslash before. Did you mean to say “a ratio”?

L230) While true, lidar signals are often far from this limit.

L261) I think that the two uncertainty ranges shown in Fig. 5 come from the two sources of uncertainty you mention here – a simple Poisson assumption and the variance of the downlinked profiles – but it would be useful for that to be stated clearly somewhere. This sentence currently implies the Poisson approximation is not used in the remainder of the paper, but you go on to mention it several times.

The first paragraph of Section 4 unnecessarily repeats the preceding paragraph.

I would mention the existence of Appendix D at the end of L343 as that’s when I asked “what’s the RMS deviation”?

In the caption for Fig. 4, don’t you mean the west coast? Also, the description of 4(b) implies that three versions of the feature mask are shown. I think you meant to say that the rows show backscatter, extinction and lidar ratio, with identical features masks shown in each frame of the rightmost column.

L418) I’d argue that the image shows the advantage of forcing L>0 rather than robustness.

L493) Is there any intention to release this data? Any possibility of funding to process the full record or become an operational product?

Appendix A largely repeats Section 3.1.

L516) ECMWF data would typically be interpolated rather than averaged as the values represent behaviour at a point rather than a grid cell average. What do you mean by “mean” here?
I found a number of grammatical corrections:

L2) the Atmospheric Laser

L2) is an Ultra Violet

L5) Being an HSRL

L10) demonstrate a predominantly

L11) information by the SCA

L15) due to effective noise

L27) respectively, and the CloudAerosol

L32) addressing high uncertainties in climate change modelling due to the indirect

L46) [commas around ‘and similarly Raman lidar observations’]

L48) increase the SNR

L68) extinction

L90) Either “revolves around”, “orbits”, or just “Aeolus is in a Sunsynchronous polar orbit”

L90) seven day repeat

L92) Earth with an offnadir

L93) Traditionally, “Earth” is the planet and “earth” is dirt so I don’t see why you use both versions here.

L93) There should be a space between numbers and their units.

L125) [remove the comma]

L135) scales dependent on

Fig 1) accumulation to observation

L164) [comma after ‘following’]

L172) into a molecular

L178) [commas around ‘in principle’]

L181) can be simplified by introducing the range

L209) Appendix A.

L281) consist of an aerosol free

L296) y_obs are generated

L302 and 563) MoorePenrose

L312) the real instrument cannot

L334) particularly

L334) by an additional

L336) simulation are shown

L350) but at the

L381) produces very

L409) cloud

L428) the lidar ratio along the plume’s horizontal

L439) The ground based

L450) aerosol

L457) be located by using extinction coefficients alone due to the fine range

L484) additional

L487) [delete comma]

L514) channels

Eq A7) [something has gone wrong with the spacing]

L541) e.g.

L559) is inverted by

L561) i.e.
I largely appreciated the relatively conversational tone of this manuscript compared to the average paper, making relatively dry material easier to read. But there are a few places where I would have made a different choice of words. As these are a matter of style, I leave it to the authors to decide if they agree with me or not.

I don’t think “Products” is necessary in the title of the paper. Then again, I would have called it “Optimization of Aerosol Optical Properties from Aeolus Profiles by MaximumLikelihood Estimation” because I’m concerned that someone might think the “Optical Properties” are possessed by Aeolus rather than by particles in the atmosphere.

Remove “(partly)” from L8. A problem is either ill posed or it isn’t; there is no partly.

L52) A white paper under preparation in the UK is hoping to call these “representation errors” rather than worry about ‘representivity’ vs ‘representativeness’. Obviously there’s no obligation to agree with us but I think it’s a cleaner phrase.

L122) “corrected for to obtain” doesn’t look strictly wrong but sounds very strange to a native near; I wouldn’t say the ‘for’.

Fig 1) While this image is technically “Exemplary” in that you are using it as an example, English has corrupted the word to typically mean “Outstanding”. I’d say “Illustrative” or “Example of” instead.

Eq 23) I know you’re using T for transmission but it hurts slightly to see temperature denoted t. I would have used \mathcal{T} for transmission.

L179) Maybe “extensive” rather than “excessive” as the latter has negative connotations.

L312) Perhaps “noise level resembles nominal” as ‘resembles’ already implies an inexact comparison.

L370) “...reacts after the backscatter coefficient, resulting in the attenuation of the first 500 m of the cloud being captured incorrectly with consequences...” I struggled to understand this sentence and have made my best guess at what you were trying to say.

L373) “Hence, the same feedback loop is triggered, as described in case I for changes of range bin thickness.” I find it difficult to understand long sentences where the verb is at the end, so I moved the verb to the start.

L383) “MLE causes extinction” as ‘locks’ is a rather strong word for what happens.

Fig 3) This caption could just say “As Figure 2 for Case II”.

Fig 4) Is it necessary to use very similar colours for the two lines? If you moved away from a rainbow colour bar, you would have more options so I don’t have to strain to see if the line has little dots in it or not.

L463) Perhaps “refactored” rather than “rephrased”? The latter usually refers to words rather than actions.

L479) Considering you go on to critique the use of a feature mask, you might want to say “it is possible to” rather than “it would be advantageous to”.

Eq C2) I have never seen the notation K^{T} to indicate an inverse transpose. Probably because it’s ambiguous if I transpose the inverse or inverse the transpose. (K^T S_Y^{1} K)^{1} is a more traditional way to write this equation.

I was taught to put a hyphen between words that together form a single adjective. I provide a nonexhaustive list of places where you could do that: L33 & 439) groundbased; L90) sevendays; L95) RayleighBrillouin; L101) doubleedge; L123) timegated; L137 & 140) socalled; 281) aerosolfree; L374) noiseinduced; L393) polarizationsensitive, dualwavelength; L541) particlefree.
Citation: https://doi.org/10.5194/amt2021212RC1 
AC1: 'Reply on RC1', Frithjof Ehlers, 01 Oct 2021
General comment:
First thing to note is that we were very glad about the thorough feedback of all Referees, who evidently made the effort to go through the details presented in our manuscript. This helped a lot to refine the text and we hope that we could make our statements more concise now. Especially the method of calculating the uncertainties in the MLE method was addressed by more than one of the Referees and required some changes that we want to present to all of Referees.
The procedure outlined in the last paragraph of Appendix C was relying on the simple equation alpha = gamma * beta with extinction coefficient alpha, lidar ratio gamma and backscatter coefficient beta. Motivated by comparing the obtained variances via (C2) with the variances one would get from a standard error propagation with constant lidar ratio sigma_a^2 = gamma^2 * sigma_b^2, we argued e.g. that the uncertainty in alpha cannot exceed 200 sr times the uncertainty of beta in a single bin, whatever the precise value of gamma is in reality, due to the upper bound in lidar ratio.
Now, after the feedback and some graphical considerations, the problem with this statement is, that it leaves aside any uncertainty or spread of the lidar ratio estimate itself, as it can lie somewhere in between 2 sr and 200 sr and hence should have nonvanishing variance sigma_g^2, causing another term to arise from the view of standard error propagation, increasing the uncertainty of alpha proportional to beta itself. This is why we noticed that the presented approximation was not correct, although we observed uncertainty values that matched approximately the SCA mid bin output for the SAL case for backscatter and extinction coefficients, which seemed realistic at first glance.
Following this, we evaluated the errors from equation (C2) directly against the uncertainties calculated by the SCA standard error propagation as in Flamant et al., 2020, and found good agreement, when the same Poisson noise variances are used in the S_y in (C2), see attached figure.
That means, that with (C2) we are only able to provide uncertainties for the unconstrained problem (as SCA solves it), but especially the soobtained extinction uncertainties are overestimated and not representative of the real uncertainties, as one can see in the comparison of SCA and MLE in the simulation cases.
In lack of a representative alternative for uncertainty estimation we decided to delete the lidar ratio and extinction uncertainties of the MLE solutions in Figures 5 and 6, for now, stating that there is no representative error estimate available yet. The challenge of reporting representative uncertainty estimates is now being stressed in the abstract and conclusion and will require future investigation.
Regarding the comments of Referee 1:
The effort you made to provide such detailed feedback on our manuscript was greatly appreciated. Your greatest concern regards the convergence of the algorithm, on which we elaborate below in the comments and which led us to refine some statements in the manuscript as well. In short: We are currently trying to tune some of the parameters and to change the scaling of the input vector for the future so that the convergence speed is optimised. As this regards a technicality of the processing, rather than the final results (since the final cost function is reasonably low in all cases), we are confident to publish the obtained data.
Below you will find all the individual answers to your comments. All grammar and spelling corrections were taken into account directly, so we dropped them below to focus mostly on the comments regarding formulation and content.
 In Eqn. 15 you say L_p > 0 but on L281 you say you start iteration at L_p=0, which is then outofbounds. Presumably you meant to say L_p >= 0. Regardless, I am surprised you chose to start at one end of your solution space. I would have started at some climatological mean value to keep the number of iterations down by giving the method the ability to increase or decrease L_p in the first step.
This is correct, we meant to state L_p >= 0. It has been corrected. It is also probably not the best choice to initialise on the boundary of the solution space, but given that the climatological value should be rather small, it does presumably not make a big difference.
 Forty thousand is a preposterous number of iterations! Are you sure L288 shouldn’t say 40? Every optimization routine I’ve worked with tends to converge in 5 to 10 iterations unless the model is extremely nonlinear. If you need more than several dozen iterations to get useable solutions, I would guess you’ve done something wrong – either a gradient is being miscalculated or an inappropriate optimization method is being used.
When just one profile is considered, as in problem (14), then the number of iterations is on the order of 50 (roughly 0.5 seconds on an office laptop) to achieve convergence (an average cost function value per bin below 1). In practice, however, we are interested in solving problem (15) for all profiles simultaneously. In this case about 5000 iterations are required (roughly 45 seconds on an office laptop) to let the cost function decrease in the same way for a whole orbit (roughly 450 profiles). This information has been added to the manuscript.
So essentially, solving (15) for the whole orbit saves us time. Though, on second thought there must be a better converging formulation, but we have had no reference on this. We made an attempt to reformulate our model in terms of log(L_p), but the convergence speed did not increase noticeably. It is very likely that a more advantageous scaling exists since the LBFGSB algorithm is not scale invariant. Unfortunately we haven’t found it and are currently tuning parameters to enhance performance.
 I don’t agree with you at L291 that artificially extending the number of iterations produces a fairer comparison to the SCA. This sounds like an attempt to avoid discussing the method of identifying convergence because those are contentious. I honestly don’t care how you do it, but I do think the paper should explain how you intend to produce data outside of the context of this validation and, ideally, give some idea of the magnitude of difference. I suspect it makes very little difference, which is why you ran a set number of iterations, but that should be stated.
You are right that for an operational use, the presented approach is not handy. So we decided to add at the end of the paragraph: “For operational use we plan to tune the number of iterations in an adhocfashion, based on if the average cost function value per bin has fallen below a value of 1.”
In contrast to OEM and regularized MLE, the only contribution to the MLE cost function is how well the raw signals y are represented by F(x*). Once they are within the estimated limits, we can essentially stop iterating, because we cannot expect to gain any additional information content from the measurements. So this will determine convergence.
 Something similar comes up again at L357 when you say the first guess “contaminates” your mean. A first guess is not a prior; it should not affect your final solution. Exceptions are where the method fails (and so you throw away the solution) or there are multiple minima (which would require a more advanced minimisation routine). Could you explain what you mean here? Is it just that you aren’t doing any quality control and so failed retrievals are present in the output? If so, you should do quality control! You don’t need a value from every single pixel.
So the problem is the following: Whenever the retrieved aerosol optical depth L_p is very low, the influence of the lidar ratio estimate on the cost function becomes increasingly insignificant, until L_p becomes zero and no lidar ratio can be provided at all (these are not included in the statistics). Wherever L_p is very low, the lidar ratio has a high error estimate and some may remain close to the first guess, which then seems to act as an ‘implicit a priori’ (see https://doiorg.tudelft.idm.oclc.org/10.1029/JD095iD05p05587, end of section 7). What our paragraph wants to say is that, if lidar ratios were averaged without weights, then the first guess would contaminate the statistics. In the shown simulation case, this would only affect the statistics when only bins with retrieved backscatter coefficients below 5*10^(2) (Mm sr)^(1) are considered.
So eventually we propose to replace the sentence “Otherwise, the first guess of $\gamma_{,p}=60$ sr would contaminate the statistics for MLE, e.g., $\text{mean}(\alpha_p/\beta_{,p}) $ would be biased towards the first guess.” with “[...], in order to disregard the influence of bins with nearly vanishing aerosol optical depth, for which no reliable lidar ratio can be retrieved.”
Additionally, lines 360361 and lines 380381 are dropped, because they are more confusing than helpful.
 The text in figures should aim to be approximately the same size as the text in its caption. If it is possible, all the figures in this paper would benefit from being regenerated with a smaller page size, such that the font is larger relative to the image.
The font size of figures 2,3,4 and D1, D2 has been adjusted.
 L68) I would say backscatter is ‘measured’ rather than ‘known’ with higher precision.
corrected
 L77) Is there a reason you preferred to constrain L to be positive rather than retrieve its logarithm? I could see this being useful in your proposed future work to produce consistent regularization in variables that span several orders of magnitude. Further, aerosol optical depth (which is what you actually retrieve) is known to be lognormally distributed, such that the log of L is a more natural basis in which to define a regularization.
There is actually no elaborate reason for the choice of L instead of log(L), but it might contribute to proper scaling of the variables, indeed. We will investigate this aspect in future investigation, also regarding the convergence speed.
 L103) I feel like this paper demonstrates that such a retrieval is possible rather than it being something you need to assert in advance.
The possibilities were discussed prelaunch and the SCA has been implemented already before. The paper’s scope is improving the precision of the retrieval.
 L204) I’ve not encountered this use of a backslash before. Did you mean to say “a ratio”?
No, we meant “or”. It has now been dropped for convenience.
 L230) While true, lidar signals are often far from this limit.
Considering also the comment of Reviewer 2, we changed the sentence to
“[...] because the discrete Poisson noise distribution can already be well fitted by a smooth Gaussian with identical mean and variance for very low (photon) counts and the aforementioned additional noise sources and their corrections, e.g., subtraction of measured solar background, will naturally smear out the discrete nature of the Poisson noise.”
 L261) I think that the two uncertainty ranges shown in Fig. 5 come from the two sources of uncertainty you mention here – a simple Poisson assumption and the variance of the downlinked profiles – but it would be useful for that to be stated clearly somewhere. This sentence currently implies the Poisson approximation is not used in the remainder of the paper, but you go on to mention it several times.
Thanks for underlining this source of incovenience. The Poisson approximations for MLE in Fig 5 and 6 have only been used for the uncertainty estimation and not in the minimization. Following the feedback of Reviewer 2 and 3, we noticed that our implementation of the boxconstraints in the error calculations was flawed. In the lack of an alternative we removed what we initially claimed to be an upper bound in Figures 5 and 6.
We also edited the last paragraph of 3.2 to explain that the uncertainties of SCA and SCA MB are calculated as in section 6 of the Algorithm Theoretical Baseline Document for the L2A and changed the name of this uncertainty to “Poisson error estimate” in Figures 5 and 6.
 The first paragraph of Section 4 unnecessarily repeats the preceding paragraph.
The paragraph 3.3 has now been deleted.
 I would mention the existence of Appendix D at the end of L343 as that’s when I asked “what’s the RMS deviation”?
According to the feedback of Reviewer 3 we promoted a modified version of figure D1 (with zoomed in boxes for better visibility below 2 km altitudes) to the main text and added a more detailed discussion of the biases and relative errors in the regime of most interest below 2 km altitude.
 In the caption for Fig. 4, don’t you mean the west coast? Also, the description of 4(b) implies that three versions of the feature mask are shown. I think you meant to say that the rows show backscatter, extinction and lidar ratio, with identical features masks shown in each frame of the rightmost column.
We corrected for both errors.
 L418) I’d argue that the image shows the advantage of forcing L>0 rather than robustness.
This sentence lacked a concise message indeed, so we changed it to say “[...] which demonstrates the advantage of the boxconstraints.”
 L493) Is there any intention to release this data? Any possibility of funding to process the full record or become an operational product?
The MLE is being considered for operational implementation.
 Appendix A largely repeats Section 3.1.
This was considered to be more reader friendly, since section 3.1 presents the bare minimum of derivations.
 L516) ECMWF data would typically be interpolated rather than averaged as the values represent behaviour at a point rather than a grid cell average. What do you mean by “mean” here?
We write “by means of” in the sence of “through” or “with the help of” and do not refer to the statistical mean.
 L32) addressing high uncertainties in climate change modelling due to the indirect
We changed this sentence to “[...] for the largest single cause of uncertainty in antropogenic radiative forcing has been reported to be from the indirect effect of aerosols on clouds (Illingworth ...)”
 L93) Traditionally, “Earth” is the planet and “earth” is dirt so I don’t see why you use both versions here.
Although amusing in its current form, we corrected the noncapital “earth”s.
 I don’t think “Products” is necessary in the title of the paper. Then again, I would have called it “Optimization of Aerosol Optical Properties from Aeolus Profiles by MaximumLikelihood Estimation” because I’m concerned that someone might think the “Optical Properties” are possessed by Aeolus rather than by particles in the atmosphere.
This is a very good point, also considering that the Journal’s scope can be on both measurement systems and the data analysis. So we agree to change the title.
 Remove “(partly)” from L8. A problem is either ill posed or it isn’t; there is no partly.
We removed illposed as what we mean is “sensitive to or ambiguous under noise influence”.
 L52) A white paper under preparation in the UK is hoping to call these “representation errors” rather than worry about ‘representivity’ vs ‘representativeness’. Obviously there’s no obligation to agree with us but I think it’s a cleaner phrase.
Fair enough, we follow this advice since the term “representativeness” seems to bulky anyhow.
 Fig 1) While this image is technically “Exemplary” in that you are using it as an example, English has corrupted the word to typically mean “Outstanding”. I’d say “Illustrative” or “Example of” instead.
Thanks! We did indeed not mean to exaggeratedly underline the presented data.
 Eq 23) I know you’re using T for transmission but it hurts slightly to see temperature denoted t. I would have used \mathcal{T} for transmission.
I agree but this way the difference is clearer and comes without unnecessary flourish. If we had used temperature frequently, we would have changed it for sure.
 L370) “...reacts after the backscatter coefficient, resulting in the attenuation of the first 500 m of the cloud being captured incorrectly with consequences...” I struggled to understand this sentence and have made my best guess at what you were trying to say.
We changed it to “the SCA extinction coefficient reacts delayed compared to the backscatter coefficient, leading to that the attenuation by the first 500m of the cloud is not captured correctly”.
 Fig 4) Is it necessary to use very similar colours for the two lines? If you moved away from a rainbow colour bar, you would have more options so I don’t have to strain to see if the line has little dots in it or not.
Our excuses, but this data has been plotted through online services. I completely agree that rainbow color bars are no optimal choice.
 L463) Perhaps “refactored” rather than “rephrased”? The latter usually refers to words rather than actions.
We chose “reformulated” instead.
 L479) Considering you go on to critique the use of a feature mask, you might want to say “it is possible to” rather than “it would be advantageous to”.
Thanks for making this statement more consistent.
 Eq C2) I have never seen the notation K^{T} to indicate an inverse transpose. Probably because it’s ambiguous if I transpose the inverse or inverse the transpose. (K^T S_Y^{1} K)^{1} is a more traditional way to write this equation.
Indeed true, thanks. We changed it to (K^{1})^T instead.


RC2: 'Comment on amt2021212', Anonymous Referee #2, 31 Aug 2021
The authors propose an alternative extinction coefficient retrieval methodology for the ESA’s Aeolus ALADIN instrument that performs better compared to the current Standard Correct Algorithm (SCA). The authors formulated the retrieval problem as a maximum likelihood estimation (MLE) problem in which the ALADIN HSRL forward models are incorporated. The benefit of the MLE methodology is that it allows for constraining the physical values of the parameters that are being estimated (e.g. backscatter coefficient and lidar ratio).
The reviewer is disappointed that the authors did not expand on the work of Xiao et al. 2020 and Marais et al. 2016 who both introduced MLE methods that employ the total variation (TV) image smoothness constraint to improve the inference of the extinction coefficient for Poisson noisy observations. The publication would have been much better if the authors at least applied the MLE TV method on spaceborne lidar data. There are readily available software implementations of the MLE TV method, one of which is called SPIRAL and it is available at http://drz.ac/code; the authors will be required to make modification to the SPIRAL code, but that should be within the expertise of the authors.
The reviewer, however, is recommending a minor revision since the authors are improving the state of the art specific to the Aeolus ALADIN instrument measurements.
Major comments:
Lines 81 to 83: In the references that are cited, the work were specific to developing inversion methods for ground based lidar instrument. The authors are proposing a spaceborne inversion HSRL method. This distinction matters because for ground based lidar instruments inaccuracies in the geometric overlap function calibration parameter can introduce biases in the extinction coefficient. Hence, the backscatter coefficient is inferred as a separate first step to isolate any geometric overlap function induced biases to the extinction coefficient. With spaceborne HSRL measurements and depending on the instrument, there is not necessarily a geometric overlap function which are confirmed by equations 2 and 3 in this paper. Thus, a spaceborne HSRL allows for the simultaneous inference of the backscatter and extinction coefficients that are not biased to due to the geometric overlap function.
For the sake of understanding how this paper's method relates to the references that are cited, the reviewer suggests that the authors mention that the geometric overlap function calibration parameter is a limiting factor for groundbased HSRL instrument when trying to simultaneously infer both the backscatter and extinction coefficients.
Line 198: It is unclear where in (Flamant et al 2017) equations 9 and 10 are defined. Since equations 9 and 10 are esoteric, it would be helpful for the reader to know where in (Flamant et al 2017) equations 9 and 10 are defined since (Flamant et al 2017) has 124 pages.
Lines 199200: Equations 9 and 10 do not make sense. For example, consider equation 9. Since L_{p,i} > 0, we have that 2L_{p,i} < 0. Hence, X_{i} < 0, since the rest of the terms in equation 9 are positive. However, this a contradiction, since X(r) in equation 5 is strictly positive.
Line 224: Should J be J_{obs}? Should S_y not be the inverse of the measurement error covariance matrix? Refer to equations 2.21 and 5.3 of (Rodgers, 2000).
Line 287: The reviewer does not agree with the statement "MLE estimates usually suffer from overfitting and noise amplification.", because it is the forward model along with the MLE that determines whether there is overfitting or "noise amplification". E.g., the averaging operator (or just averaging) can be derived via MLE and averaging operator with a sufficient number of samples does not overfit. In other words, whether a MLE overfits or does "noise amplification" depends on the MLE's parameterization with the forward model.
Line 230 to 232: Could the authors provide a reference to support the claim that the averaged signal's noise can be approximated via a Gaussian distribution? For example, are there sufficient photon counts during the nighttime at 30km altitude that are accumulated and normalized in order for the average signal's noise (at 30km) to be accurately approximated by a Gaussian distribution?
Lines 259 to 261: Poisson noise is not additive and therefore it is confusing to read \hat{sigma}_{s} = \sqrt{s + \epsilon_{s}}. For Poisson noisy observations the noise standard deviation is \sqrt{s + b_s + b_d} where b_s is the solar background radiation and b_d is the dark counts.
Lines 291 to 292: The statement "the estimate should fit as close as possible to the signal data and only solve the physical contradictions" is contradictory compared to the previous sentences in this paragraph. If the estimate should fit as close as possible to the observations (signal data), then why not let the LBFGSB converge to a solution and why bother introduce an early stopping criteria of 40 000 iterations? In other words, in this sentence the authors are suggesting that the estimator should "overfit".
Lines 431 to 433: Please add a reference that shows the range of lidar ratio values for desert dust at wavelength 355nm.
Lines 457 to 460: Are the authors reporting copolarized lidar ratios or BERs? If BERs are reported, how were the copolarized lidar ratios converted to BERs?
Sections 4.1 and 4.2: The authors should validate the uncertainty quantification as described between lines 300 and 305, since it is unclear how accurate the proposed uncertainty quantification is for observational data.
Section 4 & conclusion: The authors do not thoroughly explain why SCA methods produce biased results compared to the MLE method, and it will be insightful know why SCA methods produce biased results. Could it be that the X and Y (equations 7 and 8) are modeled by SCA (equations 9 and 10) could introduces biases in the estimates of the backscatter and extinction coefficients relative to the MLE method, since the MLE method employs equations 7 and 8?
Comment about the methods: It will also be useful know what is the performance of a method that directly algebraically solve for the lidar ratio and backscatter coefficient via equations 7 and 8. In other words, what is the performance of the MLE method without constraints? The reason why this comparison will be useful, is to gain an understanding of the low performance of the SCA methods. Specifically, is the loss of performance due to the formulation in equations 9 and 10, or because of the lack of physical constraints?
Minor comments:
Line 2: In the first paragraph of the manuscript introduction all the letters of Aladin is capitalized. To be consistent, Aladin should be all in capital letters.
Line 5: "Being and HSRL" should be "Being an HSRL".
Line 6: Backscatter coefficients and lidar ratios are normally reported without polarization.
Line 6: It will be helpful to the reader to know that ALADIN does not make crosspolariation measurements (lines 107 & 108) and therefore ALADIN is not able to make direct lidar ratio measurements.
Line 8: Do the authors mean that the inversion problem is statistically illposed? If the inversion problem is illposed, then it would not be possible to infer the extinction coefficients from ALADIN measurements without using lidar ratios.
Line 33: It is unclear what the authors mean by "because high uncertainties in climate change modelling regard the indirect effect of aerosols on clouds and anthropogenic radiative forcing (Illingworth et al., 2015).". Are the authors saying there are high uncertainties of indirect effects of aerosols on clouds in climate change modelling?
Line 4142: Lower SNRs of what? This part of the sentence is vague.
Line 47: Do the authors mean that the inversion problem is statistically illposed? If the inversion problem is illposed, then it would not be possible to infer the extinction coefficients from ALADIN measurements without using lidar ratios.
Line 69: Fine resolution of what? Image resolution?
Line 75: "reformulated as a " would read better than "rephrased into a".
Line 133: Do the authors mean detector random errors instead of wind random errors?
Line 135: Is it unclear what is meant by Basic Repeat Cycle (BRC) or Observation. Are these terminologies that are used in the field of wind lidar? If so, would it be helpful to add a reference? Does BRC mean that every 30 consecutive vertical profiles are accumulated?
Line 148: "is used as input to the optical" should be "is used as input by the optical".
Line 162: The reviewer presume that the measurement geometry is the so called geometric overlap function. It is unclear why the authors define O(r) which it is not used in equations 2 and 3?
Line 163: The range of T(r) is defined but not for \beta(r). To be consistent it will be helpful to also define the range for \beta(r) e.g. \beta(r) > 0.
Line 168: If the ALADIN instrument is measuring the copolarized backscattered energy, shouldn't the molecular backscatter coefficient also be labeled as copolarized?
Line 180: Using the word excessive sounds negative; it would be best not to use this word.
Line 190: "When" should be "where".
Line 203: What is 'this' referring to? Is 'this' referring to L_{p,sat}?
Line 205: Did the authors mean to say "or" instead of "/"?
Line 221: It is unclear what the "It" refers tp. Does the "It" refer to the second term in equation 12?
Line 222: Should J be J_{obs}?
Line 225: In regards to S_y, refer to comment of line 224.
Line 226: Should "Lidar" be "lidar"?
Line 234: In regards to S_y, refer to comment of line 224.
Lines 236 to 245: The paragraph is superfluous since 1) the authors are not employing a constraint or penalty term and 2) the paragraph does not add value to the current text. If the authors at a later stage employ a constraint or penalty term and report results in a next publication, the next publication can include this paragraph. Furthermore, the conclusion does discuss employing a regularization term.
The reviewer suggests replacing this paragraph with one line saying that equation 14 is applied on each averaged vertical profile of measurements.
Line 252: "would show off diagonal" should be "would have off diagonal".
Line 256: The last sentence on this line is unclear. To what does "following" refer to?
Lines 263 to 264: Are the 2.9km and 87km numbers referring to different horizontal resolutions?
Line 268: The citation style of Wandinger at at. 2015 is not consitent with the citation style of Illingworth et al. 2015.
Line 288: It is unclear why new sentence starts with "But".
Lines 306 to 313: Large portions of subsection 3.3 are repeated in the first paragraph of section 4. Therefore this subsection can be removed.
Line 362: It is unclear what the authors mean by "wither".
Line 382: It is unclear what the authors mean by "wither".
Figures 2 and 3: For the curtain plots; what is the horizontal axis? Profle number? Seconds?Line 427: Can the authors elobrate on how the crosspolarized lidar ratio is transformed into BER (the actual lidar ratio)?
Line 434: More robust compared to what?
Line 473 to 475: See the comments of lines 81 to 83.
Citation: https://doi.org/10.5194/amt2021212RC2 
AC2: 'Reply on RC2', Frithjof Ehlers, 01 Oct 2021
General comment:
First thing to note is that we were very glad about the thorough feedback of all Referees, who evidently made the effort to go through the details presented in our manuscript. This helped a lot to refine the text and we hope that we could make our statements more concise now. Especially the method of calculating the uncertainties in the MLE method was addressed by more than one of the Referees and required some changes that we want to present to all of Referees.
The procedure outlined in the last paragraph of Appendix C was relying on the simple equation alpha = gamma * beta with extinction coefficient alpha, lidar ratio gamma and backscatter coefficient beta. Motivated by comparing the obtained variances via (C2) with the variances one would get from a standard error propagation with constant lidar ratio sigma_a^2 = gamma^2 * sigma_b^2, we argued e.g. that the uncertainty in alpha cannot exceed 200 sr times the uncertainty of beta in a single bin, whatever the precise value of gamma is in reality, due to the upper bound in lidar ratio.
Now, after the feedback and some graphical considerations, the problem with this statement is, that it leaves aside any uncertainty or spread of the lidar ratio estimate itself, as it can lie somewhere in between 2 sr and 200 sr and hence should have nonvanishing variance sigma_g^2, causing another term to arise from the view of standard error propagation, increasing the uncertainty of alpha proportional to beta itself. This is why we noticed that the presented approximation was not correct, although we observed uncertainty values that matched approximately the SCA mid bin output for the SAL case for backscatter and extinction coefficients, which seemed realistic at first glance.
Following this, we evaluated the errors from equation (C2) directly against the uncertainties calculated by the SCA standard error propagation as in Flamant et al., 2020, and found good agreement, when the same Poisson noise variances are used in the S_y in (C2), see attached figure.
That means, that with (C2) we are only able to provide uncertainties for the unconstrained problem (as SCA solves it), but especially the soobtained extinction uncertainties are overestimated and not representative of the real uncertainties, as one can see in the comparison of SCA and MLE in the simulation cases.
In lack of a representative alternative for uncertainty estimation we decided to delete the lidar ratio and extinction uncertainties of the MLE solutions in Figures 5 and 6, for now, stating that there is no representative error estimate available yet. The challenge of reporting representative uncertainty estimates is now being stressed in the abstract and conclusion and will require future investigation.
Regarding comments of Referee 2
We thank for the thorough review of our manuscript and are hopefully able to clarify on most aspects. The main concerns of Referee 2 regard the comparability of the MLE method versus a purely algebraic inversion scheme as benchmark. We can clarify here, that the forward model of SCA and MLE is virtually identical (besides some equivalent variable transforms outlined in section 3.1). Hence, the only difference between SCA and MLE is how the forward model equations are solved/inverted. That implies, that any bias due to the forward model formulation should be shared, while any other differences in bias must then either regard the zeroflooring (only in SCA extinction) or the interaction of nonlinearities and noise levels (in SCA and SCA MB). Because this point was not made entirely clear in the manuscript, we modified the main text as outlined in the individual comments below.
All grammar and spelling errors were adopted or corrected without explicit mentioning.Lines 81 to 83: In the references that are cited, the work were specific to developing inversion methods for ground based lidar instrument. The authors are proposing a spaceborne inversion HSRL method. This distinction matters because for ground based lidar instruments inaccuracies in the geometric overlap function calibration parameter can introduce biases in the extinction coefficient. Hence, the backscatter coefficient is inferred as a separate first step to isolate any geometric overlap function induced biases to the extinction coefficient. With spaceborne HSRL measurements and depending on the instrument, there is not necessarily a geometric overlap function which are confirmed by equations 2 and 3 in this paper. Thus, a spaceborne HSRL allows for the simultaneous inference of the backscatter and extinction coefficients that are not biased to due to the geometric overlap function.
For the sake of understanding how this paper's method relates to the references that are cited, the reviewer suggests that the authors mention that the geometric overlap function calibration parameter is a limiting factor for groundbased HSRL instrument when trying to simultaneously infer both the backscatter and extinction coefficients.
We acknowledge the reviewers point of view and suggest to add the sentence: “It is important to note that such a simultaneous retrieval with ground based lidars would require the additional geometric overlap function calibration parameter. Hence, it is often preferred to retrieve backscatter coefficients independent of extinction coefficients to mitigate biases on the former due to calibration errors.”
Line 198: It is unclear where in (Flamant et al 2017) equations 9 and 10 are defined. Since equations 9 and 10 are esoteric, it would be helpful for the reader to know where in (Flamant et al 2017) equations 9 and 10 are defined since (Flamant et al 2017) has 124 pages.
We are aware that these equations are rather unappealing, but felt the need to state them once in the exact form they are used. We now refer to the equivalent equations by number in the cited document: “[...] see equations 6.356.36 in Flamant et al. (2017) and definitions above.” and also refer to equations (5) and (6) just above.
Lines 199200: Equations 9 and 10 do not make sense. For example, consider equation 9. Since L_{p,i} > 0, we have that 2L_{p,i} < 0. Hence, X_{i} < 0, since the rest of the terms in equation 9 are positive. However, this a contradiction, since X(r) in equation 5 is strictly positive.
Indeed the minus in the denominator was incorrect so it was removed in in (9) and (10) and (A2) and (A3). For your readability: The function (1exp(2x))/(2x) is almost equal to exp(x) for small x.
Line 224: Should J be J_{obs}? Should S_y not be the inverse of the measurement error covariance matrix? Refer to equations 2.21 and 5.3 of (Rodgers, 2000).
Thanks for spotting these mistakes, we corrected for them throughout the whole manuscript and added the suggested references after line 224.
Line 287: The reviewer does not agree with the statement "MLE estimates usually suffer from overfitting and noise amplification.", because it is the forward model along with the MLE that determines whether there is overfitting or "noise amplification". E.g., the averaging operator (or just averaging) can be derived via MLE and averaging operator with a sufficient number of samples does not overfit. In other words, whether a MLE overfits or does "noise amplification" depends on the MLE's parameterization with the forward model.
This is a good point and was certainly expressed with too little care in the manuscript. We now changed the wording to “MLE estimates may suffer from overfitting and noise amplification in lidar retrieval problems, so an implicit regularization is often achieved by optimal choice of the number of iterations [...]”, while keeping the existing references, in which more information can be found.
Line 230 to 232: Could the authors provide a reference to support the claim that the averaged signal's noise can be approximated via a Gaussian distribution? For example, are there sufficient photon counts during the nighttime at 30km altitude that are accumulated and normalized in order for the average signal's noise (at 30km) to be accurately approximated by a Gaussian distribution?
The reference to the central limit theorem was misleading. What we meant to say is that besides its discrete nature and asymmetry, the Poisson distribution can be decently well approximated by a Gaussian with same mean and variance. So instead we propose to change the old wording
“[...] because the Poisson noise distribution becomes indistinguishable from a Gaussian given sufficient signal accumulation (central limit theorem).”
into
“[...] because the discrete Poisson noise distribution can already be decently well approximated by a smooth Gaussian with identical mean and variance for very low (photon) counts and the aforementioned additional noise sources and their corrections, e.g., subtraction of measured solar background, will naturally smear out the discrete nature of the Poisson noise.”
Lines 259 to 261: Poisson noise is not additive and therefore it is confusing to read \hat{sigma}_{s} = \sqrt{s + \epsilon_{s}}. For Poisson noisy observations the noise standard deviation is \sqrt{s + b_s + b_d} where b_s is the solar background radiation and b_d is the dark counts.
The mathematical formulation was meant to support the understanding, but can indeed be misleading here as it was inspired by the perspective of Gaussian noise. We propose to remove this part and include “because a single draw does in general not equal the true mean and the true variance” as a reason for this bias from using the signal samples themselves.
Lines 291 to 292: The statement "the estimate should fit as close as possible to the signal data and only solve the physical contradictions" is contradictory compared to the previous sentences in this paragraph. If the estimate should fit as close as possible to the observations (signal data), then why not let the LBFGSB converge to a solution and why bother introduce an early stopping criteria of 40 000 iterations? In other words, in this sentence the authors are suggesting that the estimator should "overfit".
There is no contradiction if read the following way: Since the number of 40.000 iterations is really high (see other reviewer’s remarks), we try to maximise convergence by brute force and hence aim to eradicate any implicit regularization additional to the constraints.
Lines 431 to 433: Please add a reference that shows the range of lidar ratio values for desert dust at wavelength 355nm.
This is certainly helpful for the flow of reading. We added “(Wandinger, 2015)”.
Lines 457 to 460: Are the authors reporting copolarized lidar ratios or BERs? If BERs are reported, how were the copolarized lidar ratios converted to BERs?
No BER is calculated here, but we refer explicitly to the copolarized lidar ratio now.
Sections 4.1 and 4.2: The authors should validate the uncertainty quantification as described between lines 300 and 305, since it is unclear how accurate the proposed uncertainty quantification is for observational data.
see statement of the author’s
Section 4 & conclusion: The authors do not thoroughly explain why SCA methods produce biased results compared to the MLE method, and it will be insightful to know why SCA methods produce biased results. Could it be that the X and Y (equations 7 and 8) are modeled by SCA (equations 9 and 10) could introduces biases in the estimates of the backscatter and extinction coefficients relative to the MLE method, since the MLE method employs equations 7 and 8?
The forward model for SCA and MLE is equivalent. The MLE also relies on eq. 9 and 10, see line 212. So this is not the reason for biases.
The biases occur in our opinion mostly due to the nonlinear operations in equations (9) and (10): Take as an example the ratio operation and arbitrary random variables x and y such that mean(x)/mean(y) = 1. Now, depending on the noise amplitudes on x and y, mean(x/y) can be biased high to any value, solely because of the varying noise amplitude. Considering the nonlinear character of equations (9) and (10), it becomes explainable that biases in SCA and SCA MB can be (partly) mitigated with MLE by suppressing noise. This is now also discussed in more detail in sections 4.1 and 4.2, aided by better visualisation of the significant aerosol below 2 km with modified versions of graphs D1 and D2. We added:
“For the origin of this bias we can think of two causes: Firstly, the violation of the hypothesis of uniformly filled bins due to the strong gradient in aerosol concentration and, secondly, the nonlinear way the backscatter coefficient is calculated from $\beta_{,p,i}=Y_i\beta_{m,i}/X_i$, because here $mean(\beta_{,p,i})$ will become biased high increasingly with increasing uncertainty of $X_i$.” and "The high backscatter coefficient bias in the lowermost 2 km [in simulation case II] is more pronounced in all algorithms as compared to atmospheric simulation case I, supporting the hypothesis that it originates from the nonlinear ratio operation rather than the gradient of the aerosol concentration. Consequently, due to its noise suppression capabilities, that is likely why MLE shows less bias in the lowermost 2 km. A similar statements can be made for the extinction, because it is essentially calculated from a ratio of subsequent pure molecular signal values $X_i$."
Comment about the methods: It will also be useful know what is the performance of a method that directly algebraically solve for the lidar ratio and backscatter coefficient via equations 7 and 8. In other words, what is the performance of the MLE method without constraints? The reason why this comparison will be useful, is to gain an understanding of the low performance of the SCA methods. Specifically, is the loss of performance due to the formulation in equations 9 and 10, or because of the lack of physical constraints?
As stated above, both SCA and MLE rely on the same equations (9) and (10). So the SCA is the direct, algebraical solver for the MLE problem, as suggested. Hence, if the MLE had no constraints and was initialised with the same condition of vanishing optical depth in the first bin (as SCA is) then outputs of SCA (without zero flooring of extinction, see e.g. figure 10 in https://earth.esa.int/eogateway/documents/20142/37627/AeolusDataInnovationScienceClusterDISCLevel2Auserguide.pdf for an example) and MLE would be virtually identical. Indeed we see very similar noise patterns to this figure 10 when applying no constraints in MLE. We added the following to the text after equation (14):
“It is important to mention that it is only the apriori knowledge in form of the boxconstraints that makes the MLE solution distinct from the algebraic SCA solution (without zero flooring, see section 6.2.2.1 in Flamant et al. 2020), because this algebraic solution corresponds to the exact signal values in y and therefore to J_{obs}=0, which is the global minimum to the unconstrained counterpart of problem (14).”
Line 6: It will be helpful to the reader to know that ALADIN does not make crosspolariation measurements (lines 107 & 108) and therefore ALADIN is not able to make direct lidar ratio measurements.
We agree that this might be confusing for the lidar community, hence we added the remark “(the crosspolarized return signal is not measured)”.
Line 8: Do the authors mean that the inversion problem is statistically illposed? If the inversion problem is illposed, then it would not be possible to infer the extinction coefficients from ALADIN measurements without using lidar ratios.
This statement was based on the work of Pornsawad et al., 2008: “But From a mathematical point of view Eq. (1) is a Volterra integral equation of the first kind, which is illposed.” We find this wording misleading as well and thus decided to drop “illposed” throughout our text.
Line 33: It is unclear what the authors mean by "because high uncertainties in climate change modelling regard the indirect effect of aerosols on clouds and anthropogenic radiative forcing (Illingworth et al., 2015).". Are the authors saying there are high uncertainties of indirect effects of aerosols on clouds in climate change modelling?
That sentence was a bit bulky. The actual statement in Illingworth et al. reads: “The largest single cause of uncertainty in anthropogenic radiative forcing is from the indirect effect of aerosols on clouds.” So we reformulated our text to say “[...]; the largest single cause of uncertainty in antropogenic radiative forcing has been reported to be from the indirect effect of aerosols on clouds (Illingworth ...).”
Line 4142: Lower SNRs of what? This part of the sentence is vague.
This refers to the actual lidar signals. So the SNR of the signal. In order not to write the “signal’s signaltonoise ratio” we added “in the receive channels” for clarity.
Line 47: Do the authors mean that the inversion problem is statistically illposed? If the inversion problem is illposed, then it would not be possible to infer the extinction coefficients from ALADIN measurements without using lidar ratios.
See answer to comment on line 8.
Line 69: Fine resolution of what? Image resolution?
We mean the effective resolution in the retrieval, which is not clear. Likewise, any explanation would inflate the text here, so we kept only the statement about the higher precision.
Line 133: Do the authors mean detector random errors instead of wind random errors?
No, the primary concerns for Aeolus were the wind random errors.
Line 135: Is it unclear what is meant by Basic Repeat Cycle (BRC) or Observation. Are these terminologies that are used in the field of wind lidar? If so, would it be helpful to add a reference? Does BRC mean that every 30 consecutive vertical profiles are accumulated?
These are Aeolus mission terminologies, which are to find in the Aeolus documentation referenced in this section. But maybe it was misleading that we used “integrated”? Instead, we mean “accumulated”, which should clarify.
Line 162: The reviewer presume that the measurement geometry is the so called geometric overlap function. It is unclear why the authors define O(r) which it is not used in equations 2 and 3?
We wanted to give a general formulation, but it is true that we don’t use it. We added “For Aeolus the range overlap function $O(r)$ equals 1.”
Line 163: The range of T(r) is defined but not for \beta(r). To be consistent it will be helpful to also define the range for \beta(r) e.g. \beta(r) > 0.
This was added.
Line 168: If the ALADIN instrument is measuring the copolarized backscattered energy, shouldn't the molecular backscatter coefficient also be labeled as copolarized?
To my knowledge, it is assumed here that the molecules are a nondepolarizing target, making both the same.
Line 203: What is 'this' referring to? Is 'this' referring to L_{p,sat}?
We now explicitly referenced the equations we intend to refer to.
Line 205: Did the authors mean to say "or" instead of "/"?
Yes, we corrected this.
Line 221: It is unclear what the "It" refers tp. Does the "It" refer to the second term in equation 12?
Yes, also corrected
Line 222: Should J be J_{obs}?
Here it is equivalent, but we clarified this in the text.
Line 225: In regards to S_y, refer to comment of line 224.
corrected
Line 226: Should "Lidar" be "lidar"?
corrected for all occurrences
Lines 236 to 245: The paragraph is superfluous since 1) the authors are not employing a constraint or penalty term and 2) the paragraph does not add value to the current text. If the authors at a later stage employ a constraint or penalty term and report results in a next publication, the next publication can include this paragraph. Furthermore, the conclusion does discuss employing a regularization term.
The reviewer suggests replacing this paragraph with one line saying that equation 14 is applied on each averaged vertical profile of measurements.
We acknowledge the reviewer’s point of view, but are bound to report our method the way it is implemented. Since the actual implementation hands the problem (15) to the LBFGSB algorithm and not (14) in a for loop over single profiles, we prefer to keep the paragraph.
Line 256: The last sentence on this line is unclear. To what does "following" refer to?
We changed it to “As in SCA, this overlap is not considered in this work.”
Lines 263 to 264: Are the 2.9km and 87km numbers referring to different horizontal resolutions?
Yes, exactly, see Fig. 1 for clarification (horizontal length values added).
Line 268: The citation style of Wandinger at at. 2015 is not consitent with the citation style of Illingworth et al. 2015.
corrected
Line 288: It is unclear why new sentence starts with "But".
Pardon, that is a remnant of my German. We coupled the sentences to clarify our message.
Lines 306 to 313: Large portions of subsection 3.3 are repeated in the first paragraph of section 4. Therefore this subsection can be removed.
Many thanks for spotting this mistake. We removed subsection 3.3.
Line 362+382: It is unclear what the authors mean by "wither".
We meant to say “either”; corrected
Figures 2 and 3: For the curtain plots; what is the horizontal axis? Profle number? Seconds?
This is the profile number, but in Aeolus’ missionspecific vocabulary. See Instrument section: “[...] a total of 30 measurements are accumulated to one socalled Basic Repeat Cycle (BRC) or Observation (Aeolus mission terminology), equivalent to approximately 87 km along track distance on ground.”
Line 427: Can the authors elobrate on how the crosspolarized lidar ratio is transformed into BER (the actual lidar ratio)?
The crosspolarized BER is meant. This is added to the text now.
Line 434: More robust compared to what?
As compared to the reference algorithms. “than SCA and SCA MB” has been added.
Line 473 to 475: See the comments of lines 81 to 83.
We added: “The choice of dependent and independent retrieval of backscatter and extinction coefficients is a tradeoff between improved precision and potential biases. A coupled retrieval may improve the precision of the retrieved backscatter coefficients, but it relies on a potentially erroneous calibration as input (geometric overlap function and crosstalk correction).”

AC2: 'Reply on RC2', Frithjof Ehlers, 01 Oct 2021

RC3: 'Comment on amt2021212', Anonymous Referee #3, 08 Sep 2021
Review of "Optimization of Aeolus Optical Properties Products by MaximumLikelihood Estimation" by Ehlers et al.
The article describes a new algorithm for retrieving aerosol optical properties from Aeolus that improves various shortcomings of the existing Aeolus algorithms. Specifically, the new algorithm operates on the direct measurements and their uncertainties (rather than crosstalk corrected pseudosignals ignoring uncertainty), simultaneously fits both extinction and backscatter (rather than sequentially), and uses explicit constraints to force solutions to have nonnegative optical depth and lidar ratio within expected limits (rather than implementing these constraints as filters after the retrieval). The analysis includes the calculation of error bounds that appear to better reflect actual uncertainties than those of the standard algorithm. Comparisons between the new and existing algorithms are presented for simulated scenes and real data scenes where correlative lidar measurements are available.
I have an overall favorable impression of the manuscript and look forward to seeing it published. The improvements over the existing algorithms are useful and justify the publication of this study. An appropriate level of detail is presented and the logical flow of the manuscript is good and helpful for understanding the algorithm and results. However, there are some important aspects which need clarification and revision, particularly those relating to the uncertainty calculations. Since I see these as critically important, I am recommending major revisions, despite the overall high quality of the manuscript.
Major concerns
P3 L77. A positivity constraint on quantities where zero is a valid and common expected value will produce a bias. The authors mention this in the context of the SCA algorithm, but this is also a concern in the new algorithm. It might be better not to include the positivity constraint on backscatter and extinction in the new algorithm. Have the authors tried their MLE method without the positivity constraints?
p10 Box constraints. I think its possible that adding constraints has a smaller impact on improving the results than the other improvements: using raw measurements and uncertaintites in a coupled retrieval. Additionally, the constraints have potential negative consequencs: (1) that they could lead to bias (similar to what happens in SCA) and (2) that they make propagation of the uncertainty very difficult. Have the authors attempted the MLE retrieval without adding the box constraints? It would be useful to compare results with and without using the box constraints. If they are just as good, eliminating the box constraints would eliminate the two issues mentioned above. If they are not as good, the comparison would give a clearer view of the impact of the different improvements.
P19, L469 "This is mainly due". Similar to the previous comment. This statement is made without any analysis to suggest how the different algorithm features dominate the improvements. The authors should assess the impact of the box constraints separately, to support this statement that it is the dominant cause of improved results.
P6 Eqs 2 and 3. P21, Eqs A4 and A5. Very little is said about the crosstalk constants C1, C2, C3 and C4. The top of page 7 seems to imply the constants are known, but in fact they are very challenging to determine and have significant uncertainty associated with them. More information is needed about how these are determined and what are typical values for Aeolus. This is also important for the discussion of the correlated errors and for understanding the magnitude of that problem.
p1011 L294304, two approaches to error quantification. While in general I like the idea of analytical propagation of errors using sensitivity analysis, the propagation is seriously challenged by the difficulty of representing the impact of the constraint in the propagation equation. I'm not very convinced by this method of rescaling the variance and assuming the correlation is unchanged (Appendix C, near P23 L575). Can it be explained more clearly why the correlation matrix should remain unchanged? Also, using the lidar ratio constraint to (potentially) reduce both the extinction and backscatter error bounds separately gives an impression of circularity. The authors should add rigor and validation to make this method more convincing.
P25 Figure D1. Although the Monte Carlo methodology is discussed, it seems that the results of the Monte Carlo approach are only shown in Appendix D without any analysis or discussion. These should be promoted to main text and properly analyzed.
Both methods are available for the simulation cases, so I'd like to see a comparison of the analytical (Appendix C) method against the Monte Carlo results for the simulation cases, which might bolster my confidence in the analytic method.
It would also be useful to compare the propagated uncertaities with and without the adjustment for the constraints, to see how big this impact is.
The authors say a Monte Carlo approach is not feasible for the measurement cases, and I think this refers to the fact that they cannot vary the true measurement error. However, it is certainly possible to simulate measurement noise from the measurement error covariance matrix, S_y. Using these simulated measurement errors in a Monte Carlo approach would give an accurate view of how S_y is propagated through the retrieval, more immediately convincing than the method in Appendix C. So, going further, I'd like to see a Monte Carlo propagation of S_y, which could be used to validate or replace the current propagation.
P17, Figure 5 (also figure 6). I can't understand which error bounds are shown in Figure 5 and Figure 6. The descriptions "lower error bound" and "upper error bound" are confusing because they don't match any description in the methodolgy, and they are possibly inaccurate as well. Another terminology should be seletected, preferably one that reflects the descriptions in the methodology section. The two sets don't seem to be the "two approaches" introduced on P1011. Are they instead two formulations of the analytic propagation of errors using different characterizations of the measurement error S_y? Is the "upper error bound" the propagated uncertainty from Appendix C? (If so, calling it an upper error bound is particularly problematic since the CramerRao inequality characterizes a lower bound, not an upper bound). And how is the "lower error bound (Poisson)" calculated? I don't see that in the manuscript.
If indeed the one labeled "upper error bound" is the one produced in this analysis, and the one labeled "lower error bound" is the standard one for SCA and midbinSCA, then another emergent theme of this analysis is that the uncertainty of the standard algorithms is inaccurate as well. This should be highlighted in the manuscript as another primary impact of the new algorithm.Minor comments
P1 L8 "algebraic inversion scheme to a (partly) illposed problem and therefore sensitive to
measurement noise". The sentence should probably be reworked. "(partly) illposed" is not really informative, and even a wellposed inversion is sensitive to measurement noise.
P2 L5354 I think this point about there being no resolution between these regimes is particularly well articulated.
P3, L7879. I think the sentence should be split up into multiple sentences that clearly list the factors that led to poor results in the SCA (there are at least two, I think: the correlated noise and the incorrect removal of negative extinctions) and the factors that lead to improvements (coupled retrieval, using the measurement error covariance matrix in the retrieval, box constraints), and to explain their causal relationships. In particular, I believe it is incorrect to suggest that the box constraints are responsible for eliminating the correlated noise issue, as this sentence seems to say. Instead, I think the bias is avoided by using the measurement channels and their error covariance matrix in a coupled retrieval.
P4, L130. How is the grid spacing chosen? What are the typical values?
p7 Eqs 9 and 10. The minus sign in the denominator does not appear in Flament et al 2017. The rest of the equation looks similar but with a few more algebraic steps. I think it is otherwise correct, but I would like the authors to doublecheck to be sure.
p7 Eqs. 9 and 10. It would be helpful to have a sentence or two summarizing the derivation of the quoted equations; for example, an explanation that the equation includes an expression that explictly integrates the transmission over an extended vertical bin.
p8 near the the end of section 3.1, it would be good to have a sentence or two discussing the anticorrelated noise that's mentioned in the Appendix. This seems like an important point that is referred to both in the introduction and later, so it should be described in the main text with enough detail so a reader knows what it is about.
P9 L232. I'd like to know more about how the box constraint is implemented and specifically whether it may negatively impact results. Could it potentially bias the results (by perhaps producing a disproportionate number of solutions near the boundaries) or affect their precision of the solution (via a variable transform that changes the gradients near the solution and near the boundaries)?
p9 242. "future developments". This is really more of a response to another reviewer, but I just wanted to say that I support the authors' decision to perform single profile retrievals without including horizontal scene smoothing. It makes sense to explore the simpler solution first. Furthermore, not requiring scene continuity for the algorithm means that scene continuity can be used as a check on algorithm performance.
p9 L250 As I mentioned, I believe that using the measurements and their appropriate measurement error covariance matrix is largely responsible for avoiding the bias due to the correlated errors in the crosstalk corrected signals. Do the authors agree? If so, this paragraph might be a good place to highlight that.
p10 L272275. Is there anything in the current algorithm that addresses the bias from partially filled bins? If not, is the current algorithm compatible with the strategy implemented by Flament et al. 2017?
p10 L290292. There appear to be three rather weak arguments here to justify not using an a priori covariance matrix such as called for by OE. The weakest part of the argument, in my opinion, is saying that the algorithm has no prior information or regularization. In fact, the constraint is prior information and does provide regularization. If the authors believe the influence of the constraint is significant (which I think they do, because they attribute the improved results largely to implementing the constraint), then even suggesting that this is a weak constraint would not be justified. Therefore, the authors should acknowledge that the constraint plays the role of prior information in the retreival, but they chose not to cast this constraint in the terms required by optimal estimation, because that would require a different (probabalistic) form for the prior that isn't compatible with the desired form of the constraint. In my opinion, that is a reasonable reason to use constrained MLE rather than OE. (However, the choice leads to difficulties in working out the correct way to calculate the impact of the constraint on the uncertainties, which I discussed in the "Major" section).
p10 L295. Some clarification would be helpful here about the simulated measurement noise. Is it simulated as gaussian with the variance determined from Appendix B, or is it simulated with various separate error components? Which components are included?
p12 L347349. While I certainly agree that filtering negative results would cause a high bias that is worse in bins with low SNR, I can't clearly follow the explanation in this part of the text and I don't see a particularly clear indicator that they are triggered by the shift from coarser highaltitude bins to finer lowaltitude bins as implied by the text. Can the explanation be clarified?
p12 L354. The indicator from the averaging kernal is a good idea. Is there a quality flag in the data product related to this? It would be nice if this indicator were included in the plots to show which bins are not trustworthy. This would be particularly useful in Figure 4, for instance, where I am curious to know if the bins below the lofted aerosol are reliably retrieved. This curiosity is fueled by the fact that CALIPSO's 532 nm data frequently misses the aerosol at the bottom of attached layers due to attenuation, making them look lofted. The HSRL capability should act to prevent this problem given sufficient SNR, but on the other hand, attenuation at the shorter wavelength of 355 nm would be worse. So, to be sure, it would be good to see some indication from the retrieval algorithm that the lack of aerosol below the apparent plume bottom is reliable.
p12 L354. "The extinction bin closest to the ground cannot be well retrieved." Is this true in general, or specific to this case?
p12 L357 "Otherwise". I'm confused by this sentence, and not sure if I'm confused by content or just the wording. Does "otherwise" indicate the strategy of taking the mean lidar ratio? I agree that taking the ratio of the mean extinction and backscatter is a better strategy than the mean lidar ratio and will give potentially different results, since lidar ratio for small extinction and backscatter is not as reliable as when the SNR is higher. By why is the mean lidar ratio contaminated by the first guess? And how does taking the ratio of mean extinction and backscatter avoid the influence of the first guess?
Figure 2. What is the explanation for the low bias in the median lidar ratio from the MLE for the lowest bins?
Figure 24 and D1: The line plots are so small that I can't see important information. The data are particularly important below about 2.5 km where there is significant aerosol, but this is a very small portion of the plot and not readable due to its size and the closely spaced grid lines. Specifically, the lowest bin is called out in the text, but it is so small I cannot see the data or error bar for that lowest bin in the line plots. Please include inset boxes to show the line plot data in the lowest 2.5 km, or a second set of line plot figures that show only the lowest 2.5 km, or in some other way improve the visualization of the lowest 2.5 km.
p14 L362. The reduced uncertainty from the new retrieval compared to the SCA should be discussed and quantified. I think the new retrieval probably produces usefully smaller uncertainty, but the fact that the figures are so tight makes it nearly impossible to see the region where there is significant aerosol below ~2.5 km, and I can't even tell if the error bars are smaller than 100%. A discussion of uncertainty results is just as important as the mean tendency of the profile, because, for example, a profile that "looks right" but is indistiguishable from 0 everywhere due to its uncertainty would be rather useless.
P14, L366. Is this number a typo? exp(2*0.4) = 0.45 not 0.38.
P14, L369379. It seems that it's not just delayed (i.e. slow decay below the cloud) but the cloud is smoothed into the regions both below and above. In other words, it looks like the effective resolution of the extinction is much less than the backscatter, which make sense since it takes at least two measurements to calculate a derivative. Is there any attempt to calculate backscatter on the same coarser resolution as extinction to produce the lidar ratio?
P15, Figure 3. It appears that the bias in the mean backscatter below the cloud is actually worse in the MLE than in the SCA and SCAmidbin results. This should be discussed.
P19, L449, what causes missing values in SCA midbin?
P19, L458, Lidar ratios aren't shown in the figure for the Real Data Case II, as in other cases. Better to show them, if possible.
P19, L459460, Is this comparison between copolarized lidar ratio from AEOLUS to total (co and crosspolarized) lidar ratio from Polly? This is not the best option. Since Polly is also sensitive to polarization, wouldn't it be possible to calculate copolarized lidar ratio from Polly for a more direct comparison? In any case, it should be clearly stated what's being compared.
P19, L461. It seems strange that the quoted uncertainties are from the Poisson method that the authors are hoping to replace. It would be better to quote uncertainties from the method that the authors think are more represenatative.
P 20, Figure 6. Is there an uncertainty bound available for the "ground truth" (which is also a retrieval)?
P20, Eqs A2 and A3 are difficult to mentally derive from the previous step. It would be helpful if more intermediate steps were added to make it clearer how the equation is derived. The appendix is a good place to do this.
P22, L538. This point about anticorrelated error is interesting and informative. However, some parts of the discussion are confusing. For instance, Eqs A7 and A8 show that an error spike in one measured channel gets distributed in an anticorrelated way into the corrected signals. But how does it follow that there is correlation (or anticorrelation) in the errors in backscatter and extintion? It seems logical that if the errors in the two corrected channels are anticorrelated, the ratio would tend to be biased low, so the backscatter would tend to be biased low. However, the error in the extinction would not be correlated with it, because extinction derives from gradients in just one of these corrected channels.
P23, L557, I suggest replacing "in spirit" with something more informative, such as "except with an adaptation to account for the impact of the constraint"
P23, L561. It's not true that no a priori knowledge is imposed. The box constraints are a priori knowledge.
P25 Figure D1. "Bias" and "standard error" should be defined in equations.Grammar and word usage
I found that, although the overall flow and logic are very quite good, in some spots the word choice made the writing difficult to understand. I've marked several below. My list is probably not complete, however, so a round of editing for English usage (not just spelling and grammar) would be helpful.
Title. The title would be more informative if it contained the word "aerosol". For example: Optimization of Aeolus Aerosol Optical Properties Products by MaximumLikelihood Estimation
P1 L1. Typically "embarks" is not used this way, and used only for the much narrower circumstance of a person getting on or off a vessel or starting a journey. It could be replaced with "includes"
P1 L5. Being an HSRL
P1 L9. Consider replacing "rephrase" with "rework" or "revise"
P1 L12. Delete "equally". I think most writers would use "also" rather than "equally" here, but it really isn't necessary at all.
P1 L14. Consider moving the phrase "to consolidate and illustrate the improvements" to the beginning of the sentence. I find it easier to follow when phrases that modify the verb are near the verb.
P2 L32. "Because of high uncertainties" and "regarding the indirect effect" (add "of" and "ing")
p2 L34. Delete "aspect of"
p2 L34. Consider inserting a paragraph break after "coverage"
P2 L53. Insert "For Aeolus" at the start of the sentence that begins "There is no suitable resolution"
p3 L64. I suggest deleting the parentheses around "particle"
p3 L6970. I'm not sure I understand what is meant by "well located". Does this mean retrieved at finer resolution? (Perhaps not, because if so, it still doesn't quite make sense. I understand that extinction and backscatter are retrieved simultaneously on the same grid, but the effective resolution of extinction and lidar ratio will always be less than the finest possible backscatter resolution because it takes at least two measurements to determine a derivative.)
p4 L94. Delete "means of"
p4 L96. emitted
P4, L130. Consider adding "irregular grid" to this description. Perhaps here: "in steps of 250m, to produce an irregular grid with a total number"
p5 Figure caption. "Example" rather than "exemplary"
p7 L179. "Extensive use" rather than "Excessive use"
p7 L180. Replace "lightened" with "simplified"
p7 Equations 58. I think the variable names and subscripts could be chosen to better indicate which set of variables are the raw signals and which are the crosstalk corrected signals. It's particularly confusing that the pair with mixed Rayleigh and Mie components are subscripted "ray" and "mie" while the pair where they are actually separate is generic with no mnemonic. Using p and m subscripts for the corrected measurements in 5 and 6 might help. I would also suggest renaming the variables in 7 and 8 to avoid the ray and mie subscripts on the raw channels (although I know that suggestion might be more controversial because it's based on historical precedent with this kind of instrument).
p8 L203 Delete "can be rephrased into". The slash notation is confusing and unnecessary. In this case "are equivalent to" is the better phrase.
p8 L209 Appendix, not Annex
p8 L227, I believe the meaning will be clearer if you remove "to account for" and instead say "because additional noise contributions, such as ..., are likely to dominate over the Poisson noise"
p9 L246, "In theory" instead of "Principally"
p9 L252, "simplify" instead of "lighten"
p910, L261263, I had trouble understanding this sentence until I read the Appendix. It might be clearer to move "scaled" and break the sentence into two: "Here we use the variance measured at 2.9 km resolution, scaled to approximate the noise level in the 87 km bins. This approximation assumes the scene is homogeneous so that all the variability is due to measurement error."
p10 L269 replace "artificially increased" with "increased"
p10 L272275. I found these few sentences very confusing. I think the authors are saying the extinction bias due to underfilled bins tends to decrease the measured range of lidar ratio, but that this is not a compelling reason to reduce the lidar ratio upper bound in the algorithm, so they end up ignoring the bias found by Flament et al. 2017. If the effect is ignored, then these sentences are somewhat of a distraction and could simply be deleted. If the authors feel that it is important to keep these sentences about the bias due to underfilled bins, (and if my interpretation is correct), I think the readability could be improved by (1) signalling the contrasting thought using "On the other hand" (or "In contrast") instead of "Additionally", (2) using "aerosol partially filling a range bin" instead of "different hypothesis on the distribution of aerosol layers within a range bin" (3) specifying "underestimate" instead of "alter" and (4) correcting "copolarized particle backscatter coefficients" to "copolarized measurements". That is, the bias due to underfilled bins is a bias to extinction, not to particle backscatter coefficients.
p10 L288. Forty or forty thousand? Forty seems more likely, in which case three digits after the decimal point for an integer is a strange instance of false precision (should be just "40"). Or if forty thousand, it should be a comma not a point "40,000"  but in that case, 40,000 is a crazily large number of iterations. What is the typical number of iterations actually required for convergence?
p10 L289. "many fewer" instead of "much less"
p11 L305 Consider inserting Equations C2 here. The sentence could read "This relation is inverted to produce Eq. C2", and deleting the terminology "MoorePenrose pseudoinverse" in the main text, since it's more understandable in the Appendix where it is explained more completely.
p12 L339. "optically thin" rather than "thin"
p12 L350. "at the price of" instead of "to the price of"
Figures: I hope Figures 24 will be larger with proportionately larger text. They are quite difficult to read.
P14, L361 "either" not "wither"
P14, L381 "either" not "wither"
P14, L383, I'm not sure what "locks extinction and backscatter to appear colocated" means. Does it mean something like "MLE retrieves extinction and backscatter at the same effective resolution"?
Figure 4 caption. "CALIPSO Feature Mask". Better to use the official CALIPSO product name and be specific: is it the "Vertical Feature Mask (VFM)" or the "Atmospheric Volume Descriptor (AVD)"?
Figure 4 caption, the sentence starting "The rightmost column shows the Calipso Featuremask" is confusing. This could be rearranged to first describe the rows as backscatter, extinction, and lidar ratio. Then in a separate sentence, say the CALIPSO feature mask is repeated on each row.
Figure 5 caption, "error bounds" rather than "errors"
P17 L408409. Delete "and it might be horizontally and vertically". It is disrupted, not might be.
P18 L419. I suggest using "copolarized" rather than "copolar", to be consistent with how you have described it earlier.
P18 L427. Explicitly include copolarized lidar ratio in this sentence: "expected values of copolarized lidar ratio of 80sr  120sr for desert dust". It would be good to avoid the possibility of these expected values being taken out of context and mistaken for the more usual nonpolarized lidar ratio.
P18 L431. Replace "all values" with "expected values" and replace "the estimated error ranges" with "the error ranges estimated from Eqs C2 and C3" (or some other phrase to make it clear that it's the error ranges calculated from the new methodology. They do not fall within the "poisson" error bounds.)
P19 L449. ":Hence, " is not the right connector since it implies causality, and consequently I'm not exactly sure I know what is meant. I think the authors are only explaining what was meant in the first part, so perhaps replace with "; that is, ".
P19 L470. Replace "with this" with "in addition". Again "with this" is usually used idiomatically to indicate a causal relation.
P19 L472474. The flow of this paragraph is interrupted. I suggest moving these sentences about how the coupled retrieval improves backscatter to the earlier spot at L469 immediately after the statement that backscatter is improved along with extinction. The information about the constraints and the anticorrelated noise are separate thoughts.
P20 L483. "suitable" rather than "suited"
P21 L499. "the properties above" doesn't make sense here, since no properties have been defined yet in the Appendix.
P21 L515. Keep "are equivalent to" and delete "can be rephrased into"
P21 L521, I suggest replacing "optical depth/extinction" with "layer optical depth". Dropping "extinction from this sentence is no loss since the relationship with extinction is described in the very next sentence.Citation: https://doi.org/10.5194/amt2021212RC3 
AC3: 'Reply on RC3', Frithjof Ehlers, 01 Oct 2021
General comment:
First thing to note is that we were very glad about the thorough feedback of all Referees, who evidently made the effort to go through the details presented in our manuscript. This helped a lot to refine the text and we hope that we could make our statements more concise now. Especially the method of calculating the uncertainties in the MLE method was addressed by more than one of the Referees and required some changes that we want to present to all of Referees.
The procedure outlined in the last paragraph of Appendix C was relying on the simple equation alpha = gamma * beta with extinction coefficient alpha, lidar ratio gamma and backscatter coefficient beta. Motivated by comparing the obtained variances via (C2) with the variances one would get from a standard error propagation with constant lidar ratio sigma_a^2 = gamma^2 * sigma_b^2, we argued e.g. that the uncertainty in alpha cannot exceed 200 sr times the uncertainty of beta in a single bin, whatever the precise value of gamma is in reality, due to the upper bound in lidar ratio.
Now, after the feedback and some graphical considerations, the problem with this statement is, that it leaves aside any uncertainty or spread of the lidar ratio estimate itself, as it can lie somewhere in between 2 sr and 200 sr and hence should have nonvanishing variance sigma_g^2, causing another term to arise from the view of standard error propagation, increasing the uncertainty of alpha proportional to beta itself. This is why we noticed that the presented approximation was not correct, although we observed uncertainty values that matched approximately the SCA mid bin output for the SAL case for backscatter and extinction coefficients, which seemed realistic at first glance.
Following this, we evaluated the errors from equation (C2) directly against the uncertainties calculated by the SCA standard error propagation as in Flamant et al., 2020, and found good agreement, when the same Poisson noise variances are used in the S_y in (C2), see attached figure.
That means, that with (C2) we are only able to provide uncertainties for the unconstrained problem (as SCA solves it), but especially the soobtained extinction uncertainties are overestimated and not representative of the real uncertainties, as one can see in the comparison of SCA and MLE in the simulation cases.
In lack of a representative alternative for uncertainty estimation we decided to delete the lidar ratio and extinction uncertainties of the MLE solutions in Figures 5 and 6, for now, stating that there is no representative error estimate available yet. The challenge of reporting representative uncertainty estimates is now being stressed in the abstract and conclusion and will require future investigation.
To Referee 3: It would be helpful to get more detailed feedback on the proposed “MonteCarlo” procedure for observational data, because we are curious but did not fully understand the remark you made. However, the error estimates in section 4.3 and 4.4 are not essential for the main purpose of the investigation, because the increased precision of the MLE (and therefore its main advantage) is readily demonstrated in sections 4.14.4 against simulation cases and reference data. Due to the virtually identical forward model, the precision of the MLE is always higher than or on the order of the SCA.
Regarding the comments of Referee 3:
While the remarks on the uncertainty estimation have been very helpful, as described above, we also want to address your second concern on the main cause of the improvements of the MLE. The forward models for SCA and MLE are virtually identical, besides the included variable transforms. That implies that the only qualitative difference between the methods is the constraints. These are now acknowledged as apriori knowledge, making a stricter differentiation between an explicit cost function term and arpiori knowledge in general. Without these constraints, the MLE would lack the physical information to identify signal noise (see equations A7 and A8 in Appendix A) in the first place. In order to illustrate this influence more convincingly, we attached a comparison of real data case 1 processed with and without constraints. This underlines, that the dust plume (SAL) retrieved with MLE without constraints is disrupted similarly to SCA results in all aerosol optical properties. We hope to have made suitable modifications to to clarify this better, please find them outlined in the answers to your comments below. All noted grammar and spelling errors have been adopted without explicit mentioning.
P3 L77. A positivity constraint on quantities where zero is a valid and common expected value will produce a bias. The authors mention this in the context of the SCA algorithm, but this is also a concern in the new algorithm. It might be better not to include the positivity constraint on backscatter and extinction in the new algorithm. Have the authors tried their MLE method without the positivity constraints?
Yes, we tried this. The MLE method without any constraints shows no advantage over the SCA, since the observed signal values can be perfectly fitted to. Without the zeroflooring, the SCA produces oscillatory extinction results as in Figure 10 in https://earth.esa.int/eogateway/documents/20142/37627/AeolusDataInnovationScienceClusterDISCLevel2Auserguide.pdf. Comparably bad solutions are obtained using MLE without constraints.
We added after equation (14):
“It is important to mention that it is only the apriori knowledge in form of the boxconstraints that makes the MLE solution distinct from the algebraic SCA solution (without zero flooring, see section 6.2.2.1 in Flamant et al. 2020), because this algebraic solution corresponds to the exact signal values in y and therefore to J_{obs}=0, which is the global minimum to the unconstrained counterpart of problem (14).”
p10 Box constraints. I think its possible that adding constraints has a smaller impact on improving the results than the other improvements: using raw measurements and uncertaintites in a coupled retrieval.
The forward model equations used by SCA and by MLE are virtually identical and differ only in terms of some variable transforms between optical depth, lidar ratio, backscatter and extinction, see section 3. So it is indeed the apriori knowledge in form of the constraints that causes the improvements.
Additionally, the constraints have potential negative consequencs: (1) that they could lead to bias (similar to what happens in SCA) and (2) that they make propagation of the uncertainty very difficult. Have the authors attempted the MLE retrieval without adding the box constraints? It would be useful to compare results with and without using the box constraints. If they are just as good, eliminating the box constraints would eliminate the two issues mentioned above. If they are not as good, the comparison would give a clearer view of the impact of the different improvements.
This is a good point and has not been made entirely clear in the manuscript either, but as mentioned above: Solving the MLE without boxconstraints is essentially a more complicated way to converge to the algebraic SCA solution (without zeroflooring), which is not favourable.
P19, L469 "This is mainly due". Similar to the previous comment. This statement is made without any analysis to suggest how the different algorithm features dominate the improvements. The authors should assess the impact of the box constraints separately, to support this statement that it is the dominant cause of improved results.
For the above mentioned reasons, the SCA essentially is this benchmark. So, we do not intend to include an analysis of the “MLE without constraints”, as it would just illustrate the oscillatory behaviour outlined in section 6.2.2.1 in Flamant et al. 2017.
P6 Eqs 2 and 3. P21, Eqs A4 and A5. Very little is said about the crosstalk constants C1, C2, C3 and C4. The top of page 7 seems to imply the constants are known, but in fact they are very challenging to determine and have significant uncertainty associated with them. More information is needed about how these are determined and what are typical values for Aeolus. This is also important for the discussion of the correlated errors and for understanding the magnitude of that problem.
We agree, the typical magnitudes for C1...C4 should be added. For the calculation of the calibration coefficients please see the L2A ATBD (Flamant et al. 2020, updated 8th Juli) section 6.5 and the stated reference document RD25 “Generation of AUX_CAL Detailed Processing Model Input/Output data definition” available at https://earth.esa.int/eogateway/documents/20142/37627/AeolusAUXCALIODDDPM.pdf . To add this essential information in the text, we included just above section 3.1:
“The instrument is calibrated with measurements from dedicated instrument calibration modes (Reitebuch et al., 2018a) and the crosstalk coefficients C_{1...4} are determined according to Flamant et al.,2020 and the procedure in Dabas, 2017. At (p,t,f) = (1000 hPa, 300 K, 0 MHz) C_1 and C_4 equal 1 by definition. The other coefficients then typically take values about $C_2 \approx 0.5$ and $C_3 \approx 1.25$. For the rest of this work, we assume the calibration as known and do not include the contribution of calibration errors in the results. The calibration cannot be perfect in reality, but is likewise input to all algorithms, which guarantees a fair comparison of retrieval precision in Sections 4.3 and 4.4.”
We also include in the abstract now:
“The increased precision of MLE with respect to SCA is demonstrated by increased horizontal homogeneity and better agreement with the ground truth on real data cases, though proper uncertainty estimation of MLE results is challenged by the constraints and the accuracy of MLE and SCA retrievals can depend on calibration errors, which have not been considered.”
p1011 L294304, two approaches to error quantification. While in general I like the idea of analytical propagation of errors using sensitivity analysis, the propagation is seriously challenged by the difficulty of representing the impact of the constraint in the propagation equation. I'm not very convinced by this method of rescaling the variance and assuming the correlation is unchanged (Appendix C, near P23 L575). Can it be explained more clearly why the correlation matrix should remain unchanged? Also, using the lidar ratio constraint to (potentially) reduce both the extinction and backscatter error bounds separately gives an impression of circularity. The authors should add rigor and validation to make this method more convincing.
see statement of the author’s
P25 Figure D1. Although the Monte Carlo methodology is discussed, it seems that the results of the Monte Carlo approach are only shown in Appendix D without any analysis or discussion. These should be promoted to main text and properly analyzed.
The uncertainties in the Figures 2 and 3 are from MonteCarlo assessment, which is in the main text already. We now refer to it explicitly with having changed line 295 to say “A MonteCarlo approach as in Xiao et al. 2020 is applied to classify the uncertainties in simulation results in sections 4.1 and 4.2.” We also promoted modified versions of D1 and D2 to the main text in conjunction with a more detailed discussion, see coming answers below.
Both methods are available for the simulation cases, so I'd like to see a comparison of the analytical (Appendix C) method against the Monte Carlo results for the simulation cases, which might bolster my confidence in the analytic method.
see statement of the author’s
It would also be useful to compare the propagated uncertainties with and without the adjustment for the constraints, to see how big this impact is.
Typically, the extinction uncertainty estimate is reduced because the uncertainty of the extinction is usually much higher than 200sr times the uncertainty of the backscatter.
The authors say a Monte Carlo approach is not feasible for the measurement cases, and I think this refers to the fact that they cannot vary the true measurement error. However, it is certainly possible to simulate measurement noise from the measurement error covariance matrix, S_y. Using these simulated measurement errors in a Monte Carlo approach would give an accurate view of how S_y is propagated through the retrieval, more immediately convincing than the method in Appendix C. So, going further, I'd like to see a Monte Carlo propagation of S_y, which could be used to validate or replace the current propagation.
We cannot follow this point well. What we have is the Forward model y = F(x) going from the state space x to measurement space y. Now, what you suggest is to propagate y errors through y = F^(1)(x), which contains the implicit function F^(1). Finding this implicit function locally is the problem we solve timeconsumingly via optimization already. So, if we understand correctly, you suggest to estimate the error via multiple runs of the minimization procedure with artificially added noise from S_y to the already noisy signals S? This artificially increases our noise amplitude and makes it necessary to run the minimizer (at least 20 times) more often, which we find to be unfeasible effort for the operational product.
P17, Figure 5 (also figure 6). I can't understand which error bounds are shown in Figure 5 and Figure 6. The descriptions "lower error bound" and "upper error bound" are confusing because they don't match any description in the methodolgy, and they are possibly inaccurate as well. Another terminology should be seletected, preferably one that reflects the descriptions in the methodology section. The two sets don't seem to be the "two approaches" introduced on P1011. Are they instead two formulations of the analytic propagation of errors using different characterizations of the measurement error S_y? Is the "upper error bound" the propagated uncertainty from Appendix C? (If so, calling it an upper error bound is particularly problematic since the CramerRao inequality characterizes a lower bound, not an upper bound). And how is the "lower error bound (Poisson)" calculated? I don't see that in the manuscript.
see statement of the author’s
If indeed the one labeled "upper error bound" is the one produced in this analysis, and the one labeled "lower error bound" is the standard one for SCA and midbinSCA, then another emergent theme of this analysis is that the uncertainty of the standard algorithms is inaccurate as well. This should be highlighted in the manuscript as another primary impact of the new algorithm.
That is right. Finding the most representative errors in SCA and SCA MB estimates is under current investigation at Meteo France. That means, that the presented Poisson error estimates for SCA and SCA MB are also to be validated. This is why we decided to use signal uncertainties calculated as in Appendix B in the first place for MLE.
P1 L8 "algebraic inversion scheme to a (partly) illposed problem and therefore sensitive to measurement noise". The sentence should probably be reworked. "(partly) illposed" is not really informative, and even a wellposed inversion is sensitive to measurement noise.
This statement was based on the work of Pornsawad et al., 2008: “But From a mathematical point of view Eq. (1) is a Volterra integral equation of the first kind, which is illposed.” We find this wording misleading as well and thus decided to drop “illposed” throughout our text.
P3, L7879. I think the sentence should be split up into multiple sentences that clearly list the factors that led to poor results in the SCA (there are at least two, I think: the correlated noise and the incorrect removal of negative extinctions) and the factors that lead to improvements (coupled retrieval, using the measurement error covariance matrix in the retrieval, box constraints), and to explain their causal relationships. In particular, I believe it is incorrect to suggest that the box constraints are responsible for eliminating the correlated noise issue, as this sentence seems to say. Instead, I think the bias is avoided by using the measurement channels and their error covariance matrix in a coupled retrieval.
We keep this sentence as it is but added the respective reference to the zeroflooring and averaging operations in SCA within Appendix A. Without the boxconstraints, the MLE would not ‘know’ which fraction of any signal value likely corresponds to noise, therefore, in (A7) and (A8) the MLE would be ‘blind’ and not be able to improve anything.
P4, L130. How is the grid spacing chosen? What are the typical values?
The range bin sizes are typically coarser at high altitudes and finer close to the ground but change several times per orbit and area of interest, which makes it difficult to generalize. For better readability we changed this sentence to “Individual, vertical range bin sizes can be independently varied between 250 m and 2000 m in steps of 250 m, while remaining limited to a total number of ACCD rows and hence vertical range bins of 25 per column.”
p7 Eqs 9 and 10. The minus sign in the denominator does not appear in Flament et al 2017. The rest of the equation looks similar but with a few more algebraic steps. I think it is otherwise correct, but I would like the authors to doublecheck to be sure.
Thanks for spotting this error; the minus sign has been corrected and the equations have been double checked.
p7 Eqs. 9 and 10. It would be helpful to have a sentence or two summarizing the derivation of the quoted equations; for example, an explanation that the equation includes an expression that explictly integrates the transmission over an extended vertical bin.
For better reading flow we changed the line above these equations to say: “The following approximations for the range corrected signals (5) and (6) are made by using the mean bin properties from above, see equations (6.35)(6.36) and definitions above in Flamant et al., 2020:” We hope that clarifies? We are sorry but we cannot repeat all derivation steps from Flamant et al. as this would not add value to the manuscript. The message here is, that the identical SCA forward model equations are used (with transformed variables).
p8 near the the end of section 3.1, it would be good to have a sentence or two discussing the anticorrelated noise that's mentioned in the Appendix. This seems like an important point that is referred to both in the introduction and later, so it should be described in the main text with enough detail so a reader knows what it is about.
It is not too important and was purposely placed in the Appendix to compress the main text of the manuscript. The main insight after the algebra in 3.1 should be that the backscatter and extinction depend on both Sray and Smie at the same time.
P9 L232. I'd like to know more about how the box constraint is implemented and specifically whether it may negatively impact results. Could it potentially bias the results (by perhaps producing a disproportionate number of solutions near the boundaries) or affect their precision of the solution (via a variable transform that changes the gradients near the solution and near the boundaries)?
We are no experts in this either, but these details of the LBFGSB algorithm can be found in the work of Nocedal et al. and Zhu et al. (as referenced) on http://users.iems.northwestern.edu/~nocedal/lbfgsb.html, which is underlying the scipy.optimization package.
p9 242. "future developments". This is really more of a response to another reviewer, but I just wanted to say that I support the authors' decision to perform single profile retrievals without including horizontal scene smoothing. It makes sense to explore the simpler solution first. Furthermore, not requiring scene continuity for the algorithm means that scene continuity can be used as a check on algorithm performance.
We thank you for your remark and fully agree with this position, as otherwise no conclusion could be drawn from the comparisons in Figures 5 and 6.
p9 L250 As I mentioned, I believe that using the measurements and their appropriate measurement error covariance matrix is largely responsible for avoiding the bias due to the correlated errors in the crosstalk corrected signals. Do the authors agree? If so, this paragraph might be a good place to highlight that.
On the contrary, we will need to highlight that the apriori information in the forms of the constraints is responsible for the noise suppression capabilities. This we hope to have accounted for with the additional text added according to comment on P3 L77.
p10 L272275. Is there anything in the current algorithm that addresses the bias from partially filled bins? If not, is the current algorithm compatible with the strategy implemented by Flament et al. 2017?
No, we do not follow the strategy of the ICA, that is why SCA is used as a benchmark. While the approach reads promising, the ICA has essentially too many degrees of freedom to yield useful results and is not currently developed any further (see Aeolus L2A user guide in ESA resources) due to its very poor performance. In the data product, the only columns that seem reasonable are basically almost unchanged copies of the SCA results. Consequently, no effort has been made to tailor this implementation to the ICA. Already with the SCA, one could criticize that the effective resolution of extinction coefficients is expected to be coarser than the range bin size itself.
p10 L290292. There appear to be three rather weak arguments here to justify not using an a priori covariance matrix such as called for by OE. The weakest part of the argument, in my opinion, is saying that the algorithm has no prior information or regularization. In fact, the constraint is prior information and does provide regularization. If the authors believe the influence of the constraint is significant (which I think they do, because they attribute the improved results largely to implementing the constraint), then even suggesting that this is a weak constraint would not be justified. Therefore, the authors should acknowledge that the constraint plays the role of prior information in the retreival, but they chose not to cast this constraint in the terms required by optimal estimation, because that would require a different (probabalistic) form for the prior that isn't compatible with the desired form of the constraint. In my opinion, that is a reasonable reason to use constrained MLE rather than OE. (However, the choice leads to difficulties in working out the correct way to calculate the impact of the constraint on the uncertainties, which I discussed in the "Major" section).
Do you rather mean lines 231232? We fully agree with your understanding of the matter. What we meant was that we do not have an explicit term in the cost function. We highlight the importance of the boxconstraints but also added after (14): “[...] with boxconstraints on lidar ratio, which is prior information that cannot be exactly represented by OEM.”
p10 L295. Some clarification would be helpful here about the simulated measurement noise. Is it simulated as gaussian with the variance determined from Appendix B, or is it simulated with various separate error components? Which components are included?
The information on which error components on the signal noise the simulation results include can be found in the beginning of section 4. These noise components are to the best of our knowledge modelled realistically, not just added up variances.
p12 L347349. While I certainly agree that filtering negative results would cause a high bias that is worse in bins with low SNR, I can't clearly follow the explanation in this part of the text and I don't see a particularly clear indicator that they are triggered by the shift from coarser highaltitude bins to finer lowaltitude bins as implied by the text. Can the explanation be clarified?
According to Reviewer 2 we promoted a modified version of FIg. D1 into the main text, which we now refer to for clarity of the statement. You can see the locally highest extinction coefficient biases at 2 and 13 km altitudes in Fig. D1 exactly where the range bin settings change. An exception is the zeroth bin of course, since it is always normalized to zero extinction in SCA.
p12 L354. The indicator from the averaging kernal is a good idea. Is there a quality flag in the data product related to this? It would be nice if this indicator were included in the plots to show which bins are not trustworthy. This would be particularly useful in Figure 4, for instance, where I am curious to know if the bins below the lofted aerosol are reliably retrieved. This curiosity is fueled by the fact that CALIPSO's 532 nm data frequently misses the aerosol at the bottom of attached layers due to attenuation, making them look lofted. The HSRL capability should act to prevent this problem given sufficient SNR, but on the other hand, attenuation at the shorter wavelength of 355 nm would be worse. So, to be sure, it would be good to see some indication from the retrieval algorithm that the lack of aerosol below the apparent plume bottom is reliable.
The values that cannot be retrieved are padded with 1 in the data product. Usually, this regards only the lowermost extinction bin. Regarding the bins below the lofted layer: We do of course see increasing error estimates with increasing distance from the satellite but we fear we cannot say if the layer does only appear to be lofted in both Aeolus and CALIPSO data or if it is lofted for real.
p12 L354. "The extinction bin closest to the ground cannot be well retrieved." Is this true in general, or specific to this case?
True in general, due to the set of the used equations. We added the word ‘generally’ to the text for clarity.
p12 L357 "Otherwise". I'm confused by this sentence, and not sure if I'm confused by content or just the wording. Does "otherwise" indicate the strategy of taking the mean lidar ratio? I agree that taking the ratio of the mean extinction and backscatter is a better strategy than the mean lidar ratio and will give potentially different results, since lidar ratio for small extinction and backscatter is not as reliable as when the SNR is higher. But why is the mean lidar ratio contaminated by the first guess? And how does taking the ratio of mean extinction and backscatter avoid the influence of the first guess?
This overlaps with a point that Reviewer 1 made as well. So the problem is the following: Whenever the retrieved aerosol optical depth L_p is very low, the influence of the lidar ratio estimate on the cost function becomes increasingly insignificant, until L_p becomes zero and no lidar ratio can be provided at all (these are not included in the statistics). Wherever L_p is very low, the lidar ratio has a high error estimate and some may remain close to the first guess, which then seems to act as an ‘implicit a priori’ (see https://doiorg.tudelft.idm.oclc.org/10.1029/JD095iD05p05587, end of section 7). What our paragraph wants to say is that, if lidar ratios were averaged without weights, then the first guess would contaminate the statistics. In the shown simulation case, this would only affect the statistics when only bins with retrieved backscatter coefficients below 5*10^(2) (Mm sr)^(1) are considered.
So eventually we propose to replace the sentence “Otherwise, the first guess of $\gamma_{,p}=60$ sr would contaminate the statistics for MLE, e.g., $\text{mean}(\alpha_p/\beta_{,p}) $ would be biased towards the first guess.” with “[...], in order to disregard the influence of bins with nearly vanishing aerosol optical depth, for which no reliable lidar ratio can be retrieved.”
Additionally, lines 360361 and lines 380381 are omitted, because they are more confusing than helpful.
Figure 2. What is the explanation for the low bias in the median lidar ratio from the MLE for the lowest bins?
The pdf of MLE backscatter is asymmetric here, due to the positivity constraint. Hence, the median is not equal to the mean.
Figure 24 and D1: The line plots are so small that I can't see important information. The data are particularly important below about 2.5 km where there is significant aerosol, but this is a very small portion of the plot and not readable due to its size and the closely spaced grid lines. Specifically, the lowest bin is called out in the text, but it is so small I cannot see the data or error bar for that lowest bin in the line plots. Please include inset boxes to show the line plot data in the lowest 2.5 km, or a second set of line plot figures that show only the lowest 2.5 km, or in some other way improve the visualization of the lowest 2.5 km.
This is a really good point for consideration. So we now adjusted the font size in the mentioned figures to match better with the text and also provide and discuss an overhaul of figure D1 in section 4.1, in which inset boxes make this information on biases and errors accessible.
p14 L362. The reduced uncertainty from the new retrieval compared to the SCA should be discussed and quantified. I think the new retrieval probably produces usefully smaller uncertainty, but the fact that the figures are so tight makes it nearly impossible to see the region where there is significant aerosol below ~2.5 km, and I can't even tell if the error bars are smaller than 100%. A discussion of uncertainty results is just as important as the mean tendency of the profile, because, for example, a profile that "looks right" but is indistiguishable from 0 everywhere due to its uncertainty would be rather useless.
According to the remark above we considered your feedback and added a more detailed discussion in 3.1 by including: “In order to better illustrate this, the bias $(\text{mean}(x^{*})x_\text{true})/x_\text{true}$ and relative error $\text{std}(x^{*})/x_\text{true}$ for all retrievals with respect to the true profile are presented in Fig. D1. Here, the maximum backscatter coefficient bias in the aerosol layer below 2 km is reduced from about 43\% with SCA and SCA MB to 27\% with MLE. This bias seems to be triggered by the refined range bin setting below 2 km. For the origin of this bias we can think of two causes: Firstly, the violation of the hypothesis of uniformly filled bins due to the strong gradient in aerosol concentration and, secondly, the nonlinear way the backscatter coefficient is calculated from $\beta_{,p,i}=Y_i\beta_{m,i}/X_i$, because here $mean(\beta_{,p,i})$ will become biased high increasingly with increasing uncertainty of $X_i$. The relative error in backscatter coefficients is consistently lower for MLE compared to SCA; In the most interesting area below 2 km the relative error in backscatter coefficients reduces to 50\%30\% with MLE compared to 120\%50\% for SCA, while MLE performs only slightly better than SCA MB.”
and
“The MLE retrieves the least biased extinction coefficients over the whole profile with standard errors up to a magnitude smaller than SCA midbin product, see Fig. D1. Retrieved extinction coefficients are all biased high in the area below 2 km, with maximum bias of 500\% for SCA, 110\% for SCA MB and 70\% for MLE. Between 1.5 and 0.5 km altitude, the biases are comparable in magnitude with about 30\%. Concerning the relative extinction errors, an improvement by a factor of about 1.5 to 2 in comparison to SCA and SCA MB is achieved by MLE in the lowermost 2 km. Though the relative error is on the order of 100\% or greater in all cases.”
P14, L366. Is this number a typo? exp(2*0.4) = 0.45 not 0.38.
This has to do with Aeolus viewing angle of about 35 degree off nadir. This means exp(2*0.4/cos(35deg)) = 0.38 from the instruments point of view.
P14, L369379. It seems that it's not just delayed (i.e. slow decay below the cloud) but the cloud is smoothed into the regions both below and above. In other words, it looks like the effective resolution of the extinction is much less than the backscatter, which make sense since it takes at least two measurements to calculate a derivative. Is there any attempt to calculate backscatter on the same coarser resolution as extinction to produce the lidar ratio?
The phrasing of delay regards the curtain plot in row 3 of Fig. 3 in which you can see that the onset of the cloud is positioned at the ground truth for MLE but is delayed by a bin in SCA. We now properly indicate this in the text. The slow decay below the cloud must be a result of the zeroflooring and the noise in SCA, as we cannot see it in SCA MB and MLE. MLE relies on the same forward model and should therefore have identical effective resolution, but due to the limitations of MLE as compared to OEM, we have not determined any effective resolution.
P15, Figure 3. It appears that the bias in the mean backscatter below the cloud is actually worse in the MLE than in the SCA and SCAmidbin results. This should be discussed.
By promoting figure D2 into the main text, we also added some more detailed discussion and included this observation. "A distinct feature of the MLE result is the bias of up to 100 \% between 2 and 8 km altitude, which might be introduced by the positivity constraint in conjunction with high noise. However, the bias in the area of interest below 2 km is found to be the lowest in the MLE case."
P19, L449, what causes missing values in SCA midbin?
These values are flagged out in the operational processing since they are negative. This is the case only in this figure, but has been suppressed in the generation of the endtoend simulation statistics. For future reprocessing campaigns it is planned not to flag negative values anymore.
P19, L458, Lidar ratios aren't shown in the figure for the Real Data Case II, as in other cases. Better to show them, if possible.
The individual lidar ratio estimates on single bins are basically so noisy in the SCA cases, that this would be of no value for the reader (with zeros and negative values also due to this backscatterextinction delay problem), therefore we provide only the calculation. We referred to this noise problem in the text already.
P19, L459460, Is this comparison between copolarized lidar ratio from AEOLUS to total (co and crosspolarized) lidar ratio from Polly? This is not the best option. Since Polly is also sensitive to polarization, wouldn't it be possible to calculate copolarized lidar ratio from Polly for a more direct comparison? In any case, it should be clearly stated what's being compared.
The Polly results provided by Holger Baars do mimic the Aeolus range bin settings and the copolarized lidar ratio that Aeolus sees. We replaced all mentionings of lidar ratio in the text with copolarized lidar ratio, because this was not made clear before.
P19, L461. It seems strange that the quoted uncertainties are from the Poisson method that the authors are hoping to replace. It would be better to quote uncertainties from the method that the authors think are more represenatative.
see statement of the author’s
P 20, Figure 6. Is there an uncertainty bound available for the "ground truth" (which is also a retrieval)?
We are sorry, but this uncertainty has not been provided.
P20, Eqs A2 and A3 are difficult to mentally derive from the previous step. It would be helpful if more intermediate steps were added to make it clearer how the equation is derived. The appendix is a good place to do this.
This line has been edited the same way as in comment on p7 Eqs. 9 and 10.
P22, L538. This point about anticorrelated error is interesting and informative. However, some parts of the discussion are confusing. For instance, Eqs A7 and A8 show that an error spike in one measured channel gets distributed in an anticorrelated way into the corrected signals. But how does it follow that there is correlation (or anticorrelation) in the errors in backscatter and extintion? It seems logical that if the errors in the two corrected channels are anticorrelated, the ratio would tend to be biased low, so the backscatter would tend to be biased low. However, the error in the extinction would not be correlated with it, because extinction derives from gradients in just one of these corrected channels.
What we meant to express here is rather that:
“This means, e.g., that one noisy value in the Mie (or Rayleigh) channel disturbs both backscatter and extinction coefficients. That also implies that if an unphysical, negative value is obtained in one bin for backscatter (or extinction) in the final result, then the value for extinction (or backscatter) is definitely disturbed as well, whether it lies within physical bounds or not.”
This apparent additional information, which lies below the surface so to say, is not used after the SCA has been applied once. The text above now replaces: “Thus, whenever a negative extinction coefficient is found in bin $i$, this indicates a spurious estimate of the backscatter coefficient as well (although it might appear to be well in physical limits) and vice versa.”
P23, L557, I suggest replacing "in spirit" with something more informative, such as "except with an adaptation to account for the impact of the constraint"
We decided to drop this phrase.
P23, L561. It's not true that no a priori knowledge is imposed. The box constraints are a priori knowledge.
Thanks, we changed “apriori knowledge” to “explicit apriori term” to be more concise.
P25 Figure D1. "Bias" and "standard error" should be defined in equations.
This has been promoted into section 4.1. See answer to p14 L362.
Title. The title would be more informative if it contained the word "aerosol". For example: Optimization of Aeolus Aerosol Optical Properties Products by MaximumLikelihood Estimation
This is a good point that was also made by Referee 1, so we aim to change the title to indicate that out investigation does not concern optical properties possessed by Aeolus.
P1 L1. Typically "embarks" is not used this way, and used only for the much narrower circumstance of a person getting on or off a vessel or starting a journey. It could be replaced with "includes"
We decided for “carries”.
p3 L6970. I'm not sure I understand what is meant by "well located". Does this mean retrieved at finer resolution? (Perhaps not, because if so, it still doesn't quite make sense. I understand that extinction and backscatter are retrieved simultaneously on the same grid, but the effective resolution of extinction and lidar ratio will always be less than the finest possible backscatter resolution because it takes at least two measurements to determine a derivative.)
We modified the sentence to “Thus, the particle extinction may occur only where there is backscatter and [...]”
P4, L130. Consider adding "irregular grid" to this description. Perhaps here: "in steps of 250m, to produce an irregular grid with a total number"
We updated this mentioning individually adjustable range bin sizes.
p7 Equations 58. I think the variable names and subscripts could be chosen to better indicate which set of variables are the raw signals and which are the crosstalk corrected signals. It's particularly confusing that the pair with mixed Rayleigh and Mie components are subscripted "ray" and "mie" while the pair where they are actually separate is generic with no mnemonic. Using p and m subscripts for the corrected measurements in 5 and 6 might help. I would also suggest renaming the variables in 7 and 8 to avoid the ray and mie subscripts on the raw channels (although I know that suggestion might be more controversial because it's based on historical precedent with this kind of instrument).
We fear that having inconsistent variable naming conventions compared to the main reference (Algorithm Theoretical Baseline Document, Flamant et al. 2017/2020) would further complicate understanding. This way, it is mostly aligned. The only difference is the lowercase s for signals, so to circumvent confusion with the error covariance matrix S_y (capital).
p8 L227, I believe the meaning will be clearer if you remove "to account for" and instead say "because additional noise contributions, such as ..., are likely to dominate over the Poisson noise"
The other contributions are not likely to dominate over Poisson noise, but rather to increase it noticeably and to smear it out in a quasiGaussian fashion, so we prefer to keep the statement as it is.
p910, L261263, I had trouble understanding this sentence until I read the Appendix. It might be clearer to move "scaled" and break the sentence into two: "Here we use the variance measured at 2.9 km resolution, scaled to approximate the noise level in the 87 km bins. This approximation assumes the scene is homogeneous so that all the variability is due to measurement error."
Shorter sentences are always favourable, so we adapted the sentence.
p10 L272275. I found these few sentences very confusing. I think the authors are saying the extinction bias due to underfilled bins tends to decrease the measured range of lidar ratio, but that this is not a compelling reason to reduce the lidar ratio upper bound in the algorithm, so they end up ignoring the bias found by Flament et al. 2017. If the effect is ignored, then these sentences are somewhat of a distraction and could simply be deleted. If the authors feel that it is important to keep these sentences about the bias due to underfilled bins, (and if my interpretation is correct), I think the readability could be improved by (1) signalling the contrasting thought using "On the other hand" (or "In contrast") instead of "Additionally", (2) using "aerosol partially filling a range bin" instead of "different hypothesis on the distribution of aerosol layers within a range bin" (3) specifying "underestimate" instead of "alter" and (4) correcting "copolarized particle backscatter coefficients" to "copolarized measurements". That is, the bias due to underfilled bins is a bias to extinction, not to particle backscatter coefficients.
We changed the text to: “On the other hand, as shown in Flamant et al. 2017, aerosol partially filling a range bin can easily underestimate the true particle extinction retrieval results by a factor of 16. The applied bounds need to account for these forward model errors by an extra margin.”
p10 L288. Forty or forty thousand? Forty seems more likely, in which case three digits after the decimal point for an integer is a strange instance of false precision (should be just "40"). Or if forty thousand, it should be a comma not a point "40,000"  but in that case, 40,000 is a crazily large number of iterations. What is the typical number of iterations actually required for convergence?
When just one profile is considered, as in problem (14), then the number of iterations is on the order of 50 (roughly 0.5 seconds on an office laptop) to achieve convergence (an average cost function value per bin below 1). In practice, however, we are interested in solving problem (15) for all profiles simultaneously. In this case about 5000 iterations are required (roughly 45 seconds on an office laptop) to let the cost function decrease in the same way for a whole orbit (roughly 450 profiles). This information has been added to the manuscript.
So essentially, solving (15) for the whole orbit saves us time. Though, on second thought there must be a better converging formulation, but we have had no reference on this. We made an attempt to reformulate our model in terms of log(L_p), but the convergence speed did not increase noticeably. It is very likely that a more advantageous scaling exists since the LBFGSB algorithm is not scale invariant. Unfortunately we haven’t found it and are currently tuning parameters to enhance performance.
p11 L305 Consider inserting Equations C2 here. The sentence could read "This relation is inverted to produce Eq. C2", and deleting the terminology "MoorePenrose pseudoinverse" in the main text, since it's more understandable in the Appendix where it is explained more completely.
We dropped the phrase regarding the MoorePenrose pseudoinverse and changed it into a reference to equation C2.
p12 L339. "optically thin" rather than "thin"
Adopted.
Figures: I hope Figures 24 will be larger with proportionately larger text. They are quite difficult to read.
This has been taken care of.
P14, L383, I'm not sure what "locks extinction and backscatter to appear colocated" means. Does it mean something like "MLE retrieves extinction and backscatter at the same effective resolution"?
We changed it into “forces extinction and backscatter to appear colocated”, i.e. to appear together in the same range bin.
Figure 4 caption. "CALIPSO Feature Mask". Better to use the official CALIPSO product name and be specific: is it the "Vertical Feature Mask (VFM)" or the "Atmospheric Volume Descriptor (AVD)"?
We are sorry for the sloppy notation, we meant the vertical feature mask (VFM) and corrected the mentionings thereof in the text.
P17 L408409. Delete "and it might be horizontally and vertically". It is disrupted, not might be.
Done.
P18 L419. I suggest using "copolarized" rather than "copolar", to be consistent with how you have described it earlier.
This is now corrected for.
P18 L427. Explicitly include copolarized lidar ratio in this sentence: "expected values of copolarized lidar ratio of 80sr  120sr for desert dust". It would be good to avoid the possibility of these expected values being taken out of context and mistaken for the more usual nonpolarized lidar ratio.
This is now corrected for.
P18 L431. Replace "all values" with "expected values" and replace "the estimated error ranges" with "the error ranges estimated from Eqs C2 and C3" (or some other phrase to make it clear that it's the error ranges calculated from the new methodology. They do not fall within the "poisson" error bounds.)
The MLE error estimates have been deleted from the figure and the text.
P19 L449. ":Hence, " is not the right connector since it implies causality, and consequently I'm not exactly sure I know what is meant. I think the authors are only explaining what was meant in the first part, so perhaps replace with "; that is, ".
Is corrected for.
P19 L470. Replace "with this" with "in addition". Again "with this" is usually used idiomatically to indicate a causal relation.
Here we actually mean to indicate a causal relation, because it is only by the apriori knowledge of constraints in the first place that noise can be detected and suppressed in a second step.
P19 L472474. The flow of this paragraph is interrupted. I suggest moving these sentences about how the coupled retrieval improves backscatter to the earlier spot at L469 immediately after the statement that backscatter is improved along with extinction. The information about the constraints and the anticorrelated noise are separate thoughts.
We agree and moved the sentence on the reasons for the improvements to the end of the paragraph.
P21 L499. "the properties above" doesn't make sense here, since no properties have been defined yet in the Appendix.
Has been changed to “the properties in section 3”.
P21 L521, I suggest replacing "optical depth/extinction" with "layer optical depth". Dropping "extinction from this sentence is no loss since the relationship with extinction is described in the very next sentence.
Was adapted.

AC3: 'Reply on RC3', Frithjof Ehlers, 01 Oct 2021