the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
Intercomparison of in situ measurements of ambient NH3: instrument performance and application under field conditions
Augustinus J. C. Berkhout
Nicholas Cowan
Sabine Crunaire
Enrico Dammers
Volker Ebert
Vincent Gaudion
Marty Haaima
Christoph Häni
Lewis John
Matthew R. Jones
Bjorn Kamps
John Kentisbeer
Thomas Kupper
Sarah R. Leeson
Daiana Leuenberger
Nils O. B. Lüttschwager
Ulla Makkonen
Nicholas A. Martin
David Missler
Duncan Mounsor
Albrecht Neftel
Chad Nelson
Eiko Nemitz
Rutger Oudwater
Celine Pascale
Jean-Eudes Petit
Andrea Pogany
Nathalie Redon
Jörg Sintermann
Amy Stephens
Mark A. Sutton
Yuk S. Tang
Rens Zijlmans
Christine F. Braban
Bernhard Niederhauser
Download
- Final revised paper (published on 22 Nov 2022)
- Supplement to the final revised paper
- Preprint (discussion started on 30 Mar 2022)
- Supplement to the preprint
Interactive discussion
Status: closed
-
RC1: 'Comment on amt-2022-107', Anonymous Referee #1, 21 Apr 2022
Review of “In-situ measurements of NH3: instrument performance and applicability” by Marsailidh M. Twigg et al.
Summary: Ammonia (NH3) is not regularly monitored in networks. It is also an unregulated pollutant in many countries. Identifying a standard among the available technology for routine gas-phase ammonia monitoring is vital to our understanding of this pollutant and how future regulation policy can be shaped. This paper describes a comprehensive intercomparison of 13 instruments for sampling gas-phase NH3. The study is done in-situ at an agricultural field site in Scotland. The instruments represent a variety of currently available technologies for measuring gas-phase NH3. The authors perform a comprehensive analysis of the available data and show that variability in the ensemble is within 20%. This alone is an interesting finding from an intercomparison of 13 independent NH3 instruments of varying techniques and time responses. While there are still nuances of instrument setup, maintenance, and operations to be determined, the observations reported in this work are a step forward towards developing standardized practices for NH3 monitoring.
The paper presents findings from a new field study and the topic is highly relevant towards addressing current air quality and climate concerns. Thus, it fits the scope of AMT. I recommend this paper for publication in AMT following minor revisions.
General Comments:
In general, the flow of the paper could be improved by moving around some of the sections and refocusing the key points in the introduction to match the outcomes. For example, the introduction of this paper implies that the authors will provide recommendations on how to achieve high quality future routine monitoring of NH3. However, the conclusions do not provide specific recommendations for what the optimal inlet setup and operating/maintenance procedures could be for a monitoring site. In contrast, a key conclusion of this work that is not identified earlier in the paper is that the variability in the ensemble of 13 instruments is within 20%. Such a tight cluster of NH3 measurements from 13 independent instruments actually seems pretty good, especially considering that gas-phase NH3 can be challenging to measure. It would help to contextualized this finding better in the text in terms of what this could mean for monitoring networks comprised of a few types of NH3 instrument techniques.
There are some very detailed instrument descriptions in the methods section and some that are rather vague. It would be helpful to include a similar level of detail for all instruments, especially in cases where references are not available. Please also clarify the purpose and design of the experiments using identical instruments with different inlets earlier in the paper. This is an important factor in the final recommendation, but there is little information leading up to these results to prepare the reader to understand these findings.
Please include more context in the methods section about how the calibration sources were used in these experiments and what instruments they were used with. Were these calibrations only applied in the field, or were instruments calibrated individually in a laboratory before/after the field experiment? Please clarify in the text. It could also help the flow of the paper to move up “Section 3.9 Ammonia calibration system” to between Sections 3.2 and 3.3.
Zeros were routinely performed for the QCLAS, but what about the other instruments? Were span calibrations also performed? Are zeros and span calibrations stable over time? Were corrections associated with these zeros and spans applied to the data prior to analysis in Section 3? Please clarify in the text.
Instruments were in the field from 22 Aug to 2 Sept. Is there a reason that only 23 Aug and 29 Aug is used for the intercomparison study and not the full period? There are inconsistences in the dates in several places throughout. For example, Figures 8 and 9 and Table 4 suggests a start date of 22 Aug instead of 23 Aug. There are also inconsistencies in the dates in some of the figure captions and table headers in the SI.
A major concern for Section 3 is that the bulk of the analysis relies on comparing each instrument to an ensemble median rather than a true, validated reference (e.g., Figures 8 and 9). While this might be the best approach for this case study, it should be specified in the text that the median is not an independent variable (i.e., it is by definition driven by the range of observed values from the different instruments). If the ensemble median is not independent, then the correlation analysis might be better described using 2-sided linear least squares fits (aka. orthogonal distance regression). See additional specific comments below.
Specific Comments:
Title: Could the title be more specific to the experiments and findings? Maybe something like “Exploring best practices for gas-phase NH3 monitoring: An in-situ comparison of 13 instruments in Southeast Scotland”.
Abstract: The Abstract could be streamlined a bit by relocating some of the context to the Introduction. It could also be added that NH3 is an unregulated pollutant in many countries. Listing each instrument make and model could be solely addressed in the methods section.
P3, L15. Flip phrase to read as “not all NH3 is captured”
P3, L35. There are many similarities between this work and a prior report by von Bobrutzki et al., 2010, which is referenced several times throughout this work. From the text it seems that the main advantages of this intercomparison are the addition of newer instruments/technology and the evaluation of traceable gas standards. Are there other advancements in this work in terms of the experiment objectives, experiment design, and application of lessons learned from von Bobrutzki? For example, this work uses pairs of identical instruments outfitted with different inlets to characterize artefacts due to the setup itself. This is a unique feature of this experiment that ought to be clearly outlined as a focus of this paper in the intro. The experiments related to this comparison and any additional setup should be clearly described in the methods section.
Figure 1. This is a nice photo of the field site. It gives a lot of perspective for the experiments. It would help to add a wind rose or an arrow to show the predominant wind direction during the study period.
Table 1 is missing accuracy, precision, and range information for the miniDOAS #1 - add symbols like n/a or (-) unknown. The LGR #1 and LGR #2 response times look to be flipped. The LGR#2 with the higher flow rate should have a faster response time of 1 s. There are 13 instruments compared in this study, yet there are 15 rows in the table. Should the OGS and the ALPHA sensors be separated from this table or distinguished in some way. Maybe this is simply fixed with a footnote to clarify the usage of the OGS and ALPHA samplers. It would also be helpful to further explain the dependency of the Picarro#2 and the OGS in a footnote in the table. Further, the acronym OGS has not been defined yet in the manuscript.
Table 2. Add another row to specify if the inlet components are heated or not. For instruments that do have heated inlets, at what temperature are they maintained? For consistency, change “N” to “No” for the AiRRmonia#1 filter. Is “diameter” meant to be the inner diameter of the tubing or the nominal outer diameter? The i.d. of the tubing will be the most relevant for your residence time calculations, so that could be the better parameter to include in this table. In either case, please clarify.
Is there a reason that the manifold inlet and manifold itself are made from different materials? Is there any research on how NH3 sticks to uncoated Pyrex surfaces?
I find it interesting that all of the instruments on the common inlet use a filter. Was that planned?
Some additional explanation of the parameters for the unique inlet system associated with the QCLAS are likely needed in Table 2 and Section 2.2.3. This instrument setup is unique in that it uses a heated inertial inlet with a critical orifice to separate particles from the airstream. It also requires a rather large capacity scroll pump to create a sample flow rate of 13 l min-1. It is also important to distinguish the size of the critical orifice (~1 mm) located inside the inertial inlet compared to the size of the tubing (typically 3/8” o.d., ¼” i.d.).
P8, L9: How often are the passive ALPHA samplers collected and analyzed? Please include this information in the description in the methods section.
P8, L34: It isn't clear here why the correction was not included. I think this is explained later in section 2.3.1. It would help to add a reference here to the explaination in the later section.
P8, L37. I suggest moving the last sentence of this paragraph to Section 2.3.1 where the OGS is described. This information gets lost here and is better served in the other section.
P9, L16: Background subtraction from routinely measured zeros seems to be another unique feature of this instrument’s operations. Were any other instruments routinely zeroed throughout the measurement campaign? It seems like this would be a fairly important step for all instruments to accurately report NH3 mixing ratios. It is also interesting that N2 was used for the zeros instead of zero air. While using N2 should not impact a zero calibration, a prior study showed that span calibrations on top of N2 compared to zero air produced a spectroscopic artefact up to 10% (Pollack et al., 2019; https://amt.copernicus.org/articles/12/3717/2019/).
Section 2.3.1: What wavelength does the OGS operate at? What is the uncertainty of the absorption cross section (or line strength) at this wavelength? How does this impact the overall uncertainty of the OGS calibration system?
P11, L36: “associated with long stabilization times”
Section 2.3.2: Please specify the carrier gas used. N2 versus air can have different effects on NH3 permeation devices and spectroscopic artefacts.
Figure 2. Please add something (e.g., shaded bar, horizontal arrow, red dashed lines) to indicate the period of data used for the instrument intercomparison (which I think is 23 Aug to 29 Aug). Is there a reason that the remaining data in the timeseries up to 02 September was not included in the intercomparison?
Figures 3 and 5 make me wonder how the instruments were time aligned prior to correlation analysis. Were inlet delays account for prior to averaging to 1 hour? Please clarify in the text.
Figure S2. This is an interesting figure. It seems as if there was less atmospheric stability during the intercomparison period, yet much larger stability during the period immediately after the study period (29 Aug thru 2 Sept). How do the instruments compare during the more stable period? Are there any observable differences? This would be an interesting comparison to add to this study that would be within the scope of this study. Also, there is a typo in the caption; the date range should be from 22/08/2016 to 03/09/2016.
Figure 4 could be moved to the SI if you feel this paper is too lengthy.
Section 3.4: Can you comment on the effects of a heated vs. unheated inlet? See Ellis et al., 2010 for reference. Some additional comments about the utility of a heated inertial inlet for filter less separation of particles could be included here.
Section 3.3: Does the ensemble median include LGR#1? If yes, how does the ensemble median and related statics change if this measurement is excluded?
Figure 6. Change CV limit to 20% in caption to be consistent with figure and discussion text.
P18, L13: Is the response time truly different under ambient conditions? Without the same level of fine structure in Figure 7b as in Figure 7a, it is difficult to accept the levels of smoothing applied to the DOAS under ambient conditions to match the profiles of the AiRRmonia and Picarro instruments. Can you include another trace in 7b to show the DOAS signal with the same level of smoothing as in 7a? This would help highlight whether additional smoothing of the DOAS is indeed needed to match the features of the AiRRmonia and Picarro instruments under ambient conditions.
Figure 8. The met filter looks like it could have induced some bias in some of the fits. The met filter was applied to eliminate low wind speed and unstable conditions that could have led to inhomogeneity between the inlets at the field site. But didn’t the ALPHA samplers indicate that there was homogeneity during the study period? What do the fits look like without the met filter? At a minimum, you should comment in the text about any differences in fits with and without the met filter applied.
Section 3.6: What would you consider to be a reasonable deviation from 1 for the slopes? How does this compare to the deviation in the ensemble or with the reported measurements uncertainties in Table 1? For example, if the spread around the ensemble median is 20%, would slopes ranging from 0.80 to 1.20 be considered good?
It would also be interesting to see how the fast instruments compare with an ensemble average of only the fast instruments (like on a 1 min average timescale). Was this something you tested? Can you comment on whether the results would be different?
Data in Figures 8 and 9 are split into NH3 > 10 ppb and NH3 < 10 ppb. This was done to be able to best compare the NH3 < 10 ppb data with the findings in von Bobrutzki. Please include additional discussion of how this work compared with the findings in the prior work. Does separating the data points by NH3 mixing ratio prior to intercomparison analysis generate any bias? It would be helpful to see the results of the intercomparison fits using all data, which could be a nice figure in general to include in the SI.
At the risk of cluttering the figure, it also seems appropriate to also include the actual fits in each plot.
P24, L1: Are the least squares regressions 1-sided or 2-sided? Please specify. See general comment above.
P24, L6: It seems as if the LGR#1 instrument was having some issues during this experiment. Would it be better to exclude the data from LGR#1 from the paper altogether? Do you have specific reasons for keeping it in this intercomparison? Was the LGR#1 was included in the ensemble median? Please clarify your reasoning in the text.
Figure 10. What does the color scale correspond to? Please clarify in the caption. Add labels “a” and “b” to plots.
P26, L15: What the was the temperature of the LGR#2 inlet? Was it high enough to thermally dissociate enough NH4NO3 to impact the measurement? Gentle heating (<40 degC) might not have a huge impact on the measurement on the sampling timescale (e.g., Fig 4a in Huffman et al., 2009; https://doi.org/10.5194/acp-9-7161-2009).
Figure 11. Since the instruments compared in this figure all have relatively fast response times (1 minute or less), would it be more realistic to compare them on a 1-minute timescale?
Figure 12. Please clarify what height means in the caption. Does it have units? It would be helpful as a quick reference to include labels in the figure about what the different grouping are. You could easily add this by changing the outline colors to match a legend or a description in the caption?
P29, L11: Do you mean Figure S4?
Table 4: Fix table header to correspond with proper figures in the SI (S4 and S5). Add a footnote to highlight the relationship between Picarro#2 and OGS.
P34,L17: Based on lessons learned from this study, can you provide a recommendation for how often routine calibrations (zeros and spans) ought to be performed if one or more of these measurement techniques are used at a surface monitoring network site?
Section 4.3: Based on lessons learned in this study, can you provide additional recommendations about inlet setup? It is not surprising that instrument manufacturers do not specify a schedule for calibrating and servicing, as it is largely dependent on how, where, and under what conditions the instrument is utilized. Instead, there could be a list of indicators to watch for that could signal a user to perform routine maintenance. For example, a prior works (Pollack et al., 2019; Ellis et al, 2010) showed that increases in the response time to a step change in NH3 from a calibration source was a good indicator of when the instrument inlet needed to be cleaned.
Figure 15. There are two panel b’s in this figure.
Figure S1. There are inconsistencies in the labels used in the caption (a, i, 2, 3). The time resolutions in brackets are missing for some instruments in the figure. What do you mean by “raw” data in the timeseries? “Raw” typically implies uncalibrated data. Please clarify.
Table S3. Header has a reference error. The dates in the period do not make sense. Are these response times truly meant to be in units of minutes, or should they be seconds? E.g., 100 minutes for a 1/e time for the Picarro seems rather long.
Figure S3. Were the datapoints associated with breakthrough eliminated from any intercomparisons with the MARGA? Please clarify.
Citation: https://doi.org/10.5194/amt-2022-107-RC1 - AC1: 'Reply on RC1', Marsailidh Twigg, 07 Sep 2022
-
RC2: 'Comment on amt-2022-107', Anonymous Referee #2, 24 May 2022
GENERAL COMMENTS
The manuscript presented by Twigg et al. represents an important study towards improving the quality of in-situ ammonia (NH3) measurements. Significant challenges still exist in obtaining accurate and precise NH3 measurements, specifically in field observations where a large range of ambient conditions impact the performance of instrumentation. So far, there have been only a limited number of studies comparing different NH3 measurement methodologies under field conditions.
The authors present a comparison of 13 online instruments, spanning over the most important measurement techniques, and 1 type of passive samplers, making it to my knowledge the broadest NH3 intercomparison study conducted up to date. The authors use different statistical methods to compare the instruments determining their precision and accuracy. The comparison of closed-path systems with the open-path miniDOAS as well as employing some instruments of the same type with different inlet systems, allows the authors to investigate dampening effects caused by inlet tubing surfaces. Finally, the authors also give advice on what future measurement setup should consider such as regular calibrations (which were missing in this study).
The setup of the study, analysis of data and presentation of results was performed with great care, considering the major challenges of each instrument. Given the importance and uniqueness of the study, I suggest it to be published in AMT after addressing the comments below. The major parts where the manuscript in my view needs improvement is (1) the determination of the residence time and (2) the description of inlet tubing effects and the derivation of time constants.
SPECIFIC COMMENTS
P. 1, L. 35: The CEN EN 17346 reference protocol for passive sampling is from the year 2020, therefore, I suggest to write “first developed in the 1990s” or similar.
P. 2, L. 13/14: Other studies, which investigate ammonia emissions on PM2.5 formation on a global scale include Gu et al. (2021) and Pozzer et al. (2017)
P. 4, L. 19 ff (2.2. Instrumentation): I suggest to give similar details in the instrument description (where possible). Sometimes performance indicators (like detection limits) are given. 2.2.7 is the only section where the instrument operators are named. Providing wavenumbers for the spectroscopic techniques in 2.2.4 and 2.2.5 would be beneficial.
P. 4, L. 23: Do you mean ½” OD? In Table 2, the tubing diameter is given as 9 mm, which is less than ½”. In Table 2, I suggest to specify that the value represents the inner diameter.
P. 4, L. 24 ff & Table 2: When calculating the residence time from the flow rate and tubing dimensions, the pressure drop along the tubing and respective volume flow increase needs to be taken into account. Looking at some of the values in Table 2, is seems a constant volume flow was used for the calculation, is that correct? If so, values would need to be adjusted for the pressure drop, which would lead to faster residence times (assuming that given flow rates are at STP).
The residence time of the manifold (+inlet) is given as 1.8 s, however, in Table 2 both add up to 1.62 s. Is that correct? In Table 2, it should be specified if flow rate was at STP or different temp/pressure conditions.
P. 13, L. 13: I suggest to state the sampling interval (of the 282 ppb) to be clear that this is based on the original sampling interval and not on the 1 hr resolution (as in Table S1).
P. 17, Figure 6 caption: NH3 mixing ratios is the median value (as reported above)? If so, it should be clarified.
P. 17, L. 21 ff: The response time due to inlet effects was determined by low pass filtering the respective instrument time series to match the unattenuated miniDOAS time series. While this approach seems sound, the retrieved response times seem very high (e-folding time ranging to more than 2 hrs). Another way to determine the response time is by investigating the exponential decay/rise during calibrations (when zero or span is applied). From Figure 13 a) the OGS (i.e. Picarro #2), seems to have a much lower e-folding time. E.g. the 63% (1-1/e) increase at the step change from 0 to 10 ppb (which would represent ambient conditions) seem to be reached within 1 min as opposed to the 49.5 min with the method used here.
The dampening effects of ammonia in inlet systems is better described by a double exponential function with a fast time constant that represents the air exchange in the inlet tubing and a slower time constant that describes the adsorption and desorption effects (Ellis et al., 2010). E.g. Moravek et al. (2019) showed the evolution of the dampening over time using the double exponential function.
Even if a single exponential decay/increase is used (-> the low pass filtering method yields a single time constant), the authors should include the time constant values from the calibration measurements, also as comparison with other studies. Also, it should be explained why the time constants of both methods (i.e. low pass filtering and visual fitting to miniDOAS time series) would be so different.
P. 18, L. 10-12: Why was it discounted or what is your underlying hypothesis regarding the influence of a turbulent flow regime? I would have thought the opposite: if laminar flow conditions are increasing the dampening (due to segregation along the tubing cross section), this would explain why the Picarro #1 does not perform as well as the Picarro #2 (a similar statement is expressed by the surface/volume ratio).
P. 21, Figure 8: The y-axis offsets from the linear regressions are used to describe the accuracy (i.e. over- or underestimation) of the respective instruments at low mixing ratios. However, the offset may also result in uncertainties in the linear fit (lowest ensemble median mixing ratios are just under 2 ppb). Was this taken into account?
P. 32, L. 24 ff: The authors use the results of the linear regression between instruments to described the precision. While the R2 value is influenced by the instruments’ precisions, the slope would rather indicate the accuracy of an instrument in comparison to the ensemble median. I suggest to make clear that the precision only refers to the R2 value and not the slope.
P. 34, L. 25 ff: Next to the tubing material, contamination of the tubing surface over time can influence the time response significantly (e.g. Moravek et al., 2019). Although the experiment was probably not long enough for it to have a major influence, this point may be included in the discussion.
P. 34, L. 34 ff: Next to avoiding condensation, inlet line heating was shown to improve the time response in previous studies (e.g. Ellis et al., 2010). This should be mentioned here as well.
P. 35, L. 13 ff: Ideally, the humidity in the zero air would match the ambient air humidity levels. One way to produce zero air is by removing NH3 from ambient air through a heating catalyst (without a drying cartridge or similar). Assuming that the water vapor is conserved, the zero air would then have similar humidity levels than the ambient air. Next to its influence on spectroscopy, humidity levels can also affect the adsorption/desorption processes in the inlet line. Therefore, having humidified zero air (at least to some degree) would be a beneficial for all instruments with an inlet line.
TECHNICAL COMMENTS
P. 7, Table 2: Typo in “AiRRmonia #1” & “Operated with a filter” -> “No”
P. 7, L. 14: “are sampled” instead of “is sampled”
P. 7, L. 15: do you mean “through which air is drawn and …”?
P. 7, L. 27: (“Erisman, 2001)”
P. 22, L. 7: Insert space before “ppb”.
P. 25, Figure 10: “a)” and “b)” missing
P. 33, L. 13: Move comma to at the end of subclause (after parentheses in L. 14).
REFERENCES
Ellis, R. a., Murphy, J. G., Pattey, E., van Haarlem, R., O’Brien, J. M., and Herndon, S. C.: Characterizing a Quantum Cascade Tunable Infrared Laser Differential Absorption Spectrometer (QC-TILDAS) for measurements of atmospheric ammonia, Atmos. Meas. Tech., 3, 397–406, https://doi.org/10.5194/amt-3-397-2010, 2010.
Gu, B., Zhang, L., Dingenen, R. Van, Vieno, M., Grinsven, H. J. Van, Zhang, X., Zhang, S., Chen, Y., Wang, S., Ren, C., Rao, S., Holland, M., Winiwarter, W., Chen, D., Xu, J., and Sutton, M. A.: Abating ammonia is more cost-effective than nitrogen oxides for mitigating PM2.5 air pollution, Science (80-. )., 374, 758–762, https://doi.org/10.1126/science.abf8623, 2021.
Moravek, A., Singh, S., Pattey, E., Pelletier, L., and Murphy, J. G.: Measurements and quality control of ammonia eddy covariance fluxes: a new strategy for high-frequency attenuation correction, Atmos. Meas. Tech., 12, 6059–6078, https://doi.org/10.5194/amt-12-6059-2019, 2019.
Pozzer, A., Tsimpidi, A. P., Karydis, V. a., de Meij, A., and Lelieveld, J.: Impact of agricultural emission reductions on fine-particulate matter and public health, Atmos. Chem. Phys., 17, 12813–12826, https://doi.org/10.5194/acp-17-12813-2017, 2017.
Citation: https://doi.org/10.5194/amt-2022-107-RC2 - AC2: 'Reply on RC2', Marsailidh Twigg, 07 Sep 2022