Global Cloud Property Models for Real Time Triage Onboard Visible-Shortwave Infrared Spectrometers

New methods for optimizing data storage and transmission are required as orbital imaging spectrometers collect 10 ever-larger data volumes due to increases in optical efficiency and resolution. In Earth surface investigations, storage and downlink volumes are the most important bottleneck in the mission’s total data yield. Excising cloud-contaminated data onboard, during acquisition, can increase the value of downlinked data and significantly improve the overall science performance of the mission. Threshold-based screening algorithms can operate at the acquisition rate of the instrument but require accurate and comprehensive predictions of cloud and surface brightness. To date, the community lacks a 15 comprehensive analysis of global data to provide appropriate thresholds for screening clouds or to predict performance. Moreover, prior cloud screening studies have used universal screening criteria that do not account for the unique surface and cloud properties at different locations. To address this gap, we analyzed the Hyperion imaging spectrometer’s historical archive of global Earth reflectance data. We selected a diverse subset spanning space (with tropical, midlatitude, arctic, and Antarctic latitudes), time (2005-2017), and wavelength (400 – 2500 nm) to assure that the distributions of cloud data are representative 20 of all cases. We fit models of cloud reflectance properties gathered from the subset to predict locally and globally applicable thresholds. The distributions relate cloud reflectance properties to various surface types (land, water, and snow) and latitudinal zones. We find that taking location into account can significantly improve the efficiency of onboard cloud screening methods. Models based on this dataset will be used to screen clouds onboard orbital imaging spectrometers, effectively doubling the volume of usable science data per downlink. Models based on this dataset will be used to screen clouds onboard NASA’s 25 forthcoming mission, the Earth Mineral Dust Source InvesTigation (EMIT).

Abstract. New methods for optimizing data storage and transmission are required as orbital imaging spectrometers collect ever-larger data volumes due to increases in optical efficiency and resolution. In Earth surface investigations, storage and downlink volumes are the most important bottleneck in the mission's total data yield. Excising cloudcontaminated data on board, during acquisition, can increase the value of downlinked data and significantly improve the overall science performance of the mission. Threshold-based screening algorithms can operate at the acquisition rate of the instrument but require accurate and comprehensive predictions of cloud and surface brightness. To date, the community lacks a comprehensive analysis of global data to provide appropriate thresholds for screening clouds or to predict performance. Moreover, prior cloud-screening studies have used universal screening criteria that do not account for the unique surface and cloud properties at different locations. To address this gap, we analyzed the Hyperion imaging spectrometer's historical archive of global Earth reflectance data. We selected a diverse subset spanning space (with tropical, midlatitude, Arctic, and Antarctic latitudes), time (2005)(2006)(2007)(2008)(2009)(2010)(2011)(2012)(2013)(2014)(2015)(2016)(2017), and wavelength (400-2500 nm) to assure that the distributions of cloud data are representative of all cases. We fit models of cloud reflectance properties gathered from the subset to predict locally and globally applicable thresholds. The distributions relate cloud reflectance properties to various surface types (land, water, and snow) and latitudinal zones. We find that taking location into account can significantly improve the efficiency of onboard cloud-screening methods. Models based on this dataset will be used to screen clouds on board orbital imaging spectrometers, effectively doubling the volume of usable science data per downlink. Models based on this dataset will be used to screen clouds on board NASA's forthcoming mission, the Earth Mineral Dust Source Investigation (EMIT).

Introduction
Imaging spectrometers, also known as hyperspectral imagers, collect images in the form of three-dimensional data cubes: a two-dimensional image of the target surface in the field of view and swath of the instrument with a continuous spectrum in the third dimension. With the recent decommissioning of Hyperion, an early imaging spectrometer on board NASA's Earth Observing-1 (EO-1), many space agencies are now planning or operating a new generation of orbital imaging spectrometer missions, e.g., HISUI (Tachikawa et al., 2012), EMIT , EnMAP (Guanter et al., 2015), CHIME, PRISMA, DESIS, and TRUTHS. At the time of this writing, NASA is studying dramatically enhanced imaging spectrometer architectures to provide measurements with global coverage (National Academies of Sciences, Engineering, and Medicine, 2018). As instrument capabilities grow, the swath width and length of these instruments permit growing coverage areas. Due to the high-dimensional nature of these datasets, the duty cycle of these instruments is expected to be limited by data volume. They are operated with M. W. Sandford et al.: Global cloud property models a store-and-forward mode (Williams et al., 2002), where the data are stored on board in a limited "flight recorder" and transmitted when a ground station is within view or when a manual transfer occurs. This limits the bandwidth from the satellite to the ground and ultimately the total data yield of the mission. Consequently, optimizing the downlink from orbital remote sensing spacecraft can increase the science productivity of these missions. One promising strategy to reduce data volumes involves analyzing images on board as they are being collected, excising contaminated or irrelevant scenes and preserving good-quality data for preferential storage and downlink (e.g., Thompson et al., 2014;Doggett et al., 2006;Altinok et al., 2016).
Clouds are the most promising target for onboard screening since they are a common yet unpredictable contaminant that prevents direct observations of surface features. Previous studies indicate that clouds account for over half of the annual sky cover globally (Mercury et al., 2012;Eastman et al., 2011;King et al., 2013;Mace et al., 2009;Rossow and Schiffer, 1999). Thus, onboard cloud screening could approximately double the science productivity per downlink without changing the total stored or transmitted data volumes. Doggett et al. (2006) on board EO-1 used expert decision tree and support vector machine learning for onboard cloud classification. Altinok et al. (2016) on board IPEX and Wagstaff et al. (2018) on board EO-1 used random decision forest machine learning for onboard cloud classification. Though these were not continuous cloud-screening tools, they have determined that cloud screening is in fact a viable solution to the data reduction problem. NASA's EMIT mission has since baselined this capability  and will be the first imaging spectrometer to use continuous onboard cloud screening operationally. These newgeneration missions will build on a long history of cloudscreening algorithms in ground data systems. Most of these screening algorithms treat cloud detection as a classification problem, where attributes of the instrument data are used to determine whether clouds are present. They use a wide variety of techniques including spectroscopically estimated atmospheric properties (Taylor et al., 2016), band-specific or regionally specific threshold tests (Ackerman et al., 1998), spatial variability (Martins et al., 2002), reflectance models of surface and atmosphere (Gómez-Chova et al., 2007), or data-driven machine learning methods (Yhann and Simpson, 1995;Wang et al., 2020). Onboard algorithms carry special requirements: they must be simple, for encoding into instrument hardware or low-power spaceflight computers, and they must use only data available to the spacecraft at the time of acquisition. Finally, because the screened data are irrevocably lost, its behavior should be transparent to the operators and tunable to be more aggressive or conservative depending on their error tolerance. In addition, the algorithm's statistical properties and error rates should be well understood.
Most prior screening approaches, including those designed for onboard use, strive for global performance -a cloud clas-sifier that works the same way for all locations globally. However, the nature of the classification problem changes depending on location. Neither the coverage nor the spectral appearance of clouds is uniform. Clouds cover more of the oceans (68 %-72 %) on Earth annually than land (54 %-58 %), and tropical regions are exceptionally cloudy (Mercury et al., 2012;Eastman et al., 2011;King et al., 2013;Rossow and Schiffer, 1999). Moreover, cloud optical properties vary regionally due to the different processes involved in their formation and evolution . Finally, the optimal thresholds for a particular excision scenario also depend on the expected brightness of the surface (Thompson et al., 2014). Nevertheless, prior studies have relied on universal cloud models and excision criteria, applying the same models for use across the globe. Considering cloud fractions as a function of surface type and latitude could lead to more precise cloud detection. Based on this, we hypothesize that location-specific cloud models can improve performance relative to global methods.
To test this hypothesis, we refine the cloud-screening algorithm previously introduced in Thompson et al. (2014). The original algorithm uses three bands of interest to distinguish clouds from other surface types (El-Araby et al., 2005;Ackerman et al., 1998;Williams et al., 2002;Griffin et al., 2003). This has the desired properties of being simple, fast, transparent, and tunable by ground operators. Here, we analyze the Hyperion global archive to provide globally applicable models that are parameterized by latitude and surface type, enabling reflectance thresholds to predict the classification of cloud-contaminated data and non-cloud-contaminated data (land, water, and snow) in new scenes for EMIT and other future orbital missions.

Methods
Most cloud-screening algorithms are classifiers -they analyze independent data points and decide which ones should be flagged as cloudy. These individual data points can be pixels within a data cube (Altinok et al., 2016), in which case each location is labeled independently. Other algorithms label segments within the scene (Thompson et al., 2014), or entire image cubes. In this study, our data points are individual pixels that are single locations within a scene, or image. Each pixel is associated with a complete measured radiance spectrum, and each receives an independent classification as either clear or cloudy. We analyzed data from the Hyperion imaging spectrometer, an instrument on board NASA's EO-1 satellite, which collected a globally representative dataset over more than a decade .
After calculating top-of-the-atmosphere (TOA) reflectance, we labeled each pixel as one of four categories (land, water, snow, and clouds), which will be described in further detail in Sect. 2.2. We then accumulated brightness distributions for each pixel type, describing their respective TOA values in each of the following wavelengths of interest: 447.17, 1245.36, and 1648.90 nm (Thompson et al., 2014). Our procedure took the following steps: we acquired a dataset of representative spectra from a historical data archive, we chose channels or bands that would be used to classify pixels, we calculated pixel brightness distributions for different surface types and latitudes, and finally we optimized channel thresholds given the distributions and false-alarm requirements. These channel thresholds can predict the optimal TOA values for screening clouds in new scenes based on advance knowledge of surface types and viewing geometry.

The dataset
The Hyperion instrument was a push-broom imaging spectrometer that operated on board NASA's Earth Observing-1 (EO-1) spacecraft from 2000 to 2017 that followed a polar orbit. It acquired spectra in 220 channels spanning 357 to 2576 nm at approximately 10 nm spectral resolution. Hyperion operated in a targeted acquisition mode, acquiring images at specific locations of interest. Each map was approximately 7.7 km in width, typically 42 km in length with a spatial resolution of 30 m per pixel. During its 17-year operational lifetime, Hyperion acquired tens of thousands of maps across diverse areas including Arctic, Antarctic, oceanic, and terrestrial surfaces. We selected a subset of 102 Hyperion images over the entire time range of the mission for our study. To ensure that we sampled the entire globe, we collected approximately 25 images from each section of the globe. The following subsets were included: tropics (23.5 • S to 23.5 • N), Arctic (66.5 to 90 • N), Antarctic (66.5 to 90 • S), and midlatitudes (66.5 to 23.5 • S and 23.5 to 66.5 • N). Most Hyperion data were acquired over land, so we included a subset of longitudes spanning the Pacific Ocean (121 to 180 • E and 121 to 180 • W) to capture the spectral properties of water. We found that there were many fewer applicable images from both of the polar regions, so combining them we included 21 images. We included 30 images from the tropics region and 51 images from the midlatitudes. These numbers are not round because of the issue of resampling when including the ocean subset. The 19 images in the ocean category can be found in the other latitudinal zones, so any crossover between the two subsets was eliminated to ensure the entire sample set was unique. We transformed the radiance measurements to top-of-the-atmosphere (TOA) reflectance values to reduce the variability caused by solar geometry. Both Hyperion and EMIT are typically nadir-viewing instruments, with the EO-1 satellite in sun-synchronous orbit and EMIT on board the International Space Station.
In order to detect cloudy pixels with confidence, we selected three specific spectral bands that can distinguish clouds from other surface types. It is important to note that Earth's total TOA energy flux constitutes the total incoming solar radiation, the consequential outgoing reflected short-wave radiation from the clouds and surface, and the outgoing emitted longwave radiation from Earth's surface, atmosphere, and clouds (e.g., Trenberth et al., 2009). Some other cloud-screening algorithms do use longwave channels which provide extra sensitivity to high-altitude clouds (Mercury et al., 2012). However, this study sought to discriminate cloud and clear locations using the visible-shortwave infrared (VSWIR) region alone. This made our results directly applicable to future missions measuring solar-reflected wavelengths. We selected three wavelengths of interest, namely 447.17, 1245.36, and 1648.90 nm, based on previous studies distinguishing clouds from land, water, and snow (Ackerman et al., 1998;Griffin et al., 2003;El-Araby et al., 2005;Thompson et al., 2014). Clouds and snow had a high reflectance in the 447 nm band while land and water did not. The near-infrared (1245 nm) and shortwave-infrared (1650 nm) band reflectance values effectively discriminated between clouds and snow. Snow had a slightly lower reflectance in the 1245 nm band than clouds while the 1650 nm band showed snow as even less reflective than clouds (Griffin et al., 2003). We used just these three channels, since a small subset of bands enables threshold-based algorithms to be encoded easily into instrument electronics hardware, for real-time execution at the native frame rate of the spectrometer.

Classification
Accurate ground truth classifications were necessary to define cloud and land statistics for building the classifier. They were also useful for evaluating the resulting model's accuracy by comparing predictions against manual labels. Pixels were hand-labeled using image editing software to assure accuracy in classifying each surface type (Fig. 1). The pixels in each image of the sample set were given a color value based on their surface type (land: red; water: blue; cloud: red; snow: cyan; ambiguous: black). We then used these color values to relate the pixels' category of land, water, snow, or clouds to their TOA reflectance values. After fitting this model, the manual classification will be used to verify the model's classification accuracy. We only labeled opaque clouds; all pixels bordering various classification types were labeled as "ambiguous" to avoid misclassification.
After manual labeling, we associated the surface types with the TOA reflectance value in each of the three bands of interest. This association produced three-dimensional frequency distributions of TOA reflectance values based on wavelength and surface type (Fig. 2). They described the conditional probability of a pixel's TOA reflectance value given its classification as cloud or clear sky, c 1 and c 2 , respectively (P (y|c 1,2 )). The non-cloud distribution contained all non-cloud surface types (land, water, and snow). To assess the classification power over each surface, we tracked the brightness distributions of each surface type independently. To capture the effect of different climates, we represented  the distributions as a function of latitudinal zone, P (y|x, c), where x was the zone of interest.

Algorithms
Our cloud-screening approach predicted scene-specific thresholds in three bands for real-time use on board (Thompson et al., 2014). Our cloud-screening algorithm defined an exclusion region R ⊆ R d , i.e., a range of TOA reflectance values for which a pixel was classified as cloudy. In other words, it mapped the pixel brightness values to a binary classification c = f (y): R d −→ {c 1 , c 2 }. A vector y represented the spectrum of the pixel being classified. Thus, the decision rule for this classification was where R was defined with a set of thresholds, ϕ (in this case a triplet). Any pixel exceeding all three thresholds simultaneously is classified as cloud-contaminated (Fig. 3). We then used the following expected loss function, where α FP and α FN were the false-positive and false-negative penalties, respectively, A false-positive penalty applied to cases where clear pixels were classified as cloud-contaminated. P (y|c 1 ) and P (y|c 2 ) were the probability of encountering a cloud-contaminated pixel and a clear pixel, respectively. P (c 1 ) and P (c 2 ) were the prior probability of clouds and clear sky, respectively, based on a historical average. To eliminate any bias due to historical observations, a "uniformed" prior assigning equal probability to all classes could also be used. Minimizing this function (Eq. 2) produced the optimal threshold for the given conditions defined by x. Using Bayes' rule and assuming independence (Thompson et al., 2014), the expected loss function could be decomposed into the respective likelihoods and priors for the posterior described above, Thus, we could use the likelihood, or sampling distribution, created from the Hyperion sample set to minimize our expected loss and produce predictive thresholds for screening. As in Thompson et al. (2014), we represented probability distributions through histogram counts, creating a 3D table with one dimension for each of the three wavelengths. This allowed a fast-recursive calculation of Eq.
(3) when searching over thresholds. For a given false-positive penalty, we searched over all possible thresholds and selected the one which produced the lowest effective loss.

Results
The globally representative sample set of imaging spectroscopy data provides a comprehensive sample of TOA reflectance for various surface types in space, time, and wavelength as well as a prediction model for screening cloud-contaminated data on board orbital imaging spectrometers. This section discusses our findings concerning excision thresholds, cloud brightness, the potential improvement yield of downlink using the cloud-screening algorithm, a comparison of resulting cloud fractions in our dataset with previous literature, justifying the empirical error of our dataset, and the implications of our study and cloud-screening tool for future missions.

Cloud-screening thresholds
The model used for cloud screening was developed using cloud brightness distributions in TOA values as a function of time, space, and wavelength. The brightness distributions collected from the Hyperion sample set represent cloud and non-cloud brightness values in TOA units, for each band (447.17, 1245.36, 1648.90 nm). The sample set scenes cover representative regions across the globe (Table 1).  The output of the algorithm was a threshold triplet that defined the exclusion region, i.e., the minimum TOA reflectance values of opaque cloud-contaminated data (discussed in Sect. 2.3). The penalties in the expected loss function determined our tolerance for errors; a higher falsepositive penalty led to a more conservative threshold and a smaller exclusion region. As we hypothesized, we were able to improve performance further by considering the expected surface properties when defining the exclusion region. For example, since clouds and snow had similar reflective properties in two of the bands used, snow scenes were most challenging and dominated the threshold criterion. Consequently, one could use a more aggressive threshold triplet to screen clouds outside Arctic regions, improving performance without incurring misclassifications. The results of a conservative threshold (α FP = 1000) calculation, a moderate threshold (α FP = 100) calculation, and an aggressive threshold (α FP = 10) calculation are shown in a two-dimensional histogram representing all scenes in the sample set (Fig. 4).
The optimal thresholds, where the loss was minimized, at various false-positive rates using the Hyperion sample set are shown in Table 2. The thresholds were defined differently for each latitudinal zone. Section 3.4 describes the statistical validation tests we used to ensure the size and breadth of our subset, or sampling distribution, were sufficient to predict these thresholds for future scenes.

Cloud brightness
The Hyperion global dataset was sampled and manually classified to understand cloud brightness as a function of time, space, and wavelength. We collected the TOA values in a three-dimensional histogram, one axis for each wavelength studied. This produced a probability distribution of TOA values of clouds globally that we used to predict the classification of TOA values for future scenes. In order to verify whether classifying clouds depending on their latitudinal zone would produce a lower false-alarm rate, we subset our scenes based on latitude and surface type. Table 3 presents the mean TOA reflectance values for each latitudinal zone and wavelength. Cloudy pixels had higher TOA reflectance values while non-cloudy pixels had generally lower TOA values (Fig. 3). We found our data to align with the general properties of non-cloud surface types discussed in Ackerman et al. (1998). Specifically, pixels with snow had low TOA reflectance values at 1250 nm and even lower TOA reflectance values at 1650 nm, while having high TOA reflectance properties at 447 nm.
Separating cloud brightness values as a function of latitude is also helpful in determining the type of clouds formed in each region (Oreopoulos et al., 2014), although this is not included in our study. Studying these TOA reflectance value distributions for opaque cloud cover will be helpful to understand shortwave albedo at regional and global scales. For example, opaque clouds compose a smaller percentage of global cloud cover with respect to transparent or spatially heterogeneous clouds (e.g., Rossow and Schiffer, 1999;Stubenrauch et al., 2017), but deep convective clouds have an outsized influence on Earth's TOA radiative budget and hydrological cycle in the tropical latitudes (e.g., Jakob and Tselioudis, 2003;Tan et al., 2015).

Empirical error tests
This section evaluates the stability of the threshold estimation approach and then quantifies cloud excision performance. Based on parameters such as latitude and surface type, we determined the false-positive and false-negative performance for different observing conditions. Screening classification depended largely on the false-positive parameter (Fig. 5).
It was critical that any threshold set generalized to new scenes not yet seen. While an infinitely large dataset would be sure to capture the real statistics of the globe, we were nec-essarily limited in the number of scenes used in the analysis. To confirm this dataset was sufficiently representative to produce general thresholds, we performed two validation tests. First, we performed a leave-one-out cross-validation experiment, recalculating cloud-screening thresholds 102 times and excluding a different scene for validation from each trial. Every test conducted resulted in the same thresholds presented in Table 2. This stability is one validation that the optimal margin for the distribution for the brightness resolution of our lookup table was not sensitive to "outlier" cases but rather responded to the true statistics of global clouds. More generally, it confirmed that our dataset was sufficient in space, time, and wavelength to predict optimal thresholds for future scenes.
In the second test, we conducted an experiment to measure the variance of the cloud-screening threshold estimates using bootstrap techniques. The 102 images used in the initial experiment were used to create a sample set through sampling with replacement 500 unique times. Data concerning the threshold calculation were recorded each of the 500 times. Of particular interest is the variance in threshold calculation for each latitudinal zone ( Table 5). The variance in each latitudinal zone was low, showing that our estimation approach was robust.

Potential improvement yield of screening
To evaluate the utility of cloud screening in terms of imaging spectroscopy science investigations, we analyzed the improvement yield for two specific cases of future orbital imaging spectrometers. We simulated the data return with and without the use of onboard cloud screening.
Our first case study used the orbital parameters of the Earth Surface Mineral Dust Source Investigation (EMIT) mission . EMIT will use a visible to short wavelength infrared (VSWIR) imaging spectrometer to map the mineralogy of mineral-dust-forming regions worldwide. This will improve our understanding of the mineral composition of airborne dust particles, informing the Earth system models that simulate the dust cycle. Understanding the composition of mineral dust in Earth's atmosphere will in turn provide insight into the impact of dust on direct radiative forcing in Earth's climate. The EMIT mission will be launched to the International Space Station in 2021, with an orbit dominated by low-latitude regions. It intends to use data under clear-sky conditions without significant aerosol loading and filter scenes with an aerosol optical depth (AOD) greater than 0.4 at 550 nm. The Hyperion dataset is broadly representative of these atmospheric conditions. A case study for EMIT provides one example of the potential for onboard cloudscreening (Table 6). Since the EMIT mission plan does not entail collecting data over the latitudinal zone specified as ocean, we ignored that region in the case study.
We intersected the EMIT coverage area with historical cloud probability maps to assess the potential improve-  Figure 5. An example of pixels classified as cloudy being excised from a Hyperion image (EO1H1940712011304110T1). These images were created using the default RGB wavelengths in the ENVI program, although the algorithm to create the panels in this figure only considers the three wavelengths used in our thresholds. The panels show the following: (a) the raw image with no screening, (b) cloud screening with αFP = 1000 where 8.5 % of the total pixels were excised, (c) cloud screening with αFP = 100 where 16.4 % of the total pixels were excised, and (d) cloud screening with αFP = 10 where 25 % of the total pixels were excised. Note this is an example of excision in a particular scene. Our training set was intentionally designed to capture opaque clouds and to pass ambiguous or translucent cloud areas. ment for the instrument Chien et al., 2019). We simulated cloud cover fractions using pre-calculated global cloud probabilities (x) from historical MODIS data (Mercury et al., 2012), defined as an annual average cloud cover probability at a spatial resolution of 1 • . Then we simulated International Space Station (ISS) observations at a 10 s rate for 1 year, starting on 1 February 2022. The large improvement indicated the value of screening cloud-contaminated data. For the cloud-screening approach in this work, we predicted at least double the current return of useful data in each latitudinal zone and for the mission overall. There is a notable difference in yield when considering all regions at once and when considering one region at a time. Some areas of the globe are cloudier than others, so the benefit of using a cloud-screening tool is particular to the region(s) of interest and the sampling strategy of each mission. As previously discussed, past literature presents cloud cover fractions that are greater over land than water and that tropical regions are more cloudy than other latitudinal zones (Eastman et al., 2011;Ackerman et al., 1998). The EMIT case study cloud fractions showed that cloud cover in the tropics was greater than any other region studied. Due to the nature of our latitudinal zone sorting, we could not directly compare the ocean cloud cover fractions with land cloud cover fractions; the other latitudinal zones could also have contained regions associated with oceans. CHIME (Copernicus Hyperspectral Imaging Mission for the Environment) is an ESA (European Space Agency) mission that aims to provide routine hyperspectral observations of the Earth to aid in management of natural resources, assets, and benefits for the European Union. One main aspect of CHIME's technological advances is to facilitate increased field-of-view (> 5 • ) observations and low spatial sampling (< 30 m) with a high-data-rate (> 1 Gbit s −1 ) processor on board the instrument. In light of this requirement, an onboard processing architecture such as a cloud-screening algorithm is valuable to the CHIME mission.
The CHIME team conducted a simulated study, comparable to EMIT, where they reported simulated cloud cover based on one orbital cycle (223 orbits over 15 d) for summer 2012 with meteorological statistical inputs for cloud coverage simulation. The preliminary results of this study revealed that with a cloud-screening algorithm on board CHIME would experience about a 50 % increase in usable data (Table 6). In summary, both of the case studies considered facilitated at least double the return of viable data for each mission's desired objective. Table 5. Variance in thresholds used to screen clouds based on bootstrapping the threshold calculations described in the initial experiment 500 times. The variance is presented as a percentage of the mean of the particular zone. The false-positive setting was αFP = 1000 and the false negative αFN = 1.

Discussion and conclusion
The method described for screening cloud-contaminated data on board orbital imaging spectrometers will at least double the volume of useful data for a fixed downlink size. We collected and studied a globally representative dataset, producing optimal screening thresholds based on latitudinal zones. Using latitude as a parameter in screening clouds will help correctly classify cloud-contaminated pixels while reducing misclassifications of other surface types. The overall yield of useful data doubles when using the screening algorithm. In addition, we have produced a representation of cloud brightness in the 447.17, 1249.36, and 1650.90 nm wavelength bands, changing with latitude and surface type, based on TOA reflectance values. Our results showed that mean cloud TOA brightness differed in the bands studied as a function of latitudinal zone and surface type: tropics, Arctic, Antarctic, midlatitudes, and Pacific Ocean areas (Table 4). The starkly higher differences seen in the Arctic region show that this region needed more conservative thresholds than a global "universal" model which did not calculate unique screening thresholds based on latitude. The difference in these mean values indicates that the optimal thresholds assigned for the classification of clouds in each area should also differ (Table 2).
Given the relatively low complexity and risk involved in implementing onboard cloud screening, it is a valuable option available to mission designers trying to achieve higher yield at low cost. It can be incorporated directly into instrument hardware, saving the storage costs of downlinking cloudy data. Its operation can be tuned over time, if needed, to obtain optimal performance for the specific observing profile of the mission. Even a very conservative threshold, with almost no probability of excising good data, can significantly reduce data volumes (Thompson et al., 2014). For this reason, the EMIT mission  will use this cloud-screening method to optimize its downlink within the allowable resources provided by its position on the International Space Station. Its observation plan, which targets arid regions in the midlatitude and tropical latitudes, can use more aggressive thresholds without risking error due to snow cover. The data accumulated over its year-long mission will provide independent validations on these thresholds and brightness distributions. The methods suggested for cloud screening in this article as with Thompson et al. (2014) rely only on TOA reflectance and thresholding. Even alternative methods flown on EO-1 (Wagstaff et al., 2018) and IPEX (Altinok et al., 2016) using random decision forests are only slightly more computationally demanding (impacted by the size of the trees, number of trees voting, and neighborhood used to inform the classification). Current imaging spectrometers produce data at high rates (gigabits per second). However, because classifiers are only likely to use a small number of spectral bands, only a smaller amount of data need to be processed on the fly, making next-generation conventional processors viable. In addition, space missions are flying alternative processors that would provide dramatically enhanced computation: the IPEX CubeSat flew a Gumstix Overa , the Mars Helicopter (Grip et al., 2019) will fly a Qualcomm Snapdragon, and the ESA FSSCCAT CubeSat (ESA 2020) will fly an Intel Myriad processor. However, future computing needs for on board will continue to grow (Dally et al., 2020) well beyond these flights.
The results of this work also have applications outside screening clouds on board imaging spectrometers. Conversely, collecting solely cloud-contaminated data that are identified using the algorithm described in this paper can aid in an instrument's stray-light correction or other data products. On the other hand, the global study of cloud brightness shown in this paper has the potential to be used in various cloud studies, for example those concerning cloud type and subsequent brightness as a function of latitudinal zone (Oreopoulos et al., 2014).
Data availability. We used the Hyperion data found at Earth Observing One (EO-1; https://doi.org/10.5066/P9JXHMO2, EROS Center, 2018) to derive our data products for this study.
Author contributions. MWS collected the dataset from the Hyperion database with the help of WOD, hand-labeled pixels in the chosen images to produce mean cloud brightness statistics, performed the calculations for determining optimal screening thresholds with the guidance of DRT, and performed empirical error testing with the guidance of DRT. DRT designed the project and supervised its completion. ROG supervised the project. BHK provided necessary background and implications of cloud brightness mean values. AY and RV contributed information regarding the EMIT and CHIME case studies, respectively. SC aided in developing the background and further applications with regards to cloud screening. MWS outlined and wrote the manuscript with DRT and received direct contributions and feedback from all co-authors.