Articles | Volume 17, issue 12
https://doi.org/10.5194/amt-17-3739-2024
https://doi.org/10.5194/amt-17-3739-2024
Research article
 | 
24 Jun 2024
Research article |  | 24 Jun 2024

Revealing halos concealed by cirrus clouds

Yuji Ayatsuka
Abstract

Many types of halos (including arcs) appear in the sky. Each type of halo corresponds to the shape and orientation of ice crystals in clouds, and reflects the state of the atmosphere; therefore observing them from the ground greatly helps in understanding the state of the atmosphere. However, halos are easily obscured by the contrast of the cloud itself, making it difficult to observe them. This difficulty can be overcome by enhancing halos in images, for which various techniques have been developed. This study describes the construction of a sky-color model for halos and a new effective algorithm to reveal halos in images.

1 Introduction

Ice crystals that form clouds (mainly cirri and cirrostratus) sometimes show rings and arcs of light (often colored) around the sun or the moon, as shown in Fig. 1. These optical phenomena are called halos.1 Halos can be of many types, with each type corresponding to the shape and orientation of ice crystals in clouds (Greenler1980). Halos provide valuable information about the atmosphere. Observing halos from the ground is an important way to understand meteorological processes. It is a wide-area observation of ice crystals in clouds. For example, the difference in frequency of appearance between 22 and 46° halos suggests the ratio of pristine to non-pristine crystals in clouds (van Diedenhoven2014). There are also several studies of ice crystals and halo observations (Lynch and Schwartz1985; Sassen et al.1994; Um and McFarquhar2015; Sassen1980; Lawson et al.2006). An attempt is also being made to achieve automated observations of halos using a sun-tracking camera (Forster et al.2017) or a total sky imager (Boyd et al.2019). Image processing techniques can be used to observe even faint halos or other atmospheric optical phenomena in photographs more clearly, which can greatly aid in these types of studies (Riikonen et al.2000; Moilanen and Gritsevich2022; Großmann et al.2011).

https://amt.copernicus.org/articles/17/3739/2024/amt-17-3739-2024-f01

Figure 1Types of halos: this image contains at least nine types of halos. (All the photographs in this paper were taken by the author in Tokyo, Japan.)

Download

A halo is formed by the interaction of sunlight with the particles in a cloud. However, the contrast of the cloud itself prevents us from seeing halos clearly. To resolve this difficulty, halo observers have developed various image processing techniques that enhance halos in images and improve their observation, and a BR (blue minus red) processing proposed by Rossetto (2011) and Lefaudeux (2013) is one such technique and is known to be an effective method for enhancing halos. It involves subtracting a pixel value in the red channel from that in the blue channel and setting the difference as the gray value of the pixel. As a result, the areas in an image that are more reddish or bluish than the surrounding areas are enhanced. The author of the present study has developed a revised method, called autoBR, which was implemented in an image processing tool, named Atmospheric Optical Image Enhancer (Ayatsuka2022). In autoBR, the red and blue channels are differently weighted and the green channel is also referenced (details are described in Sect. 3). In contrast, in BR processing, the red and blue channels are equally weighted and the information in the green channel is not used.

Figure 2 shows a halo image processed with both BR processing and autoBR. It is quite noticeable that the contrasts of the clouds are flattened by the two processes, especially by autoBR. In other words, these processes are thought to flatten the “cloud colors” in the image and extract the appearance and intensity of halos. It is quite useful for observing and analyzing halos precisely through ground images.

https://amt.copernicus.org/articles/17/3739/2024/amt-17-3739-2024-f02

Figure 2Image results obtained using of the BR processing and autoBR.

Download

This study constructs and validates a sky-color model for halo display. Using the model, a new algorithm is developed, called sky-color regression, to more effectively cancel cloud colors and enhance halos in the image.

The rest of the paper is organized as follows. Section 2 discusses related work on image processing for similar purposes. Section 3 presents details of existing algorithms, BR processing, and autoBR. Section 4 proposes a sky-color model and validates it with an image of a halo display. Section 5 describes the new algorithm called sky-color regression. Section 6 discusses some remaining issues related to the new algorithm, and Sect. 7 concludes the paper.

2 Related work

As described in the previous section, the contrasts of clouds should be reduced to extract halos in an image. Similar processes have been studied for other purposes, for example, for dehaze or fog reduction. Cloud detection using color information obtained from images is another related research topic, although with an opposite purpose.

2.1 Dehaze/fog reduction

The main goal of dehaze/fog reduction is to compensate the poor camera image quality when observing land and sea scenes under hazy conditions (Chengtao et al.2015; Singh et al.2016). Building physical models of the hazed view to recover background information is one of the approaches that is similar to the one used in the present study. Shi et al. (2022) used blue and red channel information to improve images in sandstorms.

The main difference between these studies and the present one is that, in their target situation of the previous studies, complex background images are obscured by smooth haze or fog, whereas in the target situation of the present study, relatively simple patterns are obscured by clouds. Because the present study has a different purpose, which is to process the image for easier halo detection and not to compensate for the image quality, a simpler model than those described in previous studies can be employed.

Gao and Li (2017) described the removal of cirrus clouds from the ground images taken by a Landsat satellite by using a microwave imager. An image taken by a band that is highly reflected by cirrus clouds helps in effectively removing cirrus clouds from the corresponding visible light (RGB) image.

2.2 Cloud detection

Many studies have investigated the detection of cloud regions or segmentation in sky images taken from the ground for meteorological observation. Koehler et al. (1991) set a threshold on the ratio of red and blue pixels that can be utilized to judge whether a pixel is a part of clouds.

Dev et al. (2017) investigated the contribution of composite parameters for a cloud segmentation algorithm, including values in color spaces other than RGB, such as HSV and L*a*b*. They found that the saturation in HSV; the ratio of red and blue in RGB; and B-RB+R, also called the “normalized blue / red ratio” in Li et al. (2011), are important for cloud detection. These studies suggest that the blue / red ratio is also important for cloud cancellation and halo extraction in an image.

3 Existing algorithms

This section describes the existing algorithms for extracting halos in images, i.e., BR processing and autoBR. It also shows how color pixels are translated into grayscale pixels in translation maps.

3.1BR processing

BR processing, also referred as “color subtraction,” is a widely used technique for enhancing and explaining halos and other atmospheric optical phenomena2. It is introduced as an image processing technique using a tool called the “channel mixer.” It is available in Photoshop or other such photo-retouching applications. The channel mixer outputs a new pixel value of a specified channel calculated from the RGB values of the corresponding pixel in a source image. The values are added or subtracted with specified weights, and a “shift” value is used to adjusts the output range to keep the pixel values in the interested area of interest within 0–255.

For BR processing, the weight for a red channel is set to the same magnitude but with opposite sign as the weight for a blue channel. Although the magnitude is adjustable, it is often set to 200 %, the maximum value allowed by the tools. The shift value is set and adjusted by the user by looking at the resulting image. Figure 3a shows a resulting gray map when the parameters are set to B: 200 %, G: 0 %, R: 200 %, and shift: 0. The blue and red lines show the edge of the underexposed and overexposed area, respectively, while the magenta line shows pixels in the center value (128). The map depicted in Fig. 3b shows the case when the shift parameter is set to 128.

https://amt.copernicus.org/articles/17/3739/2024/amt-17-3739-2024-f03

Figure 3Color-to-gray translation maps for (a, b) BR processing and (c) autoBR in the B vs. R space: the magenta lines show the center value (128), and the blue and the red lines show the saturation points (255 and 0).

Download

The pixel-value translation in BR processing is represented as follows:

(1) I = α ( B - R ) + β ,

where I is the output intensity value, B and R are pixel values in the source image, and α and β are the “magnitude” and “shift” (I is set to 0 if I<0 and to 255 if I>255). For larger α, the contrast of the output image will be higher and the width of the area between the blue and red lines in a gray map will be narrower.

3.2 autoBR

Extending BR processing, more effective parameters have been heuristically explored and the automatic adjustment of the “shift” parameter has been implemented for the autoBR algorithm. The pixel-value translation in autoBR is represented as follows:

(2) I = α ω B B + ω G G - R + β ,

where α=2.0, ωB=1.5, and ωG=0.25. The β is set to a value such that the average of the I values (excluding too dark/bright areas in the source image) becomes 128. An example gray map for the translation is shown in Fig. 3c, where the blue, red, and magenta lines are steeper than those for BR processing.

4 Model and validation

Considering how the presented algorithms work effectively to extract halos, a model of sky color in images can be assumed. This section explains the model and validates it with some photographs.

4.1 A model for sky color

This study provides a simple physical model for celestial light, with no direct beam from the light source and no particles other than those that form clouds. Light Lsky from one direction in the sky contains the following (Fig. 4):

  1. Rayleigh scattering LR from air molecules (sky blue background)

  2. Mie scattering LM from cloud particles (white cloud)

  3. refracted/reflected light LH from cloud particles (e.g., halos).

In a digital image, Lsky, LR, LM, and LH are vectors typically containing R, G, and B intensities. If translation g that converts colors in an image to grayscale reduces the contrast of the cloud color, it satisfies the following formula:

(3) g L R + L M + L H g L R + L H .

Considering the color-to-gray translation maps in Fig. 3, the above formula implies that the direction of vector LM is almost parallel to the red, blue, and magenta lines when projected onto the B vs. R space. Figure 5 shows the model for sky color in the B vs. R space. LR in Fig. 5 represents sky color if the clouds and halos are not there. The arrows represent vectors, LM and LH, added to the vector LR (in the B vs. R space). The color of the sky at a given point in an image is the result of the summation of three vectors.

https://amt.copernicus.org/articles/17/3739/2024/amt-17-3739-2024-f04

Figure 4A physical model of the components of skylight.

Download

https://amt.copernicus.org/articles/17/3739/2024/amt-17-3739-2024-f05

Figure 5A pixel color model of skylight on an image: the arrows represent vectors, LM and LH, added to the vector LR (in the B vs. R space). The color of the sky at a given point in an image is the result of the summation of three vectors.

Download

4.2 Validation of the model

Figure 6 shows a halo display image and its color maps of it in B vs. R, R vs. G, and G vs. B spaces. The pixel colors in the maps other than black represent (one of) the real colors in the image. The colors are distributed linearly in the maps, roughly parallel to the lines shown in Fig. 3. Some colors are outside the main distribution, corresponding to the silhouette of the foreground objects in the image. The width of the main distribution is as narrow as the area between the blue and red lines depicted in Fig. 3. These reasons explain the effectiveness of BR processing and autoBR.

https://amt.copernicus.org/articles/17/3739/2024/amt-17-3739-2024-f06

Figure 6Color maps of a sky image including halos.

Download

The lines in the color maps are regression lines for the distributions of color in each space. From the slope of the B vs. R regression line, it can be expected that autoBR with a higher slope would be more effective than the BR method with a slope of 1.0. Blue minus green is a variation in BR processing (Juutilainen2015) and will be effective when the slope of the regression line in the B vs. G space is close to 1.0.

Figures 7 and 8 show partial color maps and their regression lines for the areas indicated by box lines. The partial color maps, corresponding to the red and yellow boxes, are superimposed on the full color maps shown in dimmed colors.

Figure 7 shows the color maps and regression lines for areas without halos. These maps are thinner and more linear than the full map. This means that the blue of the sky, whiteness of the light, and shadow of the clouds are also uniform in these small areas. The slopes of the regression lines for the red and yellow boxes are clearly different from each other, especially in the B vs. R space. As a result, the optimal parameters for reducing cloud contrast vary from region to region.

https://amt.copernicus.org/articles/17/3739/2024/amt-17-3739-2024-f07

Figure 7Color maps of areas without halos in the sky image: color maps for the red box are shown in the left column, and those for the yellow box are shown in the right column.

Download

Figure 8 shows the color maps of the areas with halos. The red box contains a left parhelion, and the yellow box contains part of a circumzenithal arc. Notably, the loop part of the color map seems to correspond to the parhelion. Also note that there is a bump in the lower part of the B vs. R color map, indicating the red color of the circumzenithal arc. Such features are less significant in the G vs. B space and somewhat difficult to distinguish in the R vs. G space.

https://amt.copernicus.org/articles/17/3739/2024/amt-17-3739-2024-f08

Figure 8Color maps of areas with halos in the sky image: color maps for the red box, including a parhelion, are shown in the left column, and those for the yellow box, including a circumzenithal arc, are shown in the right column.

Download

Color maps for other halo display images have similar characteristics, which supports the validity of the sky-color model shown in Fig. 5.

5 Sky-color regression algorithm

The color maps explained in the previous section suggest that an algorithm can be revised for reducing cloud contrast by adjusting the parameters according to the slope of the regression line. The maps also suggest that the parameters should be adjusted locally rather than using one set of parameters for all areas of an image. Based on these observations, a new algorithm called “sky-color regression” is constructed.

5.1 Regression-based processing

Let ωBR, ωRG, and ωGB denote the slopes of the regression lines in B vs. R, R vs. G, and G vs. B spaces. Herein, a transformation from the RGB color to grayscale is considered as follows:

(4) I = α 1 ω BR B - R + α 2 ω RG ω GB G + β .

For this study, the following conditions were set for the new algorithm: α1=2.5 and α2=1.0. However, it has not yet been discussed how the green channel affects the sky color; therefore, the values were explored heuristically. A value of β was set as in autoBR, i.e., the average of I should be 128. This process is called “regression-based processing”.

Figure 9 shows the images processed with autoBR and regression-based processing. When regression-based processing is applied to the entire image, clouds around the circumzenithal arc are reduced more than that achieved with autoBR, although the former method seemed to be less effective for the other areas. When regression-based processing was applied to a part of the image, clouds in the part were effectively reduced, and the brightness was also adjusted accordingly.

https://amt.copernicus.org/articles/17/3739/2024/amt-17-3739-2024-f09

Figure 9Comparison of regression-based processing with autoBR.

Download

5.2 Local adaptive processing

If the area to calculate the regression line is too small, the accuracy of the line will worsen. However, if the area is too large, the locality will deteriorate. Assuming a wide-angle image such as the photos shown in this paper, the default value was set to one-sixth of the long side of an image.

For best results, regression parameters ωBR, ωRG, and ωGB should be updated for every pixel in an image; however, this is computationally expensive. If the best parameters for neighboring pixels are similar, some calculations can be omitted as shown below:

  • 1.

    The parameters are updated per block of a few pixels.

  • 2.

    The regressions are performed on an appropriately shrunken image.

When implementing the sky-color regression algorithm, a target image should be divided into b × b-px blocks. The regression parameters are calculated on a shrunken image of size 1/b. The default value of b is set so that the long side of the shrunken image is approximately 1000 px, i.e., b=p/1000, where p is the long side of the target image. If the camera angle is known, the value can be optimized using this information. How to calculate the best value for each angle is a future work of this study.

An example of an image processed by the sky-color regression algorithm is shown in Fig. 10. The width and height of the image are 4024 × 6048 px, and the image was processed in 4.3 s using sky-color regression implemented in Java on a PC equipped with an Intel Core i7-1165G7 (2.80 GHz).

https://amt.copernicus.org/articles/17/3739/2024/amt-17-3739-2024-f10

Figure 10Processed with sky-color regression: in the original image (see Fig. 6), even experienced observers can hardly detect halos except for a parhelion, a circumzenithal arc, and a part of the 22° halo and the upper tangent arc.

Download

More examples are shown in Fig. 11, which compares sky-color regression with BR processing and autoBR. While the algorithms mainly enhance colored halos, a colorless parhelic circle is also visible in the image at the second row from the bottom. It suggests that a parhelic circle appears slightly bluer than the background clouds.

https://amt.copernicus.org/articles/17/3739/2024/amt-17-3739-2024-f11

Figure 11Examples of sky-color regression in comparison to BR processing and autoBR.

Download

6 Discussion

6.1 Evaluation

Figure 12 displays histograms of processed image parts by the sky-color regression, BR processing, and autoBR. Rows (a) and (c) show areas without halos, while row (b) shows an area with circumzenithal and supralateral arcs.

https://amt.copernicus.org/articles/17/3739/2024/amt-17-3739-2024-f12

Figure 12Histograms of processed image parts: rows (a) and (c) are areas without halos, while row (b) is an area with circumzenithal and supralateral arcs.

Download

With the sky-color regression and autoBR, distributions of pixel values for areas (a) and (c), without halos, are simple standard distributions, while there are two or more peaks in distributions with BR processing. Standard deviations are also smaller with the sky-color regression than with BR processing. For area (a), the standard deviation is smaller when using the sky-color regression compared to autoBR. Conversely, for area (c), the standard deviation is smaller with autoBR than with the sky-color regression. It shows that the sky-color regression is not always the most effective method for canceling out cloud and should be refined in future studies.

For area (b), all algorithms produced a histogram with three peaks corresponding to the clouds, reddish parts, and bluish parts of the halos. However, the peaks are most clearly separated with the sky-color regression. The standard deviation is also the largest with the sky-color regression.

The average values for the areas are maintained around 128, which is the midpoint value of the range 0 to 255, with the sky-color regression. It shows that the local adaptive processing works.

6.2 Other atmospheric optical phenomena

The sky-color regression algorithm can also be used efficiently to enhance other colored atmospheric optical phenomena. Figure 13a shows an example of quite faint third- and fourth-order rainbows. The algorithm extracts colored bows from the background sky without clouds. Figure 13b shows the diffraction pattern of Japanese cedar pollen corona. The algorithm also works on the sky color near sunset.

https://amt.copernicus.org/articles/17/3739/2024/amt-17-3739-2024-f13

Figure 13Applying sky-color regression for images of other atmospheric optical phenomena: (a) third- and fourth-order rainbows and (b) pollen corona (Japanese cedar).

Download

6.3 Limitations

In the sky-color regression algorithm, the mixing of the red and blue channels is calculated statistically, while the weight of the green channel is adjusted heuristically. The size of the pixel block and the area over which the regression is performed are also determined by heuristics. These parameters would also be statistically determined if further analysis was performed on the images of halo displays.

The color maps of an image change according to the white balance setting. Because of its adaptive nature, sky-color regression is less sensitive to white balance than BR processing and autoBR are. Because digital cameras often have more color settings than white balance, how the settings affect the results of the algorithms needs to be examined.

The sky-color regression algorithm has been developed and optimized mainly for the images captured by standard digital single-lens reflex (DSLR) cameras equipping lenses with focal lengths of 24 to 100 mm. Adjusting parameters may be necessary for other types of images, such as total sky images. Optimizing and evaluating the algorithm for total sky images is an area for further study.

7 Conclusions

This paper considered an image processing method to extract atmospheric optical phenomena such as halos by reducing the contrast of clouds according to their color in the sky image and constructed a model of sky color. It also validated the model with halo images and implemented a new algorithm named sky-color regression. The algorithm extracts halos, including those that are difficult to distinguish with the naked eye. Even the most popular 22° halos are sometimes camouflaged by clouds. In such cases, the new algorithm might allow automated observations. In processed images, halos, which are often geometrically shaped, are more easily detected by feature-detection algorithms. The algorithm is useful not only for accurately observing the appearance of halos, but also to measure the intensity of halos by removing cloud contrasts.

Although the purpose of this work was to process sky images for easier halo detection, the model constructed and validated in this paper and the method developed can be used for other purposes as well as cloud detection. Similar models and processing can be developed that would benefit other fields.

Code availability

Sample codes for sky-color regression in Java are available on https://doi.org/10.5281/zenodo.7716821 (Ayatsuka2023).

Data availability

All the photographs used in this manuscripts can be accessed at https://doi.org/10.5281/zenodo.12181277 (Ayatsuka2024).

Competing interests

The author has declared that there are no competing interests.

Disclaimer

Publisher’s note: Copernicus Publications remains neutral with regard to jurisdictional claims made in the text, published maps, institutional affiliations, or any other geographical representation in this paper. While Copernicus Publications makes every effort to include appropriate place names, the final responsibility lies with the authors.

Acknowledgements

I thank RainbowMustache (@HIGEHIGErainbow on Twitter) for introducing me to BR processing and Kentaro Araki for advising me on where to submit this paper.

Review statement

This paper was edited by Alyn Lambert and reviewed by four anonymous referees.

References

Ayatsuka, Y.: AOI:Atmospheric Optical Image Enhancer, http://www.asahi-net.or.jp/~cg1y-aytk/ao/aoi_e.html (last access: 10 March 2023), 2022. a

Ayatsuka, Y.: y-ayatsuka/SkyColorRegression: SCR20230310, Version SCR20230310, Zenodo [code], https://doi.org/10.5281/zenodo.7716821, 2023. a

Ayatsuka, Y.: Photos data for “Revealing Halos Concealed by CIrrus Clouds”, Zenodo [data set], https://doi.org/10.5281/zenodo.12181278, 2024. a

Boyd, S., Sorenson, S., Richard, S., King, M., and Greenslit, M.: Analysis algorithm for sky type and ice halo recognition in all-sky images, Atmos. Meas. Tech., 12, 4241–4259, https://doi.org/10.5194/amt-12-4241-2019, 2019.  a

Chengtao, C., Qiuyu, Z., and Yanhua, L.: A survey of image dehazing approaches, in: The 27th Chinese Control and Decision Conference (2015 CCDC), Qingdao, China, 23–25 May 2015, IEEE, 3964–3969, https://doi.org/10.1109/CCDC.2015.7162616, 2015. a

Dev, S., Lee, Y. H., and Winkler, S.: Color-Based Segmentation of Sky/Cloud Images From Ground-Based Cameras, IEEE J. Sel. Top. Appl., 10, 231–242, https://doi.org/10.1109/JSTARS.2016.2558474, 2017. a

Forster, L., Seefeldner, M., Wiegner, M., and Mayer, B.: Ice crystal characterization in cirrus clouds: a sun-tracking camera system and automated detection algorithm for halo displays, Atmos. Meas. Tech., 10, 2499–2516, https://doi.org/10.5194/amt-10-2499-2017, 2017. a

Gao, B.-C. and Li, R.-R.: Removal of Thin Cirrus Scattering Effects in Landsat 8 OLI Images Using the Cirrus Detecting Channel, Remote Sens.-Basel, 9, 834, https://doi.org/10.3390/rs9080834, 2017. a

Greenler, R.: Rainbows, Halos and Glories, Cambridge University Press, ISBN: 978-0521236058, 1980. a

Großmann, M., Schmidt, E., and Haußmann, A.: Photographic evidence for the third-order rainbow, Appl. Optics, 50, F134–F141, https://doi.org/10.1364/AO.50.00F134, 2011. a

Juutilainen, J.: Halokuvien B-R -käsitelty, https://www.taivaanalla.fi/2015/07/21/halokuvien-b-r-kasittely/ (last access: 10 March, 2023), 2015. a

Koehler, T. L., Johnson, R. W., and Shields, J. E.: Status of the whole sky imager database, in: Proceedings of the Cloud Impacts on DOD Operations and Systems 1991 Conference (CIDOS-91), El Segundo, California, 9–12 July 1991, Department of Defense, 77–80, 1991. a

Lawson, R. P., Baker, B. A., Zmarzly, P., O'Connor, D., Mo, Q., Gayet, J.-F., and Shcherbakov, V.: Microphysical and Optical Properties of Atmospheric Ice Crystals at South Pole Station, J. Appl. Meteorol. Clim., 45, 1505 – 1524, https://doi.org/10.1175/JAM2421.1, 2006. a

Lefaudeux, N.: “Blue minus Red” processing, http://opticsaround.blogspot.com/2013/03/le-traitement-bleu-moins-rouge-blue.html (last access: 10 March 2023), 2013. a

Li, Q., Lu, W., and Yang, J.: A Hybrid Thresholding Algorithm for Cloud Detection on Ground-Based Color Images, J. Atmos. Ocean. Tech., 28, 1286–1296, https://doi.org/10.1175/JTECH-D-11-00009.1, 2011. a

Lynch, D. K. and Schwartz, P.: Intensity profile of the 22° halo, J. Opt. Soc. Am. A, 2, 584–589, https://doi.org/10.1364/JOSAA.2.000584, 1985. a

Moilanen, J. and Gritsevich, M.: Light scattering by airborne ice crystals – An inventory of atmospheric halos, J. Quant. Spectrosc. Ra., 290, 108313, https://doi.org/10.1016/j.jqsrt.2022.108313, 2022. a

Riikonen, M., Sillanpää, M., Virta, L., Sullivan, D., Moilanen, J., and Luukkonen, I.: Halo observations provide evidence of airborne cubic ice in the Earth's atmosphere, Appl. Optics, 39, 6080–6085, https://doi.org/10.1364/AO.39.006080, 2000. a

Rossetto, N.: Complex display in Observatoire de Paris [BW], https://www.flickr.com/photos/gaukouphoto/6320708287/in/photostream (last access: 10 March 2023), 2011. a

Sassen, K.: Remote Sensing of Planar Ice Crystal Fall Attitudes, J. Meteorol. Soc. Jpn., Ser. II, 58, 422–429, https://doi.org/10.2151/jmsj1965.58.5_422, 1980. a

Sassen, K., Knight, N. C., Takano, Y., and Heymsfield, A. J.: Effects of ice-crystal structure on halo formation: cirrus cloud experimental and ray-tracing modeling studies, Appl. Optics, 33, 4590–601, https://doi.org/10.1364/AO.33.004590, 1994. a

Shi, F., Jia, Z., Lai, H., Song, S., and Wang, J.: Sand Dust Images Enhancement Based on Red and Blue Channels, Sensors, 22, 1918, https://doi.org/10.3390/s22051918, 2022. a

Singh, P., Khan, E., Upreti, H., and Kapse, G.: Survey on Image Fog Reduction Techniques, International Journal of Computer Science and Network, 5, 302–306, 2016.  a

Um, J. and McFarquhar, G. M.: Formation of atmospheric halos and applicability of geometric optics for calculating single-scattering properties of hexagonal ice crystals: Impacts of aspect ratio and ice crystal size, J. Quant. Spectrosc. Ra., 165, 134–152, https://doi.org/10.1016/j.jqsrt.2015.07.001, 2015. a

van Diedenhoven, B.: The prevalence of the 22° halo in cirrus clouds, J. Quant. Spectrosc. Ra., 146, 475–479, https://doi.org/10.1016/j.jqsrt.2014.01.012, 2014. a

1

“Halos” are typically used to refer to sun-centered rings, while “arcs” refer to the other type of atmospheric phenomena caused by ice crystals. However, in this paper, we will use “halos” to refer to both.

2

See https://atoptics.co.uk/blog/opod-helic-lowitz-arcs-france/ (last access: 20 June 2024) or https://atoptics.co.uk/blog/greek-halos-opod/ (last access: 20 June 2024) for examples.

Download
Short summary
Many types of halos appear in the sky. Each type of halo reflects the state of the atmosphere; therefore observing them from the ground greatly helps in understanding the state of the atmosphere. However, halos are easily obscured by the contrast of the cloud itself, making it difficult to observe them. This study describes the construction of a sky-color model for halos and a new effective algorithm to reveal halos in images.