the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
IceDetectNet: a rotated object detection algorithm for classifying components of aggregated ice crystals with a multi-label classification scheme
Xia Li
Fabiola Ramelli
Robert O. David
Julie Pasquier
Jan Henneberger
The shape of ice crystals affects their radiative properties, growth rate, fall speed, and collision efficiency; thus, it plays a significant role in cloud optical properties and precipitation formation. Ambient conditions, like temperature and humidity, determine the basic habit of ice crystals, while microphysical processes, such as riming and aggregation, further shape them, resulting in a diverse set of ice crystal shapes and effective densities. Current classification algorithms face two major challenges: (1) ice crystals are often classified as a whole (at the image scale), necessitating identification of the dominant component of aggregated ice crystals, and (2) single-label classifications lead to information loss because of the compromise between basic habit and microphysical process information. To address these limitations, we present a two-pronged solution here: (1) a rotated object detection algorithm (IceDetectNet) that classifies each component of an aggregated ice crystal individually and (2) a multi-label classification scheme that considers both basic habits and physical processes simultaneously. IceDetectNet was trained and tested on two independent datasets obtained by a holographic imager during the NASCENT campaign in Ny-Ålesund, Svalbard, in November 2019 and April 2020. The algorithm correctly classified 92 % of the ice crystals as either aggregate or non-aggregate and achieved an overall accuracy of 86 % for basic habits and 82 % for microphysical process classification. At the component scale, IceDetectNet demonstrated high detection and classification accuracy across all sizes, indicating its ability to effectively classify individual components of aggregated ice crystals. Furthermore, the algorithm demonstrated a good generalization ability by classifying ice crystals from an independent generalization dataset with overall accuracies above 70 %. IceDetectNet can provide a deeper understanding of ice crystal shapes, leading to better estimates of ice crystal mass, fall velocity, and radiative properties; therefore, it has the potential to improve precipitation forecasts and climate projections.
- Article
(6269 KB) - Full-text XML
- BibTeX
- EndNote
The shape of ice crystals within clouds impacts the Earth's radiation budget (Ehrlich et al., 2008; Sun and Shine, 1994; Matus and L'Ecuyer, 2017; Flanner et al., 2007; Järvinen et al., 2018). As ice crystals interact with solar and terrestrial radiation, they scatter, absorb, and emit radiation, thereby influencing the radiative properties of the atmosphere (Flanner et al., 2007; Järvinen et al., 2018; Yang et al., 2015). Furthermore, the shape of ice crystals has a substantial effect on global precipitation, influencing both the spatial distribution and precipitation rate (Sterzinger and Igel, 2021; Woods et al., 2007; Jensen et al., 2017). The growth mechanisms of ice crystals play a crucial role in precipitation formation (Wegener, 1911; Findeisen, 1938; Lohmann et al., 2016; Kalina and Puxbaum, 1994; Mosimann et al., 1993). The efficiency of these processes is largely determined by the ice crystal shape, further highlighting its importance (Heymsfield, 1972; Khvorostyanov and Curry, 2002; Bailey and Hallett, 2004; Mitchell, 1996; Mitchell et al., 1990; Wang and Ji, 2000).
The initial shape of an ice crystal, also known as its basic habit (e.g., column or plate), is governed by the ambient meteorological conditions, such as the temperature and supersaturation, that it experiences during its initial diffusional growth phase (Libbrecht, 2016; Bailey and Hallett, 2004). The change in the ambient environment, such as in a convective system, leads to a complex basic habit, such as columns on capped columns (CPCs; observed by Pasquier et al., 2023). They are further shaped by microphysical processes, including riming (i.e., supercooled cloud droplets collide and freeze on the ice crystal) and aggregation (i.e., individual ice crystals collide and stick together). This leads to a wide range of ice crystal shapes, sizes, and densities, introducing considerable challenges with respect to the systematic classification of ice crystals.
Early ice crystal classification techniques used simple features like edge complexity (Cunningham, 1978), circular deficiency (Rahman et al., 1981), the surface area and perimeter (Duroure et al., 1994), and the complexity (combined of several geometric features, such as the particle area and area ratio) (Schmitt and Heymsfield, 2014) to classify the shape of ice crystals, but they cannot distinguish between composite ice crystals, such as irregular crystals, aggregates, or bullet rosettes. More advanced techniques, like ice crystal classification with principal component analysis (Lindqvist et al., 2012) and logistic regression (Praz et al., 2017), have been developed and have achieved 80 %–90 % accuracy, but they still require manual feature extraction (e.g., aspect ratio). Furthermore, these algorithms have demonstrated limitations with respect to their ability to perform effectively on different datasets, as their classification performance is strongly influenced by the characteristics of the training dataset (Bishop and Nasrabadi, 2016; Goodfellow et al., 2016), which is defined as the generalization ability of the models. This dependency requires significant adjustments to the optimal thresholds when these algorithms are applied to new, unseen datasets.
The emergence of convolutional neural networks (CNNs) as part of deep learning algorithms has introduced significant improvements to the classification of ice crystal habits, with their capability for automated feature extraction (Li et al., 2021; Albawi et al., 2017; Touloupas et al., 2020). Although CNNs exhibit a remarkable capacity to recognize key aspects of images, they struggle when faced with complex ice crystals such as CPCs or aggregates consisting of different basic habits (Zhang, 2021). Furthermore, CNNs that are based on single-label classification schemes face the challenge of information loss when composite ice crystals are classified (Zhang, 2021; Xiao et al., 2019). For example, an aggregated column can only be labeled either as “aggregate” or column, which results in information loss with respect to either the basic habit or the microphysical process. According to the study by Korolev et al. (1999), in Arctic clouds, pristine ice habits (ice crystals that have not undergone any microphysical processes) account for only 3 % of the particles observed, which would result in losing a substantial fraction (97 %) of ice information regarding either basic habits or microphysical processes when implementing a single-label classification scheme if the ice habits are still recognizable. Moreover, in stratiform clouds, Korolev et al. (2000) found that 84 % of the ice crystals are irregular ice, which is everything except needles and dendrite. These irregular ice crystals would be either aged or aggregated by our definition (see Sect. 2 and Appendix B). To tackle this problem, Zhang et al. (2022) first proposed that the ice shape label should contain two types of information: the basic habits and the microphysical processes that the ice has experienced. Thus, they recommend that a multi-label should be assigned to one ice crystal. Jaffeux et al. (2022) combined data from the Precipitation Imaging Probe and 2DS-Stereo Probe to train CNNs to classify ice crystals according to their basic habit and the occurrence of riming and aggregation. Although their study considered potential microphysical processes for each ice crystal category manually after the CNN classification, the specific microphysical processes associated with individual components of an aggregated ice crystal remained unknown.
To summarize, there are two key limitations of current ice crystal classification algorithms:
-
Algorithms often classify the images of an ice crystal as a whole, necessitating the identification of the dominant component of an aggregated ice crystal; thus, they are not able to account for the presence of multiple basic habits in an aggregated ice crystal.
-
Single-label classification algorithms require a compromise between basic habit and microphysical process information, leading to information loss.
To address these issues, we propose a novel approach that consists of a rotated object detection algorithm (called IceDetectNet) along with a multi-label classification scheme. IceDetectNet can classify the ice crystals down to the scale of aggregated ice crystal components with both basic habit and microphysical process information, thereby eliminating the need to identify a dominant aggregated component. The multi-label classification scheme simultaneously accounts for both the basic habits and microphysical processes of an ice crystal, reducing information loss. However, like all supervised learning methods, our approach is limited to the ice categories present in the training dataset, limiting its applicability until the model is fine-tuned on a new dataset. The data used to train and test IceDetectNet are described in Sect. 2. The structure of IceDetectNet is presented in Sect. 3. The performance of our proposed algorithm is evaluated in Sect. 4. Finally, Sect. 5 and Sect. 6 present the relevant discussion and conclusions of this study, respectively.
The data used in this study were collected in Arctic mixed-phase clouds during the NASCENT campaign (Pasquier et al., 2022a) conducted in Ny-Ålesund, Norway. Ice crystal images were captured by the HOLIMO holographic imager mounted on the HoloBalloon tethered balloon system (Ramelli et al., 2020). The measured ice particle sizes ranged from 50 µm to 2.4 mm. First, the cloud particles were classified as cloud liquid droplets and ice crystals using a convolutional neural network (CNN) approach, as described in Touloupas et al. (2020). This preliminary classification served as the basis for the subsequent detailed classification of the ice crystals.
Following this initial categorization, each ice crystal was classified into one of seven basic habits: column, plate, lollipop (Pasquier et al., 2022a), CPC, irregular, frozen droplets, and small. Our seven basic habit categories were determined by their presence and distinct shape features observed in our dataset collected in the aforementioned Arctic mixed-phase clouds in Ny-Ålesund. These basic habit classes are based on the categories used in Pasquier et al. (2022a), as we used the same dataset. Additionally, up to two microphysical process attributes (i.e., aggregate and aged) were assigned to each ice crystal. Table 1 describes the seven basic habits and four microphysical process categories (i.e., pristine, aged, aggregate, and aged and aggregate). Thus, the final habit classification of an ice crystal is a combination of the basic habit and microphysical processes (i.e., the final classification = basic habit + microphysical processes). Not all combinations of basic habits and microphysical processes are feasible, resulting in a total of 19 ice classes (examples are shown in Fig. B1) rather than the theoretically possible 28 (7 × 4) categories. For instance, the small class refers ice crystals that are too small to determine their habit, making it impossible to derive their microphysical processing. Furthermore, all ice crystals in the lollipop, CPC, and irregular classes are defined as aged ice instead of pristine ice (as they are not newly produced ice), although they are still basic habit categories. For a more detailed discussion of the categorization criteria, we refer to Appendix B. However, due to data limitations, our dataset does not capture every possible basic ice habit, such as needles and rosettes, although the existence of these ice habits is well acknowledged (Kikuchi et al., 2013). This limitation is acknowledged and further discussed in Sect. 5, where we look at potential extensions to IceDetectNet. As new data containing additional ice habits become available, IceDetectNet can be updated, as it is designed to incorporate these new habits through fine-tuning, ensuring the continued evolution of the model.
(Pasquier et al., 2023)(Keppas et al., 2017)The dataset collected on 11 November 2019 (Pasquier et al., 2022b), hereafter the training dataset, was used to train IceDetectNet. During the training, it was divided into a training subset (comprising 80 % of the data) and a validation subset (made up of the remaining 20 %), using a cross-validation method (see the detailed introduction in Sect. 3.8.1). This validation subset serves a similar purpose to the traditional test sets used in other studies (Jaffeux et al., 2022; Xiao et al., 2019; Touloupas et al., 2020), providing an initial evaluation of the model's performance under known conditions. On the other hand, the generalization dataset was collected on a different date, 1 April 2020 (Pasquier et al., 2022b), and is not used during training; rather, it is employed to evaluate the generalization abilities of IceDetectNet.
Table 2 offers a summary of both the training and generalization datasets. The training dataset consists of 18 864 ice particles, where the column and CPC classes were dominant, accounting for 47.5 % of the ice crystals in the training data. Non-pristine ice, which is ice that is not freshly formed, accounts for 70.5 % of the ice crystals in the training data. Additionally, 18.8 % of the non-pristine ice crystals have undergone two microphysical processes, and aggregated ice makes up 12 % of the ice crystals in the training dataset.
In contrast, the generalization dataset has a significant fraction of irregular (47.3 %) and small (23.4 %) ice crystals (Fig. B1). Unlike the training dataset, the generalization dataset does not include any instances of the lollipop or CPC classes and, consequently, the corresponding compound categories “lollipop–aggregate” or “CPC–aggregate” do not exist. Moreover, the generalization dataset only contains three occurrences of the frozen droplet class, with small numbers for “frozen droplet–aged” (8) and “frozen droplet–aged–aggregate” (12), while “frozen droplet–aggregate” is not present. The proportion of non-pristine ice increases from 70.5 % in the training dataset to 93.9 % in the generalization dataset. The fraction of aggregate ice increases from 11.9 % in the training dataset to 37.7 % in the generalization dataset.
The difference between the training and generalization datasets is an example of the natural variability in field observations. In our case, the two datasets were collected during different seasons, resulting in variations in the environmental conditions. The training dataset was collected in the temperature range from −8 to −3 °C (mostly in the column regime), whereas the generalization dataset was collected between −23 and −15 °C (mostly in the plate regime) (Pasquier et al., 2022b). These differences allow us to assess the generalization ability of IceDetectNet and to examine its performance under diverse environmental conditions.
In this section, we first provide an overview of convolutional neural networks (CNNs), which serve as the foundation for the object detection algorithm in IceDetectNet. We start by explaining the overall structure of CNNs (Sect. 3.1), followed by an introduction to the rotated object detection algorithm developed in this study and implemented into IceDetectNet (Sect. 3.2). Subsequently, we discuss the data preparation (Sect. 3.4) and training process (Sect. 3.7), outlining the essential steps for training the model. Lastly, the evaluation metrics are introduced that are used to assess the performance of IceDetectNet on the training dataset and its ability to generalize to the unseen generalization dataset (Sect. 3.8).
3.1 Convolutional neural networks (CNNs)
CNNs are a class of neural networks widely recognized for their exceptional performance in image classification tasks (Gu et al., 2018; Albawi et al., 2017; Rawat and Wang, 2017; Touloupas et al., 2020). CNNs consist of a specific architecture designed to extract meaningful features from images. The key components of a typical CNN include convolutional layers, pooling layers, and fully connected layers (Fig. 1). These layers work together to enable effective image analysis. The convolutional layer scans the input image with a small filter or kernel, extracting low-level features such as edges and color. The pooling layer reduces the spatial size of the convolved feature (feature map shown in Fig. 1) and aims to decrease computational complexity. The fully connected layer, which is usually the final layer of a CNN, performs the classification using the flattened or pooled output from the preceding layers.
In practice, CNN structures can be much more complex than the basic CNN described above. He et al. (2016) proposed a deep residual learning approach that stores input information and propagates it directly from the first layer to the last. This approach has been successfully used in subsequent object detection algorithms that utilize the ResNet-50 structure (He et al., 2016). Due to this success, IceDetectNet (described in Sect. 3.5) is trained using the preexisting parameters of ResNet-50, which was trained on the ImageNet dataset consisting of approximately 1.3 million images labeled into 1000 categories, with the exception of the last layer due to the different number of categories in ice classification (Deng et al., 2009; He et al., 2016). This helps to speed up the training process and achieve better performance.
3.2 Rotated object detection algorithm
Building on the foundations of CNNs, object detection algorithms serve as an extension to detect and classify objects within images. While CNNs typically classify the image as a whole, object detection algorithms localize and classify specific objects within these images, providing both their location (through the bounding box) and class labels (Zhao et al., 2019). The rectangular box that tightly encloses the object of interest is called a bounding box (as shown in Fig. 2). Rotated object detection algorithms additionally predict the angle of rotation of the bounding box (Zou et al., 2023). In this study, we introduce such a rotated object detection algorithm for ice crystal classification as part of IceDetectNet. This algorithm classifies multiple components of aggregated ice crystals individually and predicts the center, the dimensions, and the rotation angle of bounding boxes enclosing the ice crystal components. This ensures that the ice components are captured within the smallest feasible rectangle, which offers a more accurate recognition of the object by minimizing the inclusion of background pixels (Ding et al., 2019). Here, we use the S2ANet network structure (Han et al., 2021) as the base structure of the rotated object detection algorithm within IceDetectNet.
3.3 Hand-labeling of bounding boxes and ice categories
Accurate hand labeling is essential for training IceDetectNet, as for all supervised learning methods. In contrast to conventional classification algorithms, rotated object detection models such as IceDetectNet require a two-step hand-labeling process: (1) locate the ice components within the images by drawing bounding boxes and (2) assign the appropriate category labels to each component based on our multi-label classification scheme.
In the present study, both the training and generalization datasets were initially hand-labeled at the image scale using our multi-label ice classification scheme (Sect. 2). For the image-scale hand-labeling, the basic habit of the largest ice component is considered the basic habit of the image. Regarding microphysical processes, any image containing an ice component with signs of aging was labeled “aged”. Additionally, images consisting of multiple ice components were categorized as “aggregated”. The image-scale hand-labeling served as the basis for the component-scale hand-labeling.
For the hand-labeling of non-aggregated ice crystals (identified using the hand label at the image scale), we applied an automated method for drawing bounding boxes. The method uses the color contrast between the typically black ice pixels and the typically gray background to identify the location of the ice component regions. Then, we calculated the minimum bounding rectangle of the ice regions automatically as the bounding box of these non-aggregated ice crystals. The non-aggregated ice crystals were assigned the same ice category labels as the corresponding labels at the image scale.
For aggregated ice crystals, the hand-labeling of the bounding boxes (i.e., drawing) was done manually using the platform provided by AngoAI (2022), as illustrated in Appendix A and Fig. A1. Bounding boxes representing the minimum enclosing rectangle of each ice component were manually drawn. Furthermore, every bounding box was visually classified in an ice category following the multi-label classification scheme introduced in Sect. 2. In total, we manually labeled 2255 aggregated ice crystals and 16 609 non-aggregated ice crystal components. Note that the component-scale hand-labeling was only done for the training dataset due to the large effort involved with the hand-labeling of bounding boxes and ice categories. The generalization dataset was only hand-labeled at the image scale.
3.4 Image preprocessing
Before the image is fed into IceDetectNet, the initial image is enlarged by 15 % (with respect to both length and width), to ensure full bounding box coverage and maintain the aspect ratio (see input in Fig. 2), by adding black pixels (pixel values = 1) around the borders. This augmentation ensures that the entire bounding box is located within the image, even when parts of the bounding box extend beyond the original image frame. To ensure consistency across the network for training and testing, all images are then uniformly resized to 512 px × 512 px by employing bilinear interpolation after the enlargement.
The input images are normalized to meet the pretrained ResNet-50 model's input specifications. For example, the pretrained ResNet-50 model requires an RGB image as input, which consists of three dimensions by default. Given that our images only have one dimension, we replicate the single dimension three times to emulate the three-dimensional structure of RGB images and to produce pseudo-RGB images.
3.5 Inference process of IceDetectNet
The structure of the IceDetectNet algorithm is shown in Fig. 2. The input for IceDetectNet (step 1 in Fig. 2) comprises the processed images undergoing the preprocessing steps described in Sect. 3.4. The algorithm uses the ResNet-50 backbone network (He et al., 2016) to extract image features at a per-pixel scale. After feature extraction, IceDetectNet predicts potential bounding boxes for individual ice components (step 2 in Fig. 2). These predicted bounding boxes contain the location, size, and rotation angle of the respective ice components. Multiple bounding boxes might be predicted for the same ice component, or some predicted bounding boxes may be too large to tightly capture an ice component, while others could be too small, missing some parts of an ice component (step 2 in Fig. 2). Prior to classification, duplicate bounding boxes are removed (step 3 in Fig. 2) by a feature alignment module (Sect. 3.6). The remaining predicted bounding boxes enclosing the individual ice components are the input for the classification module (step 4 in Fig. 2), which outputs a predicted ice label with a confidence level (step 5 in Fig. 2). After classification, a post-processing step is performed to further remove duplicate bounding boxes (step 6 in Fig. 2) by comparing all predicted bounding boxes instead of hand-labeled bounding boxes using the intersection over union (IoU) threshold and the confidence level of classification. The IoU quantifies the overlap between two predicted bounding boxes and is calculated as follows:
If the IoU between two predicted bounding boxes exceeds a threshold, the bounding box with the lower confidence level is discarded. This threshold was set to 50 % in the present study to minimize the number of aggregated ice crystals misclassified as non-aggregated and vice versa. Previous studies suggested that an IoU threshold within the range of 50 % to 75 % leads to the best performance (Zhang et al., 2019).
3.6 Training process of IceDetectNet
The training phase of IceDetectNet is a crucial and complex process, as it requires a careful balance between reducing detection errors (step 2 in Fig. 2) and classification errors (step 5 in Fig. 2), all via a process of loss minimization. The feature alignment module (step 3 in Fig. 2) is trained to reduce the difference between the predicted and hand-labeled bounding boxes. If the IoU between a predicted and a hand-labeled bounding box is above 50 %, the prediction is considered correct (Zhang et al., 2019). On the contrary, an IoU below 50 % indicates an incorrect prediction, resulting in a loss that the training process then aims to minimize. To reduce the loss, a method called backpropagation is employed. Backpropagation adjusts the model's parameters to reduce errors and refines the feature extraction to improve the accuracy of the bounding box prediction in step 2. The primary objective of the feature alignment module is to fine-tune the orientation and position of bounding boxes, especially those with an IoU below 50 %. The images within the refined bounding boxes are then processed by the classification module. Incorrectly predicted labels contribute to the model's loss in classification, which is further minimized using backpropagation as well, leading to improved classification in step 5. Steps 6 and 7 belong to post-processing and are not subject to training.
3.7 Training details
The training details of IceDetectNet and the hyperparameters used during training are described here. To prevent the model from memorizing the training data (a problem known as “overfitting”) and to lower the generalization errors when it is applied to new unseen data, we introduced transformations to our training images. More specifically, we applied a technique called data augmentation which performs random flips of the images in the horizontal, vertical, and diagonal directions with a 25 % probability. During the inference, no data augmentation was applied to prevent any distortion in the final output.
The training was executed on a computational system equipped with four RTX 2080 GPUs. A batch size of 64 was chosen to optimize computational efficiency and training stability.
The learning rate underwent a structured adaptation during the training process as follows:
-
Initially, the learning rate was set to 0 and linearly increased to 0.0025 over the initial 500 steps. Here, a “step” is defined as a single iteration in the training process, in which one batch of data is processed to update IceDetectNet's parameters.
-
After the first 500 steps, the learning rate was kept constant at 0.0025.
-
A reduction by a factor of 10 was applied at specific epochs; the learning rate was set to 0.00025 from the 64th to the 88th epochs and further reduced to 0.000025 after the 88th epoch. This decremental strategy aimed at refining model parameters with progressively smaller updates as the training advanced.
To ensure model robustness and prevent overfitting, we employed the early-stopping technique (Jabbar and Khan, 2015). Checkpoints were integrated to retain the best-performing model based on the validation dataset.
3.8 Evaluation
To evaluate the performance of IceDetectNet, we used a cross-validation approach (Sect. 3.8.1) and a range of evaluation metrics (Sect. 3.8.2): overall accuracy, precision, recall, and the confusion matrix, as highlighted in the following sections.
3.8.1 Cross-validation
IceDetectNet was validated by applying a 5-fold cross-validation approach, which is a method that minimizes the variance in performance estimation while also maximizing the use of available data for training (Browne, 2000). Here, we randomly partitioned the dataset into five equally sized subsamples or “folds”. Four of the five subsamples were used for training the model, while the remaining subsample was retained as the validation data for testing the model. This process was then repeated five times (the folds), with each of the five subsamples used exactly once for validation. Each fold was designed to include images from every class, thus ensuring that the model was trained and evaluated on a diverse set of ice crystals from all categories. This helped to prevent bias in the evaluation of the model's performance due to an unrepresentative selection of training and test data (Arlot and Celisse, 2010). The five outcomes from the folds were then averaged to produce a single estimation of IceDetectNet's performance.
In future applications, if the performance of the five individual models is similar (performance is evaluated in Sect. 4), a single model could be selected for use, simplifying the process. Alternatively, the ensemble of five models can be used to make the final predictions, increasing reliability. For example, if three models predict one ice component as a column but two predict it as a plate, the component is predicted as a column. The advantage of this method over repeated random subsampling is that all ice crystals are used for both training and validation and each ice crystal is used for validation exactly once. This method, although computationally expensive, provides a robust evaluation of IceDetectNet's performance and its ability to generalize to new, unseen data (Arlot and Celisse, 2010).
3.8.2 Evaluation metrics
To assess the performance of IceDetectNet, we employ several metrics that evaluate the model performance with regard to different aspects, including the overall accuracy, precision, recall, confusion matrix, and F1 score. The overall accuracy is defined as the ratio of the number of correct predictions to the total number of particles (Goodfellow et al., 2016). An overall accuracy of 100 % means that, for example, all ice particles were correctly predicted, whereas an overall accuracy of 0 % indicates that all particles were mispredicted. While overall accuracy provides a quick and straightforward metric to interpret the model performance, it can be misleading when dealing with imbalanced datasets where classes are not equally represented. In such cases, the model may perform well with respect to predicting the dominant classes but struggle to predict rare classes. Precision and recall both measure the accuracy of a deep learning classification model with respect to predicting a single category from two perspectives. Precision is calculated as the ratio of the number of correct predictions of a specific class to the total number of predictions (Goodfellow et al., 2016), while recall is defined by the ratio of the number of correct predictions of a specific class to the total number of this class (Goodfellow et al., 2016). A high precision score indicates effective identification of a specific class, while a high recall score indicates that the model excels in identifying instances of a particular class and is less likely to miss relevant instances that belong to the class. All of these metrics can be combined and visualized in a so-called confusion matrix (Goodfellow et al., 2016). In a confusion matrix, the diagonal, from the top left to the bottom right, corresponds to correct predictions made by the model, whereas the elements outside of this diagonal represent misclassifications. The bottom-right cell of the matrix displays the total number of ice crystals and the overall accuracy. The bottom row provides the actual counts per class and their respective per-class precision. Similarly, the rightmost column presents the predicted counts per class and the associated per-class recall. The F1 score is a harmonized metric that combines precision and recall, providing a balanced measure of a model's performance, particularly in situations where the balance between precision and recall is critical (Goodfellow et al., 2016). This score reaches its best value at 1 (indicating perfect precision and recall) and its worst value at 0. In the context of IceDetectNet, a high F1 score would indicate not only that the model accurately identifies ice particles (high precision) but also that it successfully detects the majority of actual ice particles (high recall), making it a robust metric for evaluating model performance across different classes, especially in the presence of imbalanced datasets.
4.1 Evaluation of model performance
The evaluation of IceDetectNet differs from traditional deep learning classification algorithms because both detection (Sect. 4.1.1) and classification (Sect. 4.1.2) steps need to be evaluated. As described in Sect. 3.8.1, we trained five models using a 5-fold cross-validation approach. Four folds were used for training, and the remaining fold was used for validation. A small portion of the images for which no bounding boxes were predicted (validation fold in training dataset: 11 of 3755) were labeled as “none” and excluded from the following analysis.
4.1.1 Performance of aggregate detection
The detection performance of IceDetectNet was examined by first evaluating the ability of the algorithm to distinguish between aggregate and non-aggregate ice. Images with a single bounding box were defined as non-aggregate ice, whereas images with multiple bounding boxes were defined as aggregate ice. The aggregate/non-aggregate detection was evaluated by comparing the number of predicted bounding boxes with the number of hand-labeled bounding boxes for the training dataset (Fig. 3a). Hand-labeled and predicted bounding boxes are in good agreement, with an overall accuracy of 92 %, reflecting the ability of IceDetectNet to correctly classify images as aggregated or non-aggregated ice.
To understand the source of the 8 % of misdetected aggregate and non-aggregate ice, we analyzed the number of overdetected and underdetected bounding boxes. Here, we consider it overdetection when the algorithm predicts multiple bounding boxes for an ice crystal that is hand-labeled as non-aggregate (i.e., one bounding box) and underdetection when the algorithm predicts one bounding box for an ice crystal that is hand-labeled as aggregate (i.e., multiple bounding boxes). In absolute numbers, there were 266 instances of overdetection and 63 instances of underdetection (Fig. 3b). In relative terms, 40 % of the predicted aggregates were hand-labeled as non-aggregates (overdetection), resulting in a recall of 60 % for actual non-aggregates. Conversely, only 2 % of the predicted non-aggregates were hand-labeled as aggregates (underdetection), indicating a high recall of 98 % for actual aggregates (Fig. 3c). This shows that, while it tends to overestimate the presence of aggregates, the model is highly effective at identifying actual aggregates. Considering that only 12 % of the training dataset was aggregate ice (as detailed in Sect. 2), a few mispredictions of non-aggregate ice as aggregate can significantly increase the overdetection. For example, if 2 % (66) of the non-aggregate ice crystals were misclassified as aggregate, this would lead to a 14.6 % overdetection. Until this point, we have evaluated how well the predicted categories match the actual categories. As a complement, we now turn our attention to precision, examining the accuracy of the predictions in terms of correctly identified categories. The precision is 88.7 % for aggregate (i.e., 88.7 % of the predicted aggregates were hand-labeled as aggregates) and 77.1 % for non-aggregate, consistently showing the model's tendency to overpredict the numbers of bounding boxes.
To address the issue of the overdetection or underdetection of bounding boxes, it is possible to adjust the IoU threshold in the post-processing (as introduced in Sect. 3.2). In the present study, an IoU threshold of 50 % was applied to remove duplicate bounding boxes (see Fig. 2), but the IoU threshold can be changed based on the relative composition of the ice classes in the dataset. Generally, when an overdetection problem was identified, a higher IoU threshold could be implemented to reduce the number of detected bounding boxes, and the opposite adjustment could be made if underdetection was observed. Thus, the IoU threshold can be used as a tuning parameter to reduce/increase the number of bounding boxes kept after the post-processing.
4.1.2 Performance of ice classification
The classification performance of IceDetectNet was examined by quantifying the accuracy with which detected components are categorized into their respective basic habit and microphysical process classes (following the multi-label classification scheme detailed in Sect. 2). As discussed in Sect. 3.3, the basic habit of an ice crystal (i.e., image scale) is determined by the largest bounding box. The presence or absence of an aged classification is based on the detection of aging signatures among all bounding boxes, while aggregation is defined by ice crystals with more than one bounding box. We evaluated the performance of the five trained models using the cross-validation approach described in Sect. 3.8.1. The overall accuracies for the five models; their mean values; and the mean F1 scores for ice multi-label classification (19 classes), basic habit classification (7 classes), and microphysical process classification (4 classes) are shown in Table 3. The mean overall accuracies range between 78 % (for multi-label classification) and 86 % (for basic habit classification) and, thus, indicate good classification performance. Standard deviation values of the overall accuracy (OA) below 1 % on all data demonstrate robust results among the five models. In contrast, the performance on average F1 scores, ranging from 54.9 % (for multi-label classification) to 78.8 % (for basic habit classification) is generally worse than that on average OA values, indicating that IceDetectNet performs worse on rare classes. Furthermore, the low standard deviation values of averaged F1 scores (below 2 %) on all data further indicate that the five models have relatively similar results.
To gain further insights into IceDetectNet's performance in each ice category, we analyzed the confusion matrices (mean of five models) for basic habit classification (Fig. 4) and microphysical process classification (Fig. 5). IceDetectNet achieved an overall accuracy of 86 % for the basic habit categories (Fig. 4). The confusion matrix shows that IceDetectNet performed well for the ice categories that are represented as a large fraction in the dataset, like column (precision of 90 %, 1978 instances) and small (precision of 93 %, 324 instances). However, IceDetectNet encountered challenges with respect to accurately classifying rare classes such as plate (67 %, 47 instances) and lollipop (77 %, 73 instances). The main source of misclassification for plate ice crystals was confusion with the column class (7 instances). For the microphysical process category (Fig. 5), IceDetectNet achieved an overall accuracy of 82 %. While the model performed well with respect to identifying pristine ice crystals (92 %), it showed worse performance for predicting aggregate (47 %) and aged–aggregate (49 %) ice crystals. This might be explained by the imbalanced dataset: pristine ice crystals dominated with a total contribution of 66 %. This suggests that balancing the dataset could further optimize IceDetectNet's classification performance for aggregate and aged–aggregate ice crystals in future iterations. A closer examination of the misclassified ice crystal images shows that the primary source of error was an underdetection of the number of bounding boxes. For example, in an image containing two aged columns, only one column–aged crystal was detected resulting in it being mislabeled as aged instead of aged–aggregate.
To investigate the classification performance of IceDetectNet in simpler scenarios, we evaluated the performance on non-aggregated ice and aggregated ice separately (Table 4). The evaluation on non-aggregated ice provides a benchmark, as non-aggregated ice images consist of a single ice component and, thus, allow us to compare the performance to traditional deep learning algorithms. When considering only non-aggregated ice crystals, IceDetectNet has an accuracy of 82 % for all data, 90 % for basic habits, and 85 % for microphysical processes (Table 3). Previous studies using single-label classification (Xiao et al., 2019; Jaffeux et al., 2022; Zhang, 2021) have reported overall accuracies above 90 %, which is higher compared with IceDetectNet for all data (82 %). However, for the multi-label classification, IceDetectNet classifies both basic habits and microphysical processes. While non-aggregated ice does not have an aggregation process, aging processes are still present. When one considers solely the basic habit classification, the accuracy of IceDetectNet (90 %) for non-aggregated ice closely aligns with the results reported in the aforementioned studies. Thus, under the same classification domain, IceDetectNet performs competitively with existing classification models and offers additional information regarding microphysical processes.
When shifting our focus to aggregated ice, the inherent complexity of classifying the multiple components of an aggregate becomes evident due to a decreased classification accuracy (Table 3). The accuracy drops to 50 % for multi-label, 72 % for basic habit, and 52 % for microphysical processes. The reduction in performance between non-aggregate and aggregate ice subsets was more pronounced for the classification of microphysical processes (85 %–52 %), compared with the classification of the basic habit (90 %–72 %). This suggests that the reduced performance of IceDetectNet in the all-class classification can be attributed primarily to the challenges with respect to classifying microphysical processes.
4.2 Detection and classification performance on the aggregated component scale
In the previous sections, the detection and classification performance of IceDetectNet was evaluated at the image scale. In this section, our focus shifts from the image scale to the aggregate component scale, specifically to the bounding boxes of aggregate ice crystals. To evaluate the detection and classification performance at the aggregate component level, we followed a structured approach:
-
Pairing of bounding boxes. For each hand-labeled bounding box of an aggregate, we search for the predicted bounding box with the highest IoU.
-
Detection performance. The detection was considered correct if the IoU between the paired predicted and hand-labeled bounding box was larger than 0.5, otherwise the detection was considered incorrect.
-
Classification performance. The classification was considered correct if the label of the hand-labeled bounding box matched the label of the paired predicted bounding box, otherwise it was considered incorrect.
In contrast to the previous sections where the mean performance of all five models was examined, the performance was evaluated on a single model which was randomly selected (due to the robustness among the model runs). We categorized the bounding boxes into small, medium, and large using the areas of the predicted bounding boxes. The thresholds were set at the 33 % and 66 % percentiles of all bounding box areas, corresponding to below 32 331 px, from 32 331 to 71 275 px, and above 71 275 px, respectively. For a more intuitive understanding, these ranges correspond to squares with side lengths of below 180 px, from 180 to 267 px, and above 267 px, respectively.
When evaluating the detection and classification performance of IceDetectNet at the aggregate component scale for the three bounding box size categories (small, medium, large; Fig. 6), we find good detection performance among all size categories, with accuracies ranging between 84 % (small bounding box) and 72 % (large bounding box). The detection performance decreases for larger bounding box sizes, which might be explained by an increased variability in appearance, texture, and scale for larger bounding boxes. The classification accuracies for correctly detected bounding boxes ranged between 66 % and 71 %, with medium-sized boxes achieving the highest classification accuracy of 71 %. This suggests that medium boxes may offer an ideal balance between detectability and feature richness. Thus, IceDetectNet shows good detection and classification performance for bounding box sizes down to 662 px and up to 294 903 px.
In general, the detection performance of IceDetectNet at the aggregate component scale is higher (72 %–84 %) than the classification performance (66 %–71 %). When combining the detection and classification performance (i.e., detection × classification), similar overall performance to that reported in Table 3 for the aggregate subset (i.e., image scale) is obtained (50 %). For example, the accuracy for small bounding boxes was determined to be 54.6 %, derived from the product of 65 % classification accuracy and 84 % detection accuracy. The higher performance with respect to detection compared with classification suggests that the lower performance observed in the aggregate subset compared with the non-aggregate subset (as described in Sect. 4.1.2) is primarily due to misclassifications and not misdetections.
4.3 Generalization ability of IceDetectNet
4.3.1 Generalization ability on aggregate detection
The generalization ability of IceDetectNet was evaluated by applying it to an independent generalization dataset, which was not used to train the algorithm and was collected during a different season (detailed in Sect. 2). The same evaluation for detection and classification at the image scale was performed as for the training dataset (Sect. 4.1.1).
The detection performance for aggregates of the generalization dataset shows an overall accuracy of 84 %, with an overdetection error of 26 % and an underdetection error of 8 % (Fig. 7a, b). This corresponds to a 14 % decrease in overdetection and a 6 % increase in underdetection compared with the training dataset. It is consistent across both the training (as described in Sect. 4.1.1) and generalization datasets that overdetection is a larger problem than underdetection. However, the different ice category distribution in the generalization dataset, specifically the increase in the aggregate class from 12 % in the training dataset to 37.7 % in the generalization dataset, is likely responsible for the shift in underdetection and overdetection between the training and generalization datasets. Thus, when the algorithm is applied to the generalization dataset (with a higher fraction of aggregate ice crystals), the problem of overdetection is reduced but still exists.
4.3.2 Generalization ability on ice classification
The classification accuracy on the generalization dataset showed overall accuracies ranging from 67 % (for the all data category) to 80 % (for the basic habits category) (see Fig. 7c and Table 4). Consistent with the performance on the training dataset, the values of OA are still higher than the averaged F1 scores, ranging from 48.5 % (for the all data category) to 68.7 % (for the basic habit category) (see Table 4). Compared with the training dataset, the classification performance decreased. The most significant decrease of 11 % was observed in the classification of all classes on OA, while the smallest decrease was observed in the classification of basic habits (6 %) on OA and the classification of microphysical processes (2.1 %) on the averaged F1 score. The observed decrease in performance between the training and generalization datasets is small, especially when considering a domain shift, a situation in which the data distribution differs between the training and generalization datasets (Stacke et al., 2020). Specifically, the training and generalization datasets have different compositions of non-pristine ice crystals, with 71 % in the training set and 94 % in the test set. IceDetectNet's ability to adapt to different data compositions and maintain relatively high accuracy indicates the ability of the algorithm to generalize to different dataset characteristics. Within the non-aggregate subset, IceDetectNet showed good classification performance for all classes (73 %), basic habits (86 %), and microphysical processes (78 %) on OA (as summarized in Table 4). Although the averaged F1 scores are still lower than OA, the percentage decrease in the averaged F1 scores is smaller than in OA. In the aggregate subset, performance levels were lower than in the non-aggregate subsets (50 %–68 % on OA; 45 %–68 % on averaged F1 scores). The observed reduction in the aggregate subset is mainly due to a domain shift, reflecting the difference in aggregate ice fraction between the training (12 %) and generalization (38 %) datasets. The generalization dataset, with a higher proportion of aggregated ice, presented more complexity and variability. Aggregates, with their multiple bounding boxes and variable structures, are inherently more difficult to classify than non-aggregates. This increased complexity in the test set likely contributed to the drop in performance, highlighting the challenges posed by domain shifts in the data. Consistently low standard deviation values (below 3 % on OA; below 2 % on averaged F1 scores) were observed across all models in all data subsets, indicating the stable and reproducible performance of IceDetectNet. These standard deviation values represent the variability in classification performance among different runs or configurations of the IceDetectNet model.
To gain further insights into IceDetectNet's performance in each ice category, we analyzed the confusion matrices (mean of five models) for basic habit classification (Fig. 8) and microphysical process classification (Fig. 9) for the generalization dataset as well.
IceDetectNet achieved an overall accuracy of 81 % for the basic habit categories (Fig. 8) The confusion matrix shows that IceDetectNet still performed better for the ice categories that comprise a large fraction of the dataset, like small ice crystals (precision of 90 %). However, among these, column classification performance had a large performance drop (18 % decrease in precision compared with the training dataset), and its two main misprediction sources were irregular and plate ice crystals, which represent almost all of the mispredictions (29 %). This could be due to the data distribution shift from column to plate. Under the general decreasing trend among all categories, the irregular class surprisingly has a 10 % increase in precision. The main misprediction of the irregular class comes from column ice crystals in both the training dataset and generalization data, which could be the reason that IceDetectNet learned many column features in the training dataset and, thus, distributed higher weights to these column features. In contrast, the number of column ice crystals is much lower in the generalization dataset and, thus, leads to better performance with respect to the classification of irregular ice crystals. For the missing categories, like lollipop and CPC, that had no actual occurrences, IceDetectNet still predicted 113 ice crystals as lollipop and 50 as CPC, with most misclassifications being irregular ice crystals. This problem is likely due to the model's handling of sparse data and its tendency to fit irregular into these less common categories, as irregular was associated with the most complex features and, thus, any unrecognizable shape is classed as irregular.
For the microphysical process category (Fig. 9), IceDetectNet achieved an overall accuracy of 73 % (with a 9 % drop compared with the training dataset). The model still performed well with respect to identifying pristine ice crystals (82 %). In contrast, it shows better performance with respect to predicting aggregate (17 % higher than in the training dataset) and aged–aggregate (7 % higher than in the training dataset) ice crystals. This could be due to the changes in the data distribution, especially the changes in the aggregate fraction from 12 % in the training dataset to 37.7 % in the generalization dataset, which further emphasizes the importance of the balance of the dataset. After checking the main source of misprediction, we can see that underdetection still plays an important role, for example, the main source of misprediction of aggregate ice crystals is pristine ice crystals, which is a typical misprediction problem.
There are still opportunities to improve the generalization ability of IceDetectNet, particularly in the area of accurate detection and classification of aggregated ice crystals. In the case of imbalanced datasets, balancing strategies such as oversampling or undersampling techniques can be considered. Enriching the training dataset with a more comprehensive and diverse collection of ice crystal data would improve the robustness and generalization capabilities of the algorithm. However, gathering more training data is a time-consuming process; therefore, reducing the time needed for manual labeling of bounding boxes is an important task for future research. New techniques, including contrastive learning (Le-Khac et al., 2020) and unsupervised learning algorithms, should be investigated to reduce the need for extensive manual labeling. Furthermore, the efficacy of fine-tuning as an approach to include new ice crystal classes holds considerable promise for IceDetectNet. Fine-tuning – the process of adapting a pretrained model (i.e., IceDetectNet) to a new, although still related, dataset – has been validated in a variety of fields as a means of achieving improved performance with comparatively smaller datasets (Tajbakhsh et al., 2016). This technique is consistent with the broader concept of transfer learning (Pan and Yang, 2009), which has seen widespread success in adapting models to new domains of application.
As the current dataset used for the training of IceDetectNet does not include some basic habits, such as needles and rosettes, we plan to adapt IceDetectNet to include these categories once additional datasets containing these habits are available.
In this study, we introduced IceDetectNet, a novel rotated object detection algorithm that is able to classify ice crystals not only at the image scale but also down to the aggregate component scale. The algorithm was used in combination with a multi-label classification scheme that assigns both a basic habit and microphysical processes to each ice component. The algorithm was trained and tested on two independent holographic ice crystal datasets, which were collected during the NASCENT campaign in Ny-Ålesund, Svalbard.
The performance of IceDetectNet was evaluated in terms of its detection and classification performance, at both the image and aggregate component scale. At the image scale, IceDetectNet showed good detection performance, correctly classifying 92 % of the ice crystals into the aggregate and non-aggregate classes. In terms of classification performance, it achieved an overall accuracy of 86 % for basic habits and 81 % for microphysical processes. Moreover, IceDetectNet achieved comparable classification accuracies to traditional deep learning algorithms on the non-aggregate subset, whereas the classification accuracies were lower for the aggregate subset. At the component scale, IceDetectNet showed good detection and classification performance across all bounding box sizes, indicating its ability to accurately classify components of aggregated ice crystals down to 662 px.
The generalization ability of IceDetectNet was examined on an independent generalization dataset that was collected during a different season. IceDetectNet showed good detection performance with an overall accuracy of 84 %. Although the classification accuracy decreased compared with the training dataset, the overall accuracy remained satisfactory for basic habit (81 %) and microphysical process (72 %) classification. The aggregate subset showed lower performance compared with the non-aggregate subset, possibly due to imbalances in the dataset. This highlights the potential to further optimize the generalization ability of IceDetectNet through dataset-balancing techniques, enlargement of the training dataset, or fine-tuning.
However, the ice categories used in this study are specific to the environmental and microphysical conditions present during the collection of the training data. In addition, the distinction between small and irregular ice categories combines both size and shape information, making the distinction difficult to classify. While these categories are appropriate for the current dataset, they may pose challenges when applying IceDetectNet to other datasets or comparing results with existing studies. However, adding or refining categories can be easily achieved through model fine-tuning.
IceDetectNet provides detailed shape information on the basic habit and microphysical processes of ice crystals down to the aggregate component scale; thus, it has the potential to improve the estimates of microphysical properties, such as the riming rate, aggregation rate, and ice water content. Due to the good generalization ability of IceDetectNet, we expect that it can also be applied to other cloud imaging probes in connection with fine-tuning. This will help to better understand the radiative properties of clouds and the microphysical processes leading to precipitation formation.
An essential component of training the rotated object detection algorithm is the hand-labeling of bounding boxes and ice crystals, which was done through the hand-labeling platform created by AngoAI (2022). The platform offers a graphical user interface to draw bounding boxes, adjust their size and rotation, and assign labels (Fig. A1).
The classification of ice crystals into their respective basic habits and microphysical processes is a challenging task that requires a set of rules to ensure consistency and accuracy across the dataset. Here, we describe the criteria used for the multi-label classification of ice crystals. We randomly select several images from each category as examples (see Fig. B1) and present the process of how we hand-label an ice crystal (see Fig. B2).
The classification process begins by using human judgment to determine whether the ice particle is an aggregate that contains more than one component. If an ice crystal is not aggregated, the classification process proceeds directly with the classification of the basic habit. For aggregated crystals, the process differs between the training and evaluation of IceDetectNet. In training, each component is manually located with a bounding box (i.e., smallest rectangle box) around the component, and these boxes are then classified. In the multi-label classification, only the largest visually identified component of the aggregate is classified, without drawing a bounding box. The classified basic habit of this largest component will represent the basic habit of the whole aggregate ice crystal.
The next step is to classify the basic habits of the ice crystals/components. If the basic habit is not recognizable (as defined in Table 1), the size of the ice/component is assessed by eye. Small crystals are classified as small, whereas all others are classified as irregular–aged. If the basic habit is recognizable, we classify based on shape. Special shapes, like lollipop–aged for lollipop-like crystals or frozen droplets for those with droplet features, are classified first. Rectangular-shaped ice crystals (with four distinct edges) are classified as columns, whereas rectangular-shaped ice crystals with multiple branches at the end of the maximum dimension are labeled CPC–aged. Note that the CPC–aged categories also include needle bundles with missing plate sections. Hexagonal crystals (with six distinct edges) are classified based on their aspect ratio: a high aspect ratio indicates a column, whereas a low aspect ratio indicates a plate. Crystals/components that do not fit these categories are considered irregular–aged.
Once the basic habit is determined, the appearance of the edges of the ice/component determines whether the ice/component is aged or not. As mentioned earlier (see Sect. 2), irregular, CPC, and lollipop–aged are aged by default, whereas small ice crystals are pristine by definition. Therefore, we only need to decide if column, plate and frozen drops are aged or not. Usually, when an ice/component is aged, it has some tiny bumps on the edges.
The data (https://doi.org/10.5281/zenodo.14288739, Zhang et al., 2024b), code (https://doi.org/10.5281/zenodo.14288779, Zhang et al., 2024a), and models (https://doi.org/10.5281/zenodo.14288804, Zhang et al., 2024c) are available via the links provided in the references.
HZ and XL collaborated closely to design and develop the algorithm architecture; they also conducted all of the model runs. FR and JH actively participated in discussions and gave suggestions regarding the development of the algorithm. JP collected the training and test data with assistance from ROD and JH and conducted the single-label hand-labeling at the image scale. Both HZ and XL labeled the bounding boxes for the training and test data. Subsequently, HZ relabeled the training and test data based on the proposed multi-label classification scheme. The manuscript was written by HZ, with valuable input and discussions from FR, JH, ROD, and XL. The initial idea for the classification scheme was conceived through discussions between ROD and HZ.
The contact author has declared that none of the authors has any competing interests.
Publisher’s note: Copernicus Publications remains neutral with regard to jurisdictional claims made in the text, published maps, institutional affiliations, or any other geographical representation in this paper. While Copernicus Publications makes every effort to include appropriate place names, the final responsibility lies with the authors.
The authors extend their heartfelt gratitude to Trude Storelvmo and Alexander Binder for their invaluable guidance during Huiying Zhang's master's study, which is the foundation of this work. We also acknowledge financial support from the European Research Council (ERC) within the framework of the European Union's Horizon 2020 Research and Innovation program. Finally, Robert O. David would like to acknowledge EEARO-NO-2019-0423/IceSafari, the EEA Grants and Norway Grants, and the European Union's Horizon WIDERA 2021 program for financial support.
This research has been supported by the European Research Council's Horizon 2020 Research and Innovation program (grant no. 101021272), the EEA Grants and Norway Grants (grant no. 31/2020), and the European Commission's Horizon Europe Framework program (grant no. 101079385).
This paper was edited by Linlu Mei and reviewed by three anonymous referees.
Albawi, S., Mohammed, T. A., and Al-Zawi, S.: Understanding of a convolutional neural network, in: 2017 International Conference on Engineering and Technology (ICET), Antalya, Turkey, 21–23 August 2017, IEEE, 1–6, https://doi.org/10.1109/ICEngTechnol.2017.8308186, 2017. a, b
AngoAI: Projects, https://hub.ango.ai/ (last access: 1 September 2022), 2022. a, b, c
Arlot, S. and Celisse, A.: A survey of cross-validation procedures for model selection, Statistics Surveys, 4, 40–79, https://doi.org/10.1214/09-SS054, 2010. a, b
Bailey, M. and Hallett, J.: Growth rates and habits of ice crystals between- 20 and −70 °C, J. Atmos. Sci., 61, 514–544, 2004. a, b
Bishop, C. M. and Nasrabadi, N. M.: Pattern recognition and machine learning, Springer, 4, 2 pp., ISBN 978-1-4939-3843-8, 2016. a
Browne, M. W.: Cross-Validation Methods, J. Math. Psychol., 44, 108–132, https://doi.org/10.1006/jmps.1999.1279, 2000. a
Cunningham, R. M.: Analysis of particle spectral data from optical array (PMS) 1D and 2D sensors, in: Proceedings of the Symposium on Meteorological Observations and Instrumentation (4th), 10–14 April 1978, Denver, CO, 1978. a
Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-Fei, L.: ImageNet: A large-scale hierarchical image database, in: 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009, IEEE, 248–255, https://doi.org/10.1109/CVPR.2009.5206848, 2009. a
Ding, J., Xue, N., Long, Y., Xia, G.-S., and Lu, Q.: Learning RoI transformer for oriented object detection in aerial images, in: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019, IEEE, 2849–2858, https://doi.org/10.1109/CVPR.2019.00296, 2019. a
Duroure, C., Larsen, H., Isaka, H., and Personne, P.: 2D image population analysis, Atmos. Res., 34, 195–205, https://doi.org/10.1016/0169-8095(94)90091-4, 1994. a
Ehrlich, A., Wendisch, M., Bierwirth, E., Herber, A., and Schwarzenböck, A.: Ice crystal shape effects on solar radiative properties of Arctic mixed-phase clouds – Dependence on microphysical properties, Atmos. Res., 88, 266–276, https://doi.org/10.1016/j.atmosres.2007.11.018, 2008. a
Findeisen, W.: Kolloid meteorologische Vorgänge bei Neiderschlagsbildung, Meteor. Z, 55, 121–133, 1938. a
Flanner, M. G., Zender, C. S., Randerson, J. T., and Rasch, P. J.: Present-day climate forcing and response from black carbon in snow, J. Geophys. Res.-Atmos., 112, D11202, https://doi.org/10.1029/2006JD008003, 2007. a, b
Goodfellow, I., Bengio, Y., and Courville, A.: Deep learning, MIT press, http://www.deeplearningbook.org (last access: 6 December 2024), 2016. a, b, c, d, e, f
Gu, J., Wang, Z., Kuen, J., Ma, L., Shahroudy, A., Shuai, B., Liu, T., Wang, X., Wang, G., Cai, J., and Chen, T.: Recent advances in convolutional neural networks, Pattern Recogn., 77, 354–377, 2018. a
Han, J., Ding, J., Li, J., and Xia, G.-S.: Align deep features for oriented object detection, IEEE T. Geosci. Remote, 60, 1–11, 2021. a
He, K., Zhang, X., Ren, S., and Sun, J.: Deep Residual Learning for Image Recognition, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), https://openaccess.thecvf.com/content_cvpr_2016/html/He_Deep_Residual_Learning_CVPR_2016_paper.html (last access: 6 December 2024), 2016. a, b, c, d
Heymsfield, A.: Ice crystal terminal velocities, J. Atmos. Sci., 29, 1348–1357, 1972. a
Jabbar, H. and Khan, R. Z.: Methods to avoid over-fitting and under-fitting in supervised machine learning (comparative study), Computer Science, Communication and Instrumentation Devices, 70, 978–981, 2015. a
Jaffeux, L., Schwarzenböck, A., Coutris, P., and Duroure, C.: Ice crystal images from optical array probes: classification with convolutional neural networks, Atmos. Meas. Tech., 15, 5141–5157, https://doi.org/10.5194/amt-15-5141-2022, 2022. a, b, c
Järvinen, E., Jourdan, O., Neubauer, D., Yao, B., Liu, C., Andreae, M. O., Lohmann, U., Wendisch, M., McFarquhar, G. M., Leisner, T., and Schnaiter, M.: Additional global climate cooling by clouds due to ice crystal complexity, Atmos. Chem. Phys., 18, 15767–15781, https://doi.org/10.5194/acp-18-15767-2018, 2018. a, b
Jensen, A. A., Harrington, J. Y., Morrison, H., and Milbrandt, J. A.: Predicting ice shape evolution in a bulk microphysics model, J. Atmos. Sci., 74, 2081–2104, 2017. a
Kalina, M. and Puxbaum, H.: A study of the influence of riming of ice crystals on snow chemistry during different seasons in precipitating continental clouds, Atmos. Environ., 28, 3311–3328, 1994. a
Keppas, S. C., Crosier, J., Choularton, T., and Bower, K.: Ice lollies: An ice particle generated in supercooled conveyor belts, Geophys. Res. Lett., 44, 5222–5230, 2017. a
Khvorostyanov, V. I. and Curry, J. A.: Terminal velocities of droplets and crystals: Power laws with continuous parameters over the size spectrum, J. Atmos. Sci., 59, 1872–1884, 2002. a
Kikuchi, K., Kameda, T., Higuchi, K., and Yamashita, A.: A global classification of snow crystals, ice crystals, and solid precipitation based on observations from middle latitudes to polar regions, Atmos. Res., 132, 460–472, 2013. a
Korolev, A., Isaac, G., and Hallett, J.: Ice particle habits in Arctic clouds, Geophys. Res. Lett., 26, 1299–1302, 1999. a
Korolev, A., Isaac, G., and Hallett, J.: Ice particle habits in stratiform clouds, Q. J. Roy. Meteor. Soc., 126, 2873–2902, 2000. a
Le-Khac, P. H., Healy, G., and Smeaton, A. F.: Contrastive representation learning: A framework and review, IEEE Access, 8, 193907–193934, 2020. a
Li, Z., Liu, F., Yang, W., Peng, S., and Zhou, J.: A survey of convolutional neural networks: analysis, applications, and prospects, IEEE T. Neur. Net. Lear., 33, 6999–7019, https://doi.org/10.1109/TNNLS.2021.3084827, 2021. a
Libbrecht, K.: Ken Libbrecht's Field Guide to Snowflakes, Voyageur Press, ISBN 9781616739478, 2016. a
Lindqvist, H., Muinonen, K., Nousiainen, T., Um, J., McFarquhar, G. M., Haapanala, P., Makkonen, R., and Hakkarainen, H.: Ice-cloud particle habit classification using principal components, J. Geophys. Res.-Atmos., 117, D16206, https://doi.org/10.1029/2012JD017573, 2012. a
Lohmann, U., Lüönd, F., and Mahrt, F.: An introduction to clouds: From the microscale to climate, Cambridge University Press, https://doi.org/10.1017/CBO9781139087513, 2016. a
Matus, A. V. and L'Ecuyer, T. S.: The role of cloud phase in Earth's radiation budget, J. Geophys. Res.-Atmos., 122, 2559–2578, 2017. a
Mitchell, D. L.: Use of mass-and area-dimensional power laws for determining precipitation particle terminal velocities, J. Atmos. Sci., 53, 1710–1723, 1996. a
Mitchell, D. L., Zhang, R., and Pitter, R. L.: Mass-dimensional relationships for ice particles and the influence of riming on snowfall rates, J. Appl. Meteorol. Clim., 29, 153–163, 1990. a
Mosimann, L., Steiner, M., and Henrich, W.: Prediction of snow crystal shape and riming by vertical Doppler radar, Atmos. Res., 29, 85–98, 1993. a
Pan, S. J. and Yang, Q.: A survey on transfer learning, IEEE T. Knowl. Data En., 22, 1345–1359, 2009. a
Pasquier, J. T., David, R. O., Freitas, G., Gierens, R., Gramlich, Y., Haslett, S., Li, G., Schäfer, B., Siegel, K., Wieder, J., Adachi, K., Belosi, F., Carlsen, T., Decesari, S., Ebell, K., Gilardoni, S., Gysel-Beer, M., Henneberger, J., Inoue, J., Kanji, Z. A., Koike, M., Kondo, Y., Krejci, R., Lohmann, U., Maturilli, M., Mazzolla, M., Modini, R., Mohr, C., Motos, G., Nenes, A., Nicosia, A., Ohata, S., Paglione, M., Park, S., Pileci, R. E., Ramelli, F., Rinaldi, M., Ritter, C., Sato, K., Storelvmo, T., Tobo, Y., Traversi, R., Viola, A., and Zieger, P.: The Ny-Ålesund aerosol cloud experiment (NASCENT): Overview and first results, B. Am. Meteorol. Soc., 103, E2533–E2558, 2022a. a, b, c
Pasquier, J. T., Henneberger, J., Ramelli, F., Lauber, A., David, R. O., Wieder, J., Carlsen, T., Gierens, R., Maturilli, M., and Lohmann, U.: Conditions favorable for secondary ice production in Arctic mixed-phase clouds, Atmos. Chem. Phys., 22, 15579–15601, https://doi.org/10.5194/acp-22-15579-2022, 2022b. a, b, c
Pasquier, J. T., Henneberger, J., Korolev, A., Ramelli, F., Wieder, J., Lauber, A., Li, G., David, R. O., Carlsen, T., Gierens, R., Maturilli, M., and Lohmann, U.: Understanding the history of two complex ice crystal habits deduced from a holographic imager, Geophys. Res. Lett., 50, e2022GL100247, https://doi.org/10.1029/2022GL100247, 2023. a, b
Praz, C., Roulet, Y.-A., and Berne, A.: Solid hydrometeor classification and riming degree estimation from pictures collected with a Multi-Angle Snowflake Camera, Atmos. Meas. Tech., 10, 1335–1357, https://doi.org/10.5194/amt-10-1335-2017, 2017. a
Rahman, M. M., Quincy, E. A., Jacquot, R. G., and Magee, M. J.: Feature Extraction and Selection for Pattern Recognition of Two-Dimensional Hydrometeor Images, J. Appl. Meteorol. Clim., 20, 521–535, https://doi.org/10.1175/1520-0450(1981)020<0521:FEASFP>2.0.CO;2, 1981. a
Ramelli, F., Beck, A., Henneberger, J., and Lohmann, U.: Using a holographic imager on a tethered balloon system for microphysical observations of boundary layer clouds, Atmos. Meas. Tech., 13, 925–939, https://doi.org/10.5194/amt-13-925-2020, 2020. a
Rawat, W. and Wang, Z.: Deep Convolutional Neural Networks for Image Classification: A Comprehensive Review, Neural Comput., 29, 2352–2449, https://doi.org/10.1162/neco_a_00990, 2017. a
Schmitt, C. G. and Heymsfield, A. J.: Observational quantification of the separation of simple and complex atmospheric ice particles, Geophys. Res. Lett., 41, 1301–1307, 2014. a
Stacke, K., Eilertsen, G., Unger, J., and Lundström, C.: Measuring domain shift for deep learning in histopathology, IEEE J. Biomed. Health, 25, 325–336, 2020. a
Sterzinger, L. J. and Igel, A. L.: The Effects of Ice Habit on Simulated Orographic Snowfall, J. Hydrometeorol., 22, 1649–1661, 2021. a
Sun, Z. and Shine, K. P.: Studies of the radiative properties of ice and mixed-phase clouds, Q. J. Roy. Meteor. Soc., 120, 111–137, https://doi.org/10.1002/qj.49712051508, 1994. a
Tajbakhsh, N., Shin, J. Y., Gurudu, S. R., Hurst, R. T., Kendall, C. B., Gotway, M. B., and Liang, J.: Convolutional Neural Networks for Medical Image Analysis: Full Training or Fine Tuning?, IEEE T. Med. Imaging, 35, 1299–1312, https://doi.org/10.1109/TMI.2016.2535302, 2016. a
Touloupas, G., Lauber, A., Henneberger, J., Beck, A., and Lucchi, A.: A convolutional neural network for classifying cloud particles recorded by imaging probes, Atmos. Meas. Tech., 13, 2219–2239, https://doi.org/10.5194/amt-13-2219-2020, 2020. a, b, c, d
Wang, P. K. and Ji, W.: Collision efficiencies of ice crystals at low–intermediate Reynolds numbers colliding with supercooled cloud droplets: A numerical study, J. Atmos. Sci., 57, 1001–1009, 2000. a
Wegener, A.: Thermodynamik der Atmosphäre, JA Barth, https://books.google.ch/books?id=BMMxAAAAMAAJ (last access: 6 December 2024), 1911. a
Woods, C. P., Stoelinga, M. T., and Locatelli, J. D.: The IMPROVE-1 storm of 1–2 February 2001. Part III: Sensitivity of a mesoscale model simulation to the representation of snow particle types and testing of a bulk microphysical scheme with snow habit prediction, J. Atmos. Sci., 64, 3927–3948, 2007. a
Xiao, H., Zhang, F., He, Q., Liu, P., Yan, F., Miao, L., and Yang, Z.: Classification of Ice Crystal Habits Observed From Airborne Cloud Particle Imager by Deep Transfer Learning, Earth Space Sci., 6, 1877–1886, https://doi.org/10.1029/2019EA000636, 2019. a, b, c
Yang, P., Liou, K.-N., Bi, L., Liu, C., Yi, B., and Baum, B. A.: On the radiative properties of ice clouds: Light scattering, remote sensing, and radiation parameterization, Adv. Atmos. Sci., 32, 32–63, 2015. a
Zhang, H.: Development of a classification algorithm for ice crystal habit by using deep learning, Master's thesis, http://urn.nb.no/URN:NBN:no-91875 (last access: 5 November 2024), 2021. a, b, c
Zhang, H., Chen, Y., Song, Y., Xiong, Z., Yang, Y., and Wu, Q. J.: Automatic kidney lesion detection for CT images using morphological cascade convolutional neural networks, IEEE Access, 7, 83001–83011, 2019. a, b
Zhang, H., Binder, A., Pasquier, J., Krummenacher, B., Ramelli, F., Storelvmo, T., David, R. O., and Henneberger, J.: Deep-learning based classification of ice crystals: habits and microphysical processes, EGU General Assembly 2022, Vienna, Austria, 23–27 May 2022, EGU22-8098, https://doi.org/10.5194/egusphere-egu22-8098, 2022. a
Zhang, H., Li, X., Ramelli, F., David, R., Villinger, J., and Henneberger, J.: Code for publication “IceDetectNet: a rotated object detection algorithm for classifying components of aggregated ice crystals with a multi-label classification scheme”, Zenodo [code], https://doi.org/10.5281/zenodo.14288779, 2024a. a
Zhang, H., Li, X., Ramelli, F., David, R., Villinger, J., and Henneberger, J.: Data for publication “IceDetectNet: a rotated object detection algorithm for classifying components of aggregated ice crystals with a multi-label classification scheme”, Zenodo [data set], https://doi.org/10.5281/zenodo.14288739, 2024b. a
Zhang, H., Li, X., Ramelli, F., David, R., Villinger, J., and Henneberger, J.: Models for publication “IceDetectNet: a rotated object detection algorithm for classifying components of aggregated ice crystals with a multi-label classification scheme”, Zenodo [code], https://doi.org/10.5281/zenodo.14288804, 2024c. a
Zhao, Z.-Q., Zheng, P., Xu, S.-T., and Wu, X.: Object detection with deep learning: A review, IEEE T. Neur. Net. Lear., 30, 3212–3232, 2019. a
Zou, Z., Chen, K., Shi, Z., Guo, Y., and Ye, J.: Object detection in 20 years: A survey, arXiv [preprint], https://doi.org/10.48550/arXiv.1905.05055, 18 January 2023. a
- Abstract
- Introduction
- Data description
- Methodology
- Results
- Discussion
- Conclusion
- Appendix A: Hand-labeling platform
- Appendix B: Detailed criteria for ice crystal classification
- Code and data availability
- Author contributions
- Competing interests
- Disclaimer
- Acknowledgements
- Financial support
- Review statement
- References
- Abstract
- Introduction
- Data description
- Methodology
- Results
- Discussion
- Conclusion
- Appendix A: Hand-labeling platform
- Appendix B: Detailed criteria for ice crystal classification
- Code and data availability
- Author contributions
- Competing interests
- Disclaimer
- Acknowledgements
- Financial support
- Review statement
- References