<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing with OASIS Tables v3.0 20080202//EN" "journalpub-oasis3.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:oasis="http://docs.oasis-open.org/ns/oasis-exchange/table" xml:lang="en" dtd-version="3.0" article-type="research-article"><?xmltex \makeatother\@nolinetrue\makeatletter?>
  <front>
    <journal-meta><journal-id journal-id-type="publisher">AMT</journal-id><journal-title-group>
    <journal-title>Atmospheric Measurement Techniques</journal-title>
    <abbrev-journal-title abbrev-type="publisher">AMT</abbrev-journal-title><abbrev-journal-title abbrev-type="nlm-ta">Atmos. Meas. Tech.</abbrev-journal-title>
  </journal-title-group><issn pub-type="epub">1867-8548</issn><publisher>
    <publisher-name>Copernicus Publications</publisher-name>
    <publisher-loc>Göttingen, Germany</publisher-loc>
  </publisher></journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.5194/amt-14-6695-2021</article-id><title-group><article-title>Twenty-four-hour cloud cover calculation using <?xmltex \hack{\break}?> a ground-based imager with machine learning</article-title><alt-title>24 h cloud cover calculation using a ground-based imager with machine learning</alt-title>
      </title-group><?xmltex \runningtitle{24\,h cloud cover calculation using a ground-based imager with machine learning}?><?xmltex \runningauthor{B.-Y.~Kim et al.}?>
      <contrib-group>
        <contrib contrib-type="author" corresp="yes">
          <name><surname>Kim</surname><given-names>Bu-Yo</given-names></name>
          <email>kimbuyo@korea.kr</email>
        </contrib>
        <contrib contrib-type="author" corresp="no">
          <name><surname>Cha</surname><given-names>Joo Wan</given-names></name>
          
        </contrib>
        <contrib contrib-type="author" corresp="no">
          <name><surname>Chang</surname><given-names>Ki-Ho</given-names></name>
          
        </contrib>
        <aff id="aff1"><institution>Convergence Meteorological Research Department, National Institute of
Meteorological Sciences, <?xmltex \hack{\break}?> Seogwipo, Jeju 63568, Republic of Korea</institution>
        </aff>
      </contrib-group>
      <author-notes><corresp id="corr1">Bu-Yo Kim (kimbuyo@korea.kr)</corresp></author-notes><pub-date><day>18</day><month>October</month><year>2021</year></pub-date>
      
      <volume>14</volume>
      <issue>10</issue>
      <fpage>6695</fpage><lpage>6710</lpage>
      <history>
        <date date-type="received"><day>22</day><month>June</month><year>2021</year></date>
           <date date-type="rev-request"><day>9</day><month>July</month><year>2021</year></date>
           <date date-type="rev-recd"><day>2</day><month>September</month><year>2021</year></date>
           <date date-type="accepted"><day>20</day><month>September</month><year>2021</year></date>
      </history>
      <permissions>
        <copyright-statement>Copyright: © 2021 Bu-Yo Kim et al.</copyright-statement>
        <copyright-year>2021</copyright-year>
      <license license-type="open-access"><license-p>This work is licensed under the Creative Commons Attribution 4.0 International License. To view a copy of this licence, visit <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link></license-p></license></permissions><self-uri xlink:href="https://amt.copernicus.org/articles/14/6695/2021/amt-14-6695-2021.html">This article is available from https://amt.copernicus.org/articles/14/6695/2021/amt-14-6695-2021.html</self-uri><self-uri xlink:href="https://amt.copernicus.org/articles/14/6695/2021/amt-14-6695-2021.pdf">The full text article is available as a PDF file from https://amt.copernicus.org/articles/14/6695/2021/amt-14-6695-2021.pdf</self-uri>
      <abstract><title>Abstract</title>
    <p id="d1e100">In this study, image data features and machine learning methods were used to calculate 24 h continuous cloud cover from image data obtained by a camera-based imager on the ground. The image data features were the time (Julian day and hour), solar zenith angle, and statistical characteristics of the red–blue ratio, blue–red difference, and luminance. These features were determined from the red, green, and blue brightness of images subjected to a pre-processing process involving masking removal and distortion correction. The collected image data were divided into training,
validation, and test sets and were used to optimize and evaluate the
accuracy of each machine learning method. The cloud cover calculated by each machine learning method was verified with human-eye observation data from a manned observatory. Supervised machine learning models suitable for
nowcasting, namely, support vector regression, random forest, gradient
boosting machine, <inline-formula><mml:math id="M1" display="inline"><mml:mi>k</mml:mi></mml:math></inline-formula>-nearest neighbor, artificial neural network, and multiple linear regression methods, were employed and their results were compared. The best learning results were obtained by the support vector regression model, which had an accuracy, recall, and precision of 0.94, 0.70, and 0.76, respectively. Further, bias, root mean square error, and correlation coefficient values of 0.04 tenths, 1.45 tenths, and 0.93, respectively, were obtained for the cloud cover calculated using the test set. When the difference between the calculated and observed cloud cover was allowed to range between 0, 1, and 2 tenths, high agreements of approximately 42 %, 79 %, and 91 %, respectively, were obtained. The proposed system involving a ground-based imager and machine learning methods is expected to be suitable for application as an automated system to replace human-eye observations.</p>
  </abstract>
    </article-meta>
  </front>
<body>
      

      <?xmltex \hack{\newpage}?>
<sec id="Ch1.S1" sec-type="intro">
  <label>1</label><title>Introduction</title>
      <p id="d1e121">In countries, including South Korea, that have not introduced automated
systems, ground-based cloud cover observation has been performed using the
human eye, in accordance with the normalized synoptic observation rule of
the World Meteorological Organization (WMO), and recorded in tenths or oktas
(Kim et al., 2016; Yun and Whang, 2018). However, human-eye observation of
cloud cover lacks consistency and depends on the observer conditions and the
observation term (Mantelli Neto et al., 2010; Yang et al., 2016). Further,
although continuous cloud cover observation during both day and night is
important, there is a lack of data continuity (observations with at least
1 h intervals) because a person must perform direct observations (Kim and Cha, 2020). In addition, construction of a dense cloud observation network
from observation environments with low accessibility, such as mountaintops,
is difficult. Therefore, meteorological satellites and ground-based remote
observation equipment that can continuously monitor clouds while overcoming
these problems are now being employed (Yabuki et al., 2014; Yang et al.,
2015; Kim et al., 2016; Kim and Cha, 2020).</p>
      <p id="d1e124">Geostationary satellites can observe clouds on the global scale at intervals
of several minutes; however, their spatial resolution is as large as several
kilometers (Kim et al., 2018; Lee et al., 2018). Polar satellites have spatial resolutions of several hundred meters, i.e., high resolution;
however, they can observe the same area only once or twice per day (Kim and Lee, 2019; Kim et al., 2020a). For both geostationary and polar satellites, geometric distortion problems occur during cloud cover estimation on the ground, depending on the cloud height (Mantelli Neto et al., 2010). As cloud heights and thicknesses vary, the cloud detection uncertainty also<?pagebreak page6696?> varies depending on the position of the sun or satellite (Ghonima et al., 2012). In general, cloud cover estimation using satellite data differs from the approach used for human-eye observation data, because the wide grid data around the central grid are averaged or calculated as fractions (Alonso-Montesinos, 2020; Sunil et al., 2021).</p>
      <p id="d1e127">Radar, lidar, ceilometers, and camera-based imagers can be used as
ground-based observation instruments (Boers et al., 2010). With regard to
radar, cloud radar technology such as Ka-band radar is suitable for cloud
detection but has the disadvantage of reduced detection accuracy with
increased distance from the radar apparatus (Kim et al., 2020b; Yoshida et
al., 2021). For lidar and ceilometers, the uncertainty is very large because
the cloud cover is calculated from the signal intensity of a narrow portion
of the sky (Costa-Surós et al., 2014; Peng et al., 2015; Kim and Cha,
2020). In contrast, for a camera-based imager, the sky in the surrounding
hemisphere can be observed through a fisheye lens (180<inline-formula><mml:math id="M2" display="inline"><mml:msup><mml:mi/><mml:mo>∘</mml:mo></mml:msup></mml:math></inline-formula> field of
view – FOV) mounted on the camera. Further, depending on the performance of
the imager and the operation method, clouds can be observed continuously for
24 h, i.e., through the day and night. The data can be stored as images, and
the cloud cover can be calculated from these data (Kim and Cha, 2020; Sunil
et al., 2021).</p>
      <p id="d1e139">Many studies have attempted to use camera-based imagers for automatic cloud
observation and cloud cover calculation on the ground (Dev et al., 2016;
Lothon et al., 2019; Shields et al., 2019). Those results can be used for
numerical weather analysis and forecasting; they are also very economical
and ideal for cloud monitoring over local areas (Mantelli Neto et al., 2010;
Kazantzidis et al., 2012; Ye et al., 2017; Valentín et al., 2019). In
general, cloud cover can be calculated based on the brightness of the red,
green, and blue (RGB) colors of the image taken by the imager. In detail,
the RGB brightness varies according to the light scattering from the sky and
clouds, and, using the ratio or difference between these colors, cloud can be
detected and cloud cover can be calculated (Long et al., 2006; Shields et
al., 2013; Liu et al., 2014; Yang et al., 2015; Kim et al., 2016). For example, when the red–blue ratio (RBR) is 0.6 or more or the red–blue difference (RBD) is less than 30, the corresponding pixel is classified (i.e., using a threshold method) as a cloud pixel and incorporated in the cloud cover calculation (Kreuter et al., 2009; Heinle et al., 2010; Liu et al., 2014; Azhar et al., 2021). However, using these empirical methods, it is difficult to distinguish between the sky and clouds under various weather conditions (Yang et al., 2015). This is because the colors of the sky and clouds vary with the atmospheric conditions and because the sun position and threshold conditions can change continuously (Yabuki et al., 2014; Blazek and Pata, 2015; Cazorla et al., 2015; Calbó et al., 2017). Therefore, methods of cloud detection and cloud cover calculation involving application of machine learning methods to images are now being implemented, as an alternative to empirical methods (Peng et al., 2015; Lothon et al., 2019; Al-lahham et al., 2020; Shi et al., 2021).</p>
      <p id="d1e143">Cloud cover can be calculated from camera-based imager data using a supervised machine learning method capable of regression analysis (Al-lahham
et al., 2020). Supervised learning is a method through which a prediction
model is constructed using training data which already contain the labeled
data. Examples include support vector machines (SVMs), decision trees (DTs),
gradient boosting machines (GBMs), and artificial neural networks (ANNs)
(Çınar et al., 2020; Shin et al., 2020). Deep learning methods that
repeatedly learn data features by sub-sampling image data at each
convolution step for gradient descent are also available, such as
convolutional neural networks (Dev et al., 2019; Shi et al., 2019; Xie et
al., 2020). However, this approach is difficult to utilize for nowcasting
because considerable physical resources and time are consumed by the
learning and prediction processes (Al Banna et al., 2020; Kim et al., 2021).</p>
      <p id="d1e146">In this study, cloud cover was calculated continuously for 24 h from image
data obtained by a camera-based imager on the ground, using image data
features and machine learning methods. ANN, GBM, <inline-formula><mml:math id="M3" display="inline"><mml:mi>k</mml:mi></mml:math></inline-formula>-nearest neighbor (kNN),
multiple linear regression (MLR), support vector regression (SVR), and
random forest (RF) methods suitable for nowcasting were used for calculation. For each of these methods, an optimal prediction model is constructed by setting hyper-parameters. The machine learning model most suitable for cloud cover calculation is then selected by comparing the prediction performance of each model on training and validation datasets. The cloud cover calculated from the selected machine learning model is then compared with human-eye observation data and the results are analyzed. The remainder of this paper is organized as follows. The image and observation data used in this study are described in Sect. 2, and the machine learning methods and their sets are summarized in Sect. 3. The prediction performance evaluation for each machine learning method and the calculation result verification are reported in Sect. 4. Finally, the summary and conclusion are given in Sect. 5.</p>
</sec>
<sec id="Ch1.S2">
  <label>2</label><title>Research data and methods</title>
<sec id="Ch1.S2.SS1">
  <label>2.1</label><title>Ground-based imager</title>
      <p id="d1e171">In this study, a digital camera-based automatic cloud observation system (ACOS) was developed using a Canon EOS 6D camera to detect and calculate cloud cover for 24 h, as shown in Fig. 1. This system was developed by the National Institute of Meteorological Sciences (NIMS)/Korea Meteorological Administration (KMA) and A &amp; D <inline-formula><mml:math id="M4" display="inline"><mml:mo>⋅</mml:mo></mml:math></inline-formula> 3D Co., Ltd. (Kim and Cha, 2020). The ACOS was installed at the Daejeon Regional Office of Meteorology (DROM; 36.37<inline-formula><mml:math id="M5" display="inline"><mml:msup><mml:mi/><mml:mo>∘</mml:mo></mml:msup></mml:math></inline-formula> N, 127.37<inline-formula><mml:math id="M6" display="inline"><mml:msup><mml:mi/><mml:mo>∘</mml:mo></mml:msup></mml:math></inline-formula> E), a manned observatory in which cloud cover observation by human eye is performed. The detailed ACOS<?pagebreak page6697?> specifications are listed in Table 1. The International Organization for Standardization (ISO) values of the complementary metal oxide semiconductor (CMOS) sensor employed in the camera are 100 (day)–25 600 (night), and the sensitivity is adjusted according to the image brightness. In this study, the camera shutter speed was set to <inline-formula><mml:math id="M7" display="inline"><mml:mrow><mml:mn mathvariant="normal">1</mml:mn><mml:mo>/</mml:mo><mml:mn mathvariant="normal">1000</mml:mn></mml:mrow></mml:math></inline-formula> s (day)–5 s (night), considering the long exposure for object detection required at night. The F-stop value was set to F8 (day)–F11 (night), and the sky-dome object was taken with a large depth of field (Peng et al., 2015; Dev et al., 2017). The camera lens was installed at a height of 1.8 m, similar to human-eye height, and a fisheye lens (EF 8–15 mm F/4L fisheye USM) was installed to capture the entire surroundings, including the sky and clouds, within a 180<inline-formula><mml:math id="M8" display="inline"><mml:msup><mml:mi/><mml:mo>∘</mml:mo></mml:msup></mml:math></inline-formula> FOV. To perform 24 h continuous observation, heating (below <inline-formula><mml:math id="M9" display="inline"><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">2</mml:mn></mml:mrow></mml:math></inline-formula> <inline-formula><mml:math id="M10" display="inline"><mml:msup><mml:mi/><mml:mo>∘</mml:mo></mml:msup></mml:math></inline-formula>C) and ventilation devices were installed inside the ACOS body to facilitate image acquisition without manned management (Dev et al., 2015; Kim and Cha, 2020).</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F1"><?xmltex \currentcnt{1}?><?xmltex \def\figurename{Figure}?><label>Figure 1</label><caption><p id="d1e242">ACOS appearance <bold>(a)</bold> and installation environment <bold>(b)</bold> (Kim and Cha, 2020).</p></caption>
          <?xmltex \igopts{width=236.157874pt}?><graphic xlink:href="https://amt.copernicus.org/articles/14/6695/2021/amt-14-6695-2021-f01.jpg"/>

        </fig>

<?xmltex \floatpos{t}?><table-wrap id="Ch1.T1"><?xmltex \currentcnt{1}?><label>Table 1</label><caption><p id="d1e260">Detailed ACOS specifications.</p></caption><oasis:table frame="topbot"><?xmltex \begin{scaleboxenv}{.85}[.85]?><oasis:tgroup cols="2">
     <oasis:colspec colnum="1" colname="col1" align="left"/>
     <oasis:colspec colnum="2" colname="col2" align="left"/>
     <oasis:thead>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1">Function</oasis:entry>
         <oasis:entry colname="col2">Description</oasis:entry>
       </oasis:row>
     </oasis:thead>
     <oasis:tbody>
       <oasis:row>
         <oasis:entry colname="col1">Size</oasis:entry>
         <oasis:entry colname="col2">264 mm (<inline-formula><mml:math id="M11" display="inline"><mml:mi>L</mml:mi></mml:math></inline-formula>) <inline-formula><mml:math id="M12" display="inline"><mml:mo>×</mml:mo></mml:math></inline-formula> 264 mm (<inline-formula><mml:math id="M13" display="inline"><mml:mi>W</mml:mi></mml:math></inline-formula>) <inline-formula><mml:math id="M14" display="inline"><mml:mo>×</mml:mo></mml:math></inline-formula> 250 mm (<inline-formula><mml:math id="M15" display="inline"><mml:mi>H</mml:mi></mml:math></inline-formula>), 6.5 kg</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Pixels</oasis:entry>
         <oasis:entry colname="col2"><inline-formula><mml:math id="M16" display="inline"><mml:mrow><mml:mn mathvariant="normal">2432</mml:mn><mml:mo>×</mml:mo><mml:mn mathvariant="normal">2432</mml:mn></mml:mrow></mml:math></inline-formula></oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Focal length</oasis:entry>
         <oasis:entry colname="col2">8 mm, 180<inline-formula><mml:math id="M17" display="inline"><mml:msup><mml:mi/><mml:mo>∘</mml:mo></mml:msup></mml:math></inline-formula> fisheye lens</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Sensor</oasis:entry>
         <oasis:entry colname="col2">CMOS</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Aperture</oasis:entry>
         <oasis:entry colname="col2">F8 (day)–F11 (night)</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Shutter speeds</oasis:entry>
         <oasis:entry colname="col2"><inline-formula><mml:math id="M18" display="inline"><mml:mrow><mml:mn mathvariant="normal">1</mml:mn><mml:mo>/</mml:mo><mml:mn mathvariant="normal">1000</mml:mn></mml:mrow></mml:math></inline-formula> s (day)–5 s (night)</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">ISO</oasis:entry>
         <oasis:entry colname="col2">100 (day)–25 600 (night)</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Observation periods</oasis:entry>
         <oasis:entry colname="col2">24 h operation, 10 min interval observation</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Additional settings</oasis:entry>
         <oasis:entry colname="col2">24 h automatic heating (below <inline-formula><mml:math id="M19" display="inline"><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">2</mml:mn></mml:mrow></mml:math></inline-formula><inline-formula><mml:math id="M20" display="inline"><mml:msup><mml:mi/><mml:mo>∘</mml:mo></mml:msup></mml:math></inline-formula>) and ventilation</oasis:entry>
       </oasis:row>
     </oasis:tbody>
   </oasis:tgroup><?xmltex \end{scaleboxenv}?></oasis:table></table-wrap>

</sec>
<sec id="Ch1.S2.SS2">
  <label>2.2</label><title>Cloud cover calculation and validation</title>
      <p id="d1e460">The image data captured by ACOS were processed by converting each RGB channel of each image pixel to a brightness of 0–255. Although the camera-lens FOV was 180<inline-formula><mml:math id="M21" display="inline"><mml:msup><mml:mi/><mml:mo>∘</mml:mo></mml:msup></mml:math></inline-formula>, only pixel data within the zenith angle of 80<inline-formula><mml:math id="M22" display="inline"><mml:msup><mml:mi/><mml:mo>∘</mml:mo></mml:msup></mml:math></inline-formula> (FOV 160<inline-formula><mml:math id="M23" display="inline"><mml:msup><mml:mi/><mml:mo>∘</mml:mo></mml:msup></mml:math></inline-formula>) were used. This condition was in consideration of the permanent masking area of the horizontal plane due to surrounding objects (buildings, trees, equipment, etc.) (Kazantzidis et al., 2012; Shields et al., 2019; Kim and Cha, 2020). For cloud cover calculation using the ACOS images, image data taken at 1 h intervals from 1 January to 31 December 2019 were used. The cloud cover was calculated using the statistical characteristics of the RGB brightness ratio (i.e., the red–blue ratio – RBR), difference (i.e., the blue–red difference – BRD), and luminance (<inline-formula><mml:math id="M24" display="inline"><mml:mi>Y</mml:mi></mml:math></inline-formula>), which vary for each image (Sect. 2.3), as well as supervised machine learning methods (Sect. 3). Here, <inline-formula><mml:math id="M25" display="inline"><mml:mi>Y</mml:mi></mml:math></inline-formula> was calculated as <inline-formula><mml:math id="M26" display="inline"><mml:mrow><mml:mi>Y</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">0.2126</mml:mn><mml:mi>R</mml:mi><mml:mo>+</mml:mo><mml:mn mathvariant="normal">0.7152</mml:mn><mml:mi>G</mml:mi><mml:mo>+</mml:mo><mml:mn mathvariant="normal">0.0722</mml:mn><mml:mi>B</mml:mi></mml:mrow></mml:math></inline-formula> (Sazzad et al., 2013; Shimoji et al., 2016). The calculated cloud cover was compared with human-eye observation data from DROM. As the cloud cover was calculated as a percentage between 0 % and 100 %, the result was converted to an integer (tenth) between 0 and 10 (Table 2) for comparison with the human-eye-based cloud cover values. As the ACOS was installed at DROM, there were no location differences between observers; thus, the same clouds were captured (Kim and Cha, 2020). At DROM, night observations were performed at 1 h intervals during inclement weather (rainfall, snowfall, etc.), but otherwise at 3 h intervals. The night period varied with the season. Considering this, a total of 7402 images of concurrent human observations were collected, excluding missing cases, from the ACOS.</p>

<?xmltex \floatpos{t}?><table-wrap id="Ch1.T2" specific-use="star"><?xmltex \currentcnt{2}?><label>Table 2</label><caption><p id="d1e534">ACOS cloud cover (%) to DROM human-eye-observed cloud cover (tenths) conversion table.</p></caption><oasis:table frame="topbot"><oasis:tgroup cols="12">
     <oasis:colspec colnum="1" colname="col1" align="left"/>
     <oasis:colspec colnum="2" colname="col2" align="center"/>
     <oasis:colspec colnum="3" colname="col3" align="center"/>
     <oasis:colspec colnum="4" colname="col4" align="center"/>
     <oasis:colspec colnum="5" colname="col5" align="center"/>
     <oasis:colspec colnum="6" colname="col6" align="center"/>
     <oasis:colspec colnum="7" colname="col7" align="center"/>
     <oasis:colspec colnum="8" colname="col8" align="center"/>
     <oasis:colspec colnum="9" colname="col9" align="center"/>
     <oasis:colspec colnum="10" colname="col10" align="center"/>
     <oasis:colspec colnum="11" colname="col11" align="center"/>
     <oasis:colspec colnum="12" colname="col12" align="center"/>
     <oasis:thead>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1">%</oasis:entry>
         <oasis:entry colname="col2"><inline-formula><mml:math id="M27" display="inline"><mml:mrow><mml:mo>≤</mml:mo><mml:mn mathvariant="normal">5</mml:mn></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col3">5–15</oasis:entry>
         <oasis:entry colname="col4">15–25</oasis:entry>
         <oasis:entry colname="col5">25–35</oasis:entry>
         <oasis:entry colname="col6">35–45</oasis:entry>
         <oasis:entry colname="col7">45–55</oasis:entry>
         <oasis:entry colname="col8">55–65</oasis:entry>
         <oasis:entry colname="col9">65–75</oasis:entry>
         <oasis:entry colname="col10">75–85</oasis:entry>
         <oasis:entry colname="col11">85–95</oasis:entry>
         <oasis:entry colname="col12"><inline-formula><mml:math id="M28" display="inline"><mml:mrow><mml:mo>&lt;</mml:mo><mml:mn mathvariant="normal">95</mml:mn></mml:mrow></mml:math></inline-formula></oasis:entry>
       </oasis:row>
     </oasis:thead>
     <oasis:tbody>
       <oasis:row>
         <oasis:entry colname="col1">Tenth</oasis:entry>
         <oasis:entry colname="col2">0</oasis:entry>
         <oasis:entry colname="col3">1</oasis:entry>
         <oasis:entry colname="col4">2</oasis:entry>
         <oasis:entry colname="col5">3</oasis:entry>
         <oasis:entry colname="col6">4</oasis:entry>
         <oasis:entry colname="col7">5</oasis:entry>
         <oasis:entry colname="col8">6</oasis:entry>
         <oasis:entry colname="col9">7</oasis:entry>
         <oasis:entry colname="col10">8</oasis:entry>
         <oasis:entry colname="col11">9</oasis:entry>
         <oasis:entry colname="col12">10</oasis:entry>
       </oasis:row>
     </oasis:tbody>
   </oasis:tgroup></oasis:table></table-wrap>

      <?pagebreak page6698?><p id="d1e672">The entire collected dataset was randomly sampled without replacement. Overall, 50 % (3701 cases) of the total data elements were configured as
a training set, 30 % (2221 cases) as a validation set, and 20 % (1480 cases) as a test set (Xiong et al., 2020). The training set was used to
train the machine learning algorithms, and the prediction performance of each machine learning method was assessed using the validation set. Optimal hyper-parameters were set for each machine learning method through the
training and validation sets. The results of each machine learning method were compared. In this process, the test set was input to the machine learning model that exhibited the best prediction performance, and the calculated results and human-eye observation data were compared. The
accuracy, recall, precision, bias, root mean square error (RMSE), and
correlation coefficient (<inline-formula><mml:math id="M29" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula>) were analyzed according to Eqs. (1)–(6); hence, the prediction performance of each machine learning method was determined and compared based on the human-eye observation data.

                <disp-formula specific-use="align" content-type="numbered"><mml:math id="M30" display="block"><mml:mtable displaystyle="true"><mml:mlabeledtr id="Ch1.E1"><mml:mtd><mml:mtext>1</mml:mtext></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" class="stylechange"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:mi mathvariant="normal">Accuracy</mml:mi><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:mi mathvariant="normal">TP</mml:mi><mml:mo>+</mml:mo><mml:mi mathvariant="normal">TN</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">TP</mml:mi><mml:mo>+</mml:mo><mml:mi mathvariant="normal">TN</mml:mi><mml:mo>+</mml:mo><mml:mi mathvariant="normal">FP</mml:mi><mml:mo>+</mml:mo><mml:mi mathvariant="normal">FN</mml:mi></mml:mrow></mml:mfrac></mml:mstyle></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E2"><mml:mtd><mml:mtext>2</mml:mtext></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" class="stylechange"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:mi mathvariant="normal">Recall</mml:mi><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mi mathvariant="normal">TP</mml:mi><mml:mrow><mml:mi mathvariant="normal">TP</mml:mi><mml:mo>+</mml:mo><mml:mi mathvariant="normal">FN</mml:mi></mml:mrow></mml:mfrac></mml:mstyle></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E3"><mml:mtd><mml:mtext>3</mml:mtext></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" class="stylechange"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:mi mathvariant="normal">Precision</mml:mi><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mi mathvariant="normal">TP</mml:mi><mml:mrow><mml:mi mathvariant="normal">TP</mml:mi><mml:mo>+</mml:mo><mml:mi mathvariant="normal">FP</mml:mi></mml:mrow></mml:mfrac></mml:mstyle></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E4"><mml:mtd><mml:mtext>4</mml:mtext></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" class="stylechange"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle class="stylechange" displaystyle="true"/><mml:mi mathvariant="normal">Bias</mml:mi><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:mo>∑</mml:mo><mml:mo>(</mml:mo><mml:mi>M</mml:mi><mml:mo>-</mml:mo><mml:mi>O</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mi>N</mml:mi></mml:mfrac></mml:mstyle></mml:mrow></mml:mtd></mml:mlabeledtr></mml:mtable></mml:math></disp-formula>

            <?xmltex \hack{\newpage}?><?xmltex \hack{\vspace*{-6mm}}?>

                <disp-formula specific-use="align" content-type="numbered"><mml:math id="M31" display="block"><mml:mtable displaystyle="true"><mml:mlabeledtr id="Ch1.E5"><mml:mtd><mml:mtext>5</mml:mtext></mml:mtd><mml:mtd><mml:mstyle class="stylechange" displaystyle="true"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:mi mathvariant="normal">RMSE</mml:mi><mml:mo>=</mml:mo><mml:msqrt><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:mo>∑</mml:mo><mml:mo>(</mml:mo><mml:mi>M</mml:mi><mml:mo>-</mml:mo><mml:mi>O</mml:mi><mml:msup><mml:mo>)</mml:mo><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow><mml:mi>N</mml:mi></mml:mfrac></mml:mstyle></mml:msqrt></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E6"><mml:mtd><mml:mtext>6</mml:mtext></mml:mtd><mml:mtd><mml:mstyle class="stylechange" displaystyle="true"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle class="stylechange" displaystyle="true"/><mml:mi>R</mml:mi><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:mo>∑</mml:mo><mml:mfenced open="(" close=")"><mml:mrow><mml:mi>M</mml:mi><mml:mo>-</mml:mo><mml:mover accent="true"><mml:mi>M</mml:mi><mml:mo mathvariant="normal">‾</mml:mo></mml:mover></mml:mrow></mml:mfenced><mml:mfenced open="(" close=")"><mml:mrow><mml:mi>O</mml:mi><mml:mo>-</mml:mo><mml:mover accent="true"><mml:mi>O</mml:mi><mml:mo mathvariant="normal">‾</mml:mo></mml:mover></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:mo>∑</mml:mo><mml:msup><mml:mfenced open="(" close=")"><mml:mrow><mml:mi>M</mml:mi><mml:mo>-</mml:mo><mml:mover accent="true"><mml:mi>M</mml:mi><mml:mo mathvariant="normal">‾</mml:mo></mml:mover></mml:mrow></mml:mfenced><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:msqrt><mml:msqrt><mml:mrow><mml:mo>∑</mml:mo><mml:msup><mml:mfenced open="(" close=")"><mml:mrow><mml:mi>O</mml:mi><mml:mo>-</mml:mo><mml:mover accent="true"><mml:mi>O</mml:mi><mml:mo mathvariant="normal">‾</mml:mo></mml:mover></mml:mrow></mml:mfenced><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac></mml:mstyle></mml:mrow></mml:mtd></mml:mlabeledtr></mml:mtable></mml:math></disp-formula>

            Here, TP, TN, FP, and FN are the number of true positives (reference: yes,
prediction: yes), true negatives (reference: yes, prediction: no), false
positives (reference: no, prediction: yes), and false negatives (reference:
no, prediction: no), respectively. Further, <inline-formula><mml:math id="M32" display="inline"><mml:mi>M</mml:mi></mml:math></inline-formula>, <inline-formula><mml:math id="M33" display="inline"><mml:mi>O</mml:mi></mml:math></inline-formula>, and <inline-formula><mml:math id="M34" display="inline"><mml:mi>N</mml:mi></mml:math></inline-formula> are the cloud cover calculated by the employed machine learning method, the human-eye-observed cloud cover, and the number of data, respectively.</p>
</sec>
<sec id="Ch1.S2.SS3">
  <label>2.3</label><title>Machine learning input data</title>
      <p id="d1e953">The data input to the machine learning algorithms for cloud cover calculation using the ACOS images were produced as follows. First, as the ACOS image was taken with a fisheye lens, the image was distorted. That is, objects at the edge were smaller than those at the center of the image (Chauvin et al., 2015; Yang et al., 2015; Lothon et al., 2019). Therefore, the relative size of each object in the image was adjusted through orthogonal projection distortion correction according to the method expressed in Eqs. (7)–(11) (Kim and Cha, 2020).

                <disp-formula specific-use="align" content-type="numbered"><mml:math id="M35" display="block"><mml:mtable displaystyle="true"><mml:mlabeledtr id="Ch1.E7"><mml:mtd><mml:mtext>7</mml:mtext></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" class="stylechange"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle class="stylechange" displaystyle="true"/><mml:mi>r</mml:mi><mml:mo>=</mml:mo><mml:msqrt><mml:mrow><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>-</mml:mo><mml:mi>c</mml:mi><mml:mi>x</mml:mi><mml:msup><mml:mo>)</mml:mo><mml:mn mathvariant="normal">2</mml:mn></mml:msup><mml:mo>+</mml:mo><mml:mo>(</mml:mo><mml:mi>y</mml:mi><mml:mo>-</mml:mo><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:msup><mml:mo>)</mml:mo><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:msqrt><mml:mo>,</mml:mo></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E8"><mml:mtd><mml:mtext>8</mml:mtext></mml:mtd><mml:mtd><mml:mstyle class="stylechange" displaystyle="true"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:mi mathvariant="italic">θ</mml:mi><mml:mo>=</mml:mo><mml:mi mathvariant="normal">asin</mml:mi><mml:mo>(</mml:mo><mml:mi>r</mml:mi><mml:mo>/</mml:mo><mml:mi mathvariant="normal">radi</mml:mi><mml:mo>)</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E9"><mml:mtd><mml:mtext>9</mml:mtext></mml:mtd><mml:mtd><mml:mstyle class="stylechange" displaystyle="true"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:mi mathvariant="italic">ϕ</mml:mi><mml:mo>=</mml:mo><mml:mi mathvariant="normal">asin</mml:mi><mml:mo>(</mml:mo><mml:mo>(</mml:mo><mml:mi>y</mml:mi><mml:mo>-</mml:mo><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mo>)</mml:mo><mml:mo>/</mml:mo><mml:mi>r</mml:mi><mml:mo>)</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E10"><mml:mtd><mml:mtext>10</mml:mtext></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" class="stylechange"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle class="stylechange" displaystyle="true"/><mml:msup><mml:mi>x</mml:mi><mml:mo>′</mml:mo></mml:msup><mml:mo>=</mml:mo><mml:mi>c</mml:mi><mml:mi>x</mml:mi><mml:mo>+</mml:mo><mml:mi>r</mml:mi><mml:mo>×</mml:mo><mml:mi mathvariant="italic">θ</mml:mi><mml:mo>×</mml:mo><mml:mi>cos⁡</mml:mi><mml:mo>(</mml:mo><mml:mi mathvariant="italic">ϕ</mml:mi><mml:mo>)</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E11"><mml:mtd><mml:mtext>11</mml:mtext></mml:mtd><mml:mtd><mml:mstyle class="stylechange" displaystyle="true"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle class="stylechange" displaystyle="true"/><mml:msup><mml:mi>y</mml:mi><mml:mo>′</mml:mo></mml:msup><mml:mo>=</mml:mo><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mo>+</mml:mo><mml:mi>r</mml:mi><mml:mo>×</mml:mo><mml:mi mathvariant="italic">θ</mml:mi><mml:mo>×</mml:mo><mml:mi>sin⁡</mml:mi><mml:mo>(</mml:mo><mml:mi mathvariant="italic">ϕ</mml:mi><mml:mo>)</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:mtd></mml:mlabeledtr></mml:mtable></mml:math></disp-formula>

            where <inline-formula><mml:math id="M36" display="inline"><mml:mi>r</mml:mi></mml:math></inline-formula> is the distance between the center pixel (<inline-formula><mml:math id="M37" display="inline"><mml:mrow><mml:mi>c</mml:mi><mml:mi>x</mml:mi></mml:mrow></mml:math></inline-formula>, <inline-formula><mml:math id="M38" display="inline"><mml:mrow><mml:mi>c</mml:mi><mml:mi>y</mml:mi></mml:mrow></mml:math></inline-formula>) of the original image and each pixel (<inline-formula><mml:math id="M39" display="inline"><mml:mi>x</mml:mi></mml:math></inline-formula>, <inline-formula><mml:math id="M40" display="inline"><mml:mi>y</mml:mi></mml:math></inline-formula>), <inline-formula><mml:math id="M41" display="inline"><mml:mi mathvariant="italic">θ</mml:mi></mml:math></inline-formula> is the solar zenith angle (SZA), “radi” is the image radius (distance between center and edge pixel of circular images), <inline-formula><mml:math id="M42" display="inline"><mml:mi mathvariant="italic">ϕ</mml:mi></mml:math></inline-formula> is the azimuth, and <inline-formula><mml:math id="M43" display="inline"><mml:mrow><mml:msup><mml:mi>x</mml:mi><mml:mo>′</mml:mo></mml:msup></mml:mrow></mml:math></inline-formula> and <inline-formula><mml:math id="M44" display="inline"><mml:mrow><mml:msup><mml:mi>y</mml:mi><mml:mo>′</mml:mo></mml:msup></mml:mrow></mml:math></inline-formula> are the coordinates of each pixel after distortion correction.</p>
      <p id="d1e1237">Second, surrounding masks such as buildings, trees, and equipment, as well as light sources such as the sun, moon, and stars, were removed from the image (building, tree, and equipment: masking was performed when the mean RGB brightness was less than 60 in the daytime on a clear day; light source: masking was performed when the mean RGB brightness exceeded 240). These
objects directly mask the sky and clouds or make it difficult to distinguish
them; therefore, they must be removed when calculating cloud cover (Yabuki
et al., 2014; Kim et al., 2016; Kim and Cha, 2020). Third, the RBR, BRD, and <inline-formula><mml:math id="M45" display="inline"><mml:mi>Y</mml:mi></mml:math></inline-formula> frequency distributions were calculated using the RGB brightness of each pixel of image data subjected to pre-processing (i.e., masking removal and distortion correction). The class interval sizes of the RBR, BRD, and <inline-formula><mml:math id="M46" display="inline"><mml:mi>Y</mml:mi></mml:math></inline-formula> frequency distributions were set to 0.02, 2, and 2, respectively, and classes with frequencies of less than 100 were ignored. Statistical characteristics of the mean, mode, frequency of mode, kurtosis, skewness, and quantile (<inline-formula><mml:math id="M47" display="inline"><mml:mrow><mml:msub><mml:mi>Q</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula>–<inline-formula><mml:math id="M48" display="inline"><mml:mrow><mml:msub><mml:mi>Q</mml:mi><mml:mn mathvariant="normal">4</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula>: 0 %, 25 %, 50 %, 75 %, and 100 %) data obtained for each frequency distribution were used as input for each machine learning method. As input data for machine learning, time information (Julian day and hour) allowed differentiating seasons and day and night. Further, SZA should be considered because the colors of the sky and clouds change according to the position of the sun (Blazek and Pata, 2015; Cazorla et al., 2015; Azhar et al., 2021). As these image data features have different appearances under different conditions (cloud cover, day, night, etc.), they constitute an important variable in machine learning regression for cloud cover calculation (Heinle et al., 2010; Li et al., 2011).</p>

      <?xmltex \floatpos{p}?><fig id="Ch1.F2" specific-use="star"><?xmltex \currentcnt{2}?><?xmltex \def\figurename{Figure}?><label>Figure 2</label><caption><p id="d1e1278">Distortion-corrected images and RBR <bold>(d, j)</bold>, BRD <bold>(e, k)</bold>, and <inline-formula><mml:math id="M49" display="inline"><mml:mi>Y</mml:mi></mml:math></inline-formula> <bold>(f, l)</bold> relative frequency distributions for cloud-free (0 tenths), overcast (10 tenths), and partly cloudy (5 tenths) cases at day and night. The daytime cloud-free <bold>(a)</bold>, overcast <bold>(b)</bold>, and partly cloudy <bold>(c)</bold> data were obtained at
14:00 LST on 8 March, 12:00 LST on 15 July, and 15:00 LST on 28 September 2019. Cloud-free <bold>(g)</bold>, overcast <bold>(h)</bold>, and partly cloudy <bold>(i)</bold> nighttime data were obtained at 03:00 LST on 24 January, 20:00 LST on 18 February, and 22:00 LST on 30 April 2019. The green and red areas are masked to remove surrounding masks (i.e., buildings, trees, and equipment) and light sources (i.e., the sun and moon), respectively.</p></caption>
          <?xmltex \igopts{width=455.244094pt}?><graphic xlink:href="https://amt.copernicus.org/articles/14/6695/2021/amt-14-6695-2021-f02.png"/>

        </fig>

      <?pagebreak page6700?><p id="d1e1323">Figure 2 shows distortion-corrected images for day and night cloud-free,
overcast, and partly cloudy cases and the RBR, BRD, and <inline-formula><mml:math id="M50" display="inline"><mml:mi>Y</mml:mi></mml:math></inline-formula> relative frequency distributions. The frequency distributions were expressed as percentages over approximately 310 000 px excluding the masked area. Human-eye observations at DROM yielded cloud-free (Fig. 2a and g), overcast (Fig. 2b and h), and partly cloudy (Fig. 2c and i) case values of 0, 10, and 5 tenths, respectively. As for the RBR frequency distribution during the day,
larger RBR distributions were observed for the overcast than cloud-free case, and bimodal distributions including both (i.e., overcast and cloud-free) distributions were obtained for the partly cloudy case. The variance was large in the partly cloudy case. With regard to the BRD frequency distribution, the blue-channel brightness increased with Rayleigh scattering, such that the cloud-free case with many sky pixels had larger BRD distribution than the overcast case (Ghonima et al., 2012; Kim et al., 2016). In contrast, the <inline-formula><mml:math id="M51" display="inline"><mml:mi>Y</mml:mi></mml:math></inline-formula> frequency distribution was relatively large for the overcast case, which involved many cloud pixels. Although the RBR frequency distributions at night and day were similar, the RBR was larger at night because the red-channel brightness increased under the influence of Mie scattering (Kyba et al., 2012; Kim and Cha, 2020). A negative BRD distribution was obtained from the cloud pixels. At night, there is no light source such as the sun. Therefore, in this study, RGB brightness close to
black (0, 0, 0) was distributed in the cloud-free case, yielding small <inline-formula><mml:math id="M52" display="inline"><mml:mi>Y</mml:mi></mml:math></inline-formula>. As the images obtained through ACOS had different RBR, BRD, and <inline-formula><mml:math id="M53" display="inline"><mml:mi>Y</mml:mi></mml:math></inline-formula> frequency
distribution classes and shapes for each case, it was necessary to train
these data features (i.e., the mean, mode, frequency of mode, kurtosis,
skewness, and quantile of each frequency distribution) on a machine learning
model to calculate the cloud cover.</p>
</sec>
</sec>
<sec id="Ch1.S3">
  <label>3</label><title>Machine learning methods</title>
      <p id="d1e1363">Depending on the machine learning method, even if the accuracy, recall,
precision, and <inline-formula><mml:math id="M54" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula> of the trained model are high and the bias and RMSE are
small, overfitting problems may occur when data other than training data are
used for prediction; these problems can yield low prediction performance
(Ying, 2019). Therefore, in this study, optimal hyper-parameters were set by
iteratively changing the hyper-parameter for each machine learning method
using the training and validation sets (Bergstra and Bengio, 2012). The optimal hyper-parameter was determined based on the accuracy, recall,
precision, bias, RMSE, and <inline-formula><mml:math id="M55" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula>, which were prediction performance indicators for each iteration. The details and hyper-parameter settings of each supervised machine learning method used in this study are described in
Sect. 3.1 to 3.6. The prediction results of each machine learning method are compared in Sect. 4.1.</p>
<sec id="Ch1.S3.SS1">
  <label>3.1</label><title>Multiple linear regression (MLR)</title>
      <p id="d1e1387">The method in which the relationship of the dependent variable to the independent variable is regressed by considering one independent variable
only is called simple linear regression, and the method in which the change
in the dependent variable is predicted based on the changes in two or more
independent variables is called MLR. An MLR model with <inline-formula><mml:math id="M56" display="inline"><mml:mi>k</mml:mi></mml:math></inline-formula> independent variables predicts the dependent variable as shown in Eq. (12), using the
least squares method which minimizes the predictor variable and the sum of
squared errors (Fig. 3a) (Olive, 2017). In this study, we used the R “glm”
package (Geyer, 2003).
            <disp-formula id="Ch1.E12" content-type="numbered"><label>12</label><mml:math id="M57" display="block"><mml:mrow><?xmltex \hack{\hbox\bgroup\fontsize{9.2}{9.2}\selectfont$\displaystyle}?><mml:msub><mml:mi>Y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">1</mml:mn></mml:msub><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mn mathvariant="normal">1</mml:mn><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msub><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mn mathvariant="normal">2</mml:mn><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:mspace width="0.125em" linebreak="nobreak"/><mml:mi mathvariant="normal">⋯</mml:mi><mml:mspace width="0.125em" linebreak="nobreak"/><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mspace linebreak="nobreak" width="0.25em"/><mml:mspace width="0.25em" linebreak="nobreak"/><mml:mspace linebreak="nobreak" width="0.25em"/><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn><mml:mo>,</mml:mo><mml:mn mathvariant="normal">2</mml:mn><mml:mo>,</mml:mo><mml:mspace width="0.125em" linebreak="nobreak"/><mml:mi mathvariant="normal">⋯</mml:mi><mml:mo>,</mml:mo><mml:mspace linebreak="nobreak" width="0.125em"/><mml:mi>N</mml:mi><?xmltex \hack{$\egroup}?><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
          where <inline-formula><mml:math id="M58" display="inline"><mml:mrow><mml:msub><mml:mi mathvariant="italic">β</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula> are the population coefficients (i.e., parameters), and
<inline-formula><mml:math id="M59" display="inline"><mml:mrow><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> is the <inline-formula><mml:math id="M60" display="inline"><mml:mi>k</mml:mi></mml:math></inline-formula>th predictor of the <inline-formula><mml:math id="M61" display="inline"><mml:mi>i</mml:mi></mml:math></inline-formula>th observation (a value that describes the variable <inline-formula><mml:math id="M62" display="inline"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula> to be predicted). In this study, the independent variables were the RGB, BRD, and <inline-formula><mml:math id="M63" display="inline"><mml:mi>Y</mml:mi></mml:math></inline-formula> mean, mode, frequency of mode, skewness, kurtosis, quantile, as well as the Julian day, hour, and SZA. The dependent variable was the cloud cover observed by human eyes.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F3" specific-use="star"><?xmltex \currentcnt{3}?><?xmltex \def\figurename{Figure}?><label>Figure 3</label><caption><p id="d1e1558">Schematic of each machine learning method: MLR <bold>(a)</bold>, kNN <bold>(b)</bold>, SVR <bold>(c)</bold>, ANN <bold>(d)</bold>, RF <bold>(e)</bold>, and GBM <bold>(f)</bold>.</p></caption>
          <?xmltex \igopts{width=341.433071pt}?><graphic xlink:href="https://amt.copernicus.org/articles/14/6695/2021/amt-14-6695-2021-f03.png"/>

        </fig>

<?xmltex \hack{\newpage}?>
</sec>
<sec id="Ch1.S3.SS2">
  <label>3.2</label><?xmltex \opttitle{$k$-nearest neighbor~(kNN)}?><title><inline-formula><mml:math id="M64" display="inline"><mml:mi>k</mml:mi></mml:math></inline-formula>-nearest neighbor (kNN)</title>
      <p id="d1e1603">The kNN method involves non-parametric, instance-based learning and is one
of the simplest predictive models in machine learning. The kNN algorithm
finds the <inline-formula><mml:math id="M65" display="inline"><mml:mi>k</mml:mi></mml:math></inline-formula>-nearest neighbors to the query in the data feature space, as
shown in Fig. 3b, and then predicts the query with distance-based weights
(S. Zhang et al., 2018). That is, a set of independent variables is
constructed as a cluster, and values corresponding to each neighbor are
weighted according to the Euclidean distance and predicted (Martínez et
al., 2019). In this study, the R “class” package (Ripley and Venables,
2021a) was used, and the hyper-parameter setting was <inline-formula><mml:math id="M66" display="inline"><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">15</mml:mn></mml:mrow></mml:math></inline-formula>.</p>
</sec>
<sec id="Ch1.S3.SS3">
  <label>3.3</label><title>Support vector regression (SVR)</title>
      <p id="d1e1633">SVR is an extended method that can be used for regression analysis by introducing an <inline-formula><mml:math id="M67" display="inline"><mml:mi mathvariant="italic">ε</mml:mi></mml:math></inline-formula>-insensitive loss function to an SVM. As shown
in Fig. 3c, a hyperplane consisting of support vectors that can classify the
maximum margin for the distance between vectors is found (Gani et al., 2010;
Taghizadeh-Mehrjardi et al., 2017). The optimal hyperplane is obtained by
finding <inline-formula><mml:math id="M68" display="inline"><mml:mi>w</mml:mi></mml:math></inline-formula> and <inline-formula><mml:math id="M69" display="inline"><mml:mi>b</mml:mi></mml:math></inline-formula> that minimize the mapping function (<inline-formula><mml:math id="M70" display="inline"><mml:mrow><mml:mi mathvariant="normal">Φ</mml:mi><mml:mo>(</mml:mo><mml:mi>w</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>), as shown in Eq. (13) (Meyer, 2001). The constraints are shown in Eq. (14). Then, as in Eq. (15), the kernel is applied and mapped to a higher dimension. Here, <inline-formula><mml:math id="M71" display="inline"><mml:mi mathvariant="italic">ε</mml:mi></mml:math></inline-formula> determines the threshold of margin, <inline-formula><mml:math id="M72" display="inline"><mml:mi mathvariant="italic">ξ</mml:mi></mml:math></inline-formula> is a slack variable to allow error, and <inline-formula><mml:math id="M73" display="inline"><mml:mi>C</mml:mi></mml:math></inline-formula> is the allowable cost that can violate the constraint of Eq. (14). In this study, the R “e1071” package (Meyer et al., 2021) was used, the SVR kernel was set as a radial basis function (RBF), and the hyper-parameters were set to <italic>epsilon</italic> <inline-formula><mml:math id="M74" display="inline"><mml:mrow><mml:mo>(</mml:mo><mml:mi mathvariant="italic">ε</mml:mi><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:mn mathvariant="normal">0.12</mml:mn></mml:mrow></mml:math></inline-formula>, <italic>gamma</italic> <inline-formula><mml:math id="M75" display="inline"><mml:mrow><mml:mo>(</mml:mo><mml:mi mathvariant="italic">γ</mml:mi><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:mn mathvariant="normal">0.04</mml:mn></mml:mrow></mml:math></inline-formula>, and <italic>cost</italic> <inline-formula><mml:math id="M76" display="inline"><mml:mrow><mml:mo>(</mml:mo><mml:mi>C</mml:mi><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:mn mathvariant="normal">5</mml:mn></mml:mrow></mml:math></inline-formula>.

                <disp-formula specific-use="gather" content-type="numbered"><mml:math id="M77" display="block"><mml:mtable displaystyle="true"><mml:mlabeledtr id="Ch1.E13"><mml:mtd><mml:mtext>13</mml:mtext></mml:mtd><mml:mtd><mml:mrow><mml:mstyle class="stylechange" displaystyle="true"/><mml:mrow><mml:mi mathvariant="normal">Φ</mml:mi><mml:mo>(</mml:mo><mml:mi>w</mml:mi><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:mi mathvariant="normal">min</mml:mi><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mn mathvariant="normal">1</mml:mn><mml:mn mathvariant="normal">2</mml:mn></mml:mfrac></mml:mstyle><mml:msup><mml:mfenced close="∥" open="∥"><mml:mi>w</mml:mi></mml:mfenced><mml:mn mathvariant="normal">2</mml:mn></mml:msup><mml:mo>+</mml:mo><mml:mi>C</mml:mi><mml:munderover><mml:mo movablelimits="false">∑</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mi>n</mml:mi></mml:munderover><mml:mfenced open="(" close=")"><mml:mrow><mml:msub><mml:mi mathvariant="italic">ξ</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>+</mml:mo><mml:msubsup><mml:mi mathvariant="italic">ξ</mml:mi><mml:mi>i</mml:mi><mml:mo>*</mml:mo></mml:msubsup></mml:mrow></mml:mfenced><mml:mo>,</mml:mo></mml:mrow></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E14"><mml:mtd><mml:mtext>14</mml:mtext></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:mtable rowspacing="0.2ex" class="split" displaystyle="true" columnalign="right left"><mml:mtr><mml:mtd><mml:mrow><mml:mfenced close=")" open="("><mml:mrow><mml:msup><mml:mi>w</mml:mi><mml:mi>T</mml:mi></mml:msup><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>+</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:mfenced></mml:mrow></mml:mtd><mml:mtd><mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>≤</mml:mo><mml:mi mathvariant="italic">ε</mml:mi><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="italic">ξ</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>,</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd/><mml:mtd><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>-</mml:mo><mml:mfenced open="(" close=")"><mml:mrow><mml:msup><mml:mi>w</mml:mi><mml:mi>T</mml:mi></mml:msup><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>+</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:mfenced><mml:mo>-</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>≤</mml:mo><mml:mi mathvariant="italic">ε</mml:mi><mml:mo>+</mml:mo><mml:msubsup><mml:mi mathvariant="italic">ξ</mml:mi><mml:mi>i</mml:mi><mml:mo>*</mml:mo></mml:msubsup><mml:mo>,</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd/><mml:mtd><mml:mrow><mml:msub><mml:mi mathvariant="italic">ξ</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi mathvariant="italic">ξ</mml:mi><mml:mi>i</mml:mi><mml:mo>*</mml:mo></mml:msubsup><mml:mo>≥</mml:mo><mml:mn mathvariant="normal">0</mml:mn><mml:mo>,</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E15"><mml:mtd><mml:mtext>15</mml:mtext></mml:mtd><mml:mtd><mml:mrow><mml:mstyle class="stylechange" displaystyle="true"/><mml:mrow><mml:mi>K</mml:mi><mml:mfenced open="(" close=")"><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:mfenced><mml:mo>=</mml:mo><mml:mi>exp⁡</mml:mi><mml:mfenced open="(" close=")"><mml:mrow><mml:mo>-</mml:mo><mml:mi mathvariant="italic">γ</mml:mi><mml:msup><mml:mfenced open="(" close=")"><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:mfenced><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:mfenced><mml:mo>,</mml:mo></mml:mrow></mml:mrow></mml:mtd></mml:mlabeledtr></mml:mtable></mml:math></disp-formula>

            where subscript <inline-formula><mml:math id="M78" display="inline"><mml:mi>i</mml:mi></mml:math></inline-formula> and <inline-formula><mml:math id="M79" display="inline"><mml:mi>j</mml:mi></mml:math></inline-formula> are <inline-formula><mml:math id="M80" display="inline"><mml:mi>i</mml:mi></mml:math></inline-formula>th and <inline-formula><mml:math id="M81" display="inline"><mml:mi>j</mml:mi></mml:math></inline-formula>th data point, respectively, and
<inline-formula><mml:math id="M82" display="inline"><mml:mi mathvariant="italic">γ</mml:mi></mml:math></inline-formula> is a parameter that controls the RBF kernel width.</p>
</sec>
<sec id="Ch1.S3.SS4">
  <label>3.4</label><title>Artificial neural network (ANN)</title>
      <p id="d1e2044">An ANN is a mathematical model that mimics a neuron; i.e., the signal
transmission system of a biological neural network. As shown in Fig. 3d,
this model consists of an input layer that receives input data, an output
layer that outputs prediction results, and an invisible hidden layer between
the two layers (Rosa et al., 2020). The hidden node of the hidden layer acts
like a neuron in a neural network and is composed<?pagebreak page6701?> of weight, bias, and an
activation function. In this study, we used the R “nnet” package (Ripley
and Venables, 2021b), which is based on feed-forward neural networks with a
single hidden layer that can rapidly learn and predict while considering
nowcasting. The hyper-parameters of this package were set as follows: <italic>size</italic> (number of hidden nodes) <inline-formula><mml:math id="M83" display="inline"><mml:mo>=</mml:mo></mml:math></inline-formula> 7, <italic>maxit</italic> (maximum number of iterations) <inline-formula><mml:math id="M84" display="inline"><mml:mo>=</mml:mo></mml:math></inline-formula> 700, and <italic>decay</italic> (weight decay parameter) <inline-formula><mml:math id="M85" display="inline"><mml:mo>=</mml:mo></mml:math></inline-formula> 0.05.</p>
</sec>
<sec id="Ch1.S3.SS5">
  <label>3.5</label><title>Random forest (RF)</title>
      <p id="d1e2086">The RF method composes <inline-formula><mml:math id="M86" display="inline"><mml:mi>N</mml:mi></mml:math></inline-formula> decision trees by combining randomly selected
variables from each node to grow a regression tree, as shown in Fig. 3e. An
ensemble of the results of each decision tree is obtained, and hence a
prediction result is provided (Wright and Ziegler, 2017). That is, in the RF ensemble learning method, every individual tree of the decision tree contributes to the final prediction (Shin et al., 2020; Kim et al., 2021). In this study, the R “Ranger” package (Wright et al., 2020) was used, and the hyper-parameters were set to <italic>num.trees</italic> (the number of trees) <inline-formula><mml:math id="M87" display="inline"><mml:mo>=</mml:mo></mml:math></inline-formula> 510, <italic>mtry</italic> (the number of variables randomly sampled from each node) <inline-formula><mml:math id="M88" display="inline"><mml:mo>=</mml:mo></mml:math></inline-formula> 7, and <italic>min.node.size</italic> (minimal node size) <inline-formula><mml:math id="M89" display="inline"><mml:mo>=</mml:mo></mml:math></inline-formula> 5.</p>

<?xmltex \floatpos{t}?><table-wrap id="Ch1.T3" specific-use="star"><?xmltex \currentcnt{3}?><label>Table 3</label><caption><p id="d1e2130">Prediction performance for all cloud cover of machine learning methods using training and validation sets.</p></caption><oasis:table frame="topbot"><oasis:tgroup cols="8">
     <oasis:colspec colnum="1" colname="col1" align="left"/>
     <oasis:colspec colnum="2" colname="col2" align="left"/>
     <oasis:colspec colnum="3" colname="col3" align="center"/>
     <oasis:colspec colnum="4" colname="col4" align="center"/>
     <oasis:colspec colnum="5" colname="col5" align="center"/>
     <oasis:colspec colnum="6" colname="col6" align="right"/>
     <oasis:colspec colnum="7" colname="col7" align="center"/>
     <oasis:colspec colnum="8" colname="col8" align="center"/>
     <oasis:thead>
       <oasis:row rowsep="1">

         <oasis:entry colname="col1">Model</oasis:entry>

         <oasis:entry colname="col2">Set</oasis:entry>

         <oasis:entry colname="col3">Accuracy</oasis:entry>

         <oasis:entry colname="col4">Recall</oasis:entry>

         <oasis:entry colname="col5">Precision</oasis:entry>

         <oasis:entry colname="col6">Bias</oasis:entry>

         <oasis:entry colname="col7">RMSE</oasis:entry>

         <oasis:entry colname="col8"><inline-formula><mml:math id="M90" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula></oasis:entry>

       </oasis:row>
     </oasis:thead>
     <oasis:tbody>
       <oasis:row>

         <oasis:entry rowsep="1" colname="col1" morerows="1">SVR</oasis:entry>

         <oasis:entry colname="col2">Training</oasis:entry>

         <oasis:entry colname="col3">0.94</oasis:entry>

         <oasis:entry colname="col4">0.70</oasis:entry>

         <oasis:entry colname="col5">0.76</oasis:entry>

         <oasis:entry colname="col6">0.07</oasis:entry>

         <oasis:entry colname="col7">1.05</oasis:entry>

         <oasis:entry colname="col8">0.96</oasis:entry>

       </oasis:row>
       <oasis:row rowsep="1">

         <oasis:entry colname="col2">Validation</oasis:entry>

         <oasis:entry colname="col3">0.88</oasis:entry>

         <oasis:entry colname="col4">0.41</oasis:entry>

         <oasis:entry colname="col5">0.51</oasis:entry>

         <oasis:entry colname="col6">0.06</oasis:entry>

         <oasis:entry colname="col7">1.51</oasis:entry>

         <oasis:entry colname="col8">0.93</oasis:entry>

       </oasis:row>
       <oasis:row>

         <oasis:entry rowsep="1" colname="col1" morerows="1">RF</oasis:entry>

         <oasis:entry colname="col2">Training</oasis:entry>

         <oasis:entry colname="col3">0.93</oasis:entry>

         <oasis:entry colname="col4">0.67</oasis:entry>

         <oasis:entry colname="col5">0.76</oasis:entry>

         <oasis:entry colname="col6">0.02</oasis:entry>

         <oasis:entry colname="col7">0.71</oasis:entry>

         <oasis:entry colname="col8">0.98</oasis:entry>

       </oasis:row>
       <oasis:row rowsep="1">

         <oasis:entry colname="col2">Validation</oasis:entry>

         <oasis:entry colname="col3">0.86</oasis:entry>

         <oasis:entry colname="col4">0.35</oasis:entry>

         <oasis:entry colname="col5">0.53</oasis:entry>

         <oasis:entry colname="col6"><inline-formula><mml:math id="M91" display="inline"><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">0.03</mml:mn></mml:mrow></mml:math></inline-formula></oasis:entry>

         <oasis:entry colname="col7">1.55</oasis:entry>

         <oasis:entry colname="col8">0.92</oasis:entry>

       </oasis:row>
       <oasis:row>

         <oasis:entry rowsep="1" colname="col1" morerows="1">GBM</oasis:entry>

         <oasis:entry colname="col2">Training</oasis:entry>

         <oasis:entry colname="col3">0.89</oasis:entry>

         <oasis:entry colname="col4">0.47</oasis:entry>

         <oasis:entry colname="col5">0.59</oasis:entry>

         <oasis:entry colname="col6"><inline-formula><mml:math id="M92" display="inline"><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">0.00</mml:mn></mml:mrow></mml:math></inline-formula></oasis:entry>

         <oasis:entry colname="col7">1.03</oasis:entry>

         <oasis:entry colname="col8">0.97</oasis:entry>

       </oasis:row>
       <oasis:row rowsep="1">

         <oasis:entry colname="col2">Validation</oasis:entry>

         <oasis:entry colname="col3">0.86</oasis:entry>

         <oasis:entry colname="col4">0.36</oasis:entry>

         <oasis:entry colname="col5">0.50</oasis:entry>

         <oasis:entry colname="col6"><inline-formula><mml:math id="M93" display="inline"><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">0.06</mml:mn></mml:mrow></mml:math></inline-formula></oasis:entry>

         <oasis:entry colname="col7">1.58</oasis:entry>

         <oasis:entry colname="col8">0.92</oasis:entry>

       </oasis:row>
       <oasis:row>

         <oasis:entry rowsep="1" colname="col1" morerows="1">kNN</oasis:entry>

         <oasis:entry colname="col2">Training</oasis:entry>

         <oasis:entry colname="col3">0.88</oasis:entry>

         <oasis:entry colname="col4">0.41</oasis:entry>

         <oasis:entry colname="col5">0.57</oasis:entry>

         <oasis:entry colname="col6">0.15</oasis:entry>

         <oasis:entry colname="col7">1.41</oasis:entry>

         <oasis:entry colname="col8">0.94</oasis:entry>

       </oasis:row>
       <oasis:row rowsep="1">

         <oasis:entry colname="col2">Validation</oasis:entry>

         <oasis:entry colname="col3">0.87</oasis:entry>

         <oasis:entry colname="col4">0.37</oasis:entry>

         <oasis:entry colname="col5">0.51</oasis:entry>

         <oasis:entry colname="col6">0.12</oasis:entry>

         <oasis:entry colname="col7">1.78</oasis:entry>

         <oasis:entry colname="col8">0.90</oasis:entry>

       </oasis:row>
       <oasis:row>

         <oasis:entry rowsep="1" colname="col1" morerows="1">ANN</oasis:entry>

         <oasis:entry colname="col2">Training</oasis:entry>

         <oasis:entry colname="col3">0.86</oasis:entry>

         <oasis:entry colname="col4">0.33</oasis:entry>

         <oasis:entry colname="col5">0.49</oasis:entry>

         <oasis:entry colname="col6">0.03</oasis:entry>

         <oasis:entry colname="col7">1.69</oasis:entry>

         <oasis:entry colname="col8">0.91</oasis:entry>

       </oasis:row>
       <oasis:row rowsep="1">

         <oasis:entry colname="col2">Validation</oasis:entry>

         <oasis:entry colname="col3">0.85</oasis:entry>

         <oasis:entry colname="col4">0.31</oasis:entry>

         <oasis:entry colname="col5">0.46</oasis:entry>

         <oasis:entry colname="col6">0.01</oasis:entry>

         <oasis:entry colname="col7">1.92</oasis:entry>

         <oasis:entry colname="col8">0.88</oasis:entry>

       </oasis:row>
       <oasis:row>

         <oasis:entry colname="col1" morerows="1">MLR</oasis:entry>

         <oasis:entry colname="col2">Training</oasis:entry>

         <oasis:entry colname="col3">0.84</oasis:entry>

         <oasis:entry colname="col4">0.27</oasis:entry>

         <oasis:entry colname="col5">0.46</oasis:entry>

         <oasis:entry colname="col6"><inline-formula><mml:math id="M94" display="inline"><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">0.02</mml:mn></mml:mrow></mml:math></inline-formula></oasis:entry>

         <oasis:entry colname="col7">1.90</oasis:entry>

         <oasis:entry colname="col8">0.88</oasis:entry>

       </oasis:row>
       <oasis:row>

         <oasis:entry colname="col2">Validation</oasis:entry>

         <oasis:entry colname="col3">0.84</oasis:entry>

         <oasis:entry colname="col4">0.27</oasis:entry>

         <oasis:entry colname="col5">0.46</oasis:entry>

         <oasis:entry colname="col6"><inline-formula><mml:math id="M95" display="inline"><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">0.02</mml:mn></mml:mrow></mml:math></inline-formula></oasis:entry>

         <oasis:entry colname="col7">1.94</oasis:entry>

         <oasis:entry colname="col8">0.87</oasis:entry>

       </oasis:row>
     </oasis:tbody>
   </oasis:tgroup></oasis:table></table-wrap>

</sec>
<sec id="Ch1.S3.SS6">
  <label>3.6</label><title>Gradient boosting machine (GBM)</title>
      <p id="d1e2562">The GBM uses boosting instead of bagging during resampling and ensemble
processes. As shown in Fig. 3f, a model with improved predictive power is
created by gradually improving upon the parts that the previous model could
not predict while sequentially generating weak models. The final prediction
is calculated from the weighted mean of these<?pagebreak page6702?> results (Friedman, 2001). In
other words, gradient boosting updates the weights iteratively to minimize
the difference from the function <inline-formula><mml:math id="M96" display="inline"><mml:mrow><mml:mi>f</mml:mi><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> that predicts the actual observation
using gradient descent (Ridgeway, 2020). In this study, the R “gbm”
package (Greenwell et al., 2020) was used; the GBM kernel was set to a
Gaussian distribution function; and the hyper-parameters were set to
<italic>n.trees</italic> (number of trees) <inline-formula><mml:math id="M97" display="inline"><mml:mo>=</mml:mo></mml:math></inline-formula> 500, <italic>interaction.depth</italic> (maximum depth of binary tree) <inline-formula><mml:math id="M98" display="inline"><mml:mo>=</mml:mo></mml:math></inline-formula> 5, and <italic>shrinkage</italic> (learning rate) <inline-formula><mml:math id="M99" display="inline"><mml:mo>=</mml:mo></mml:math></inline-formula> 0.1.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F4" specific-use="star"><?xmltex \currentcnt{4}?><?xmltex \def\figurename{Figure}?><label>Figure 4</label><caption><p id="d1e2612">Scatter plots of observed cloud cover and that predicted by machine learning methods – SVR <bold>(a)</bold>, RF <bold>(b)</bold>, GBM <bold>(c)</bold>, kNN <bold>(d)</bold>, ANN <bold>(e)</bold>, and MLR <bold>(f)</bold> – on the training set. The number of observations for each observed cloud cover are 0: 992, 1: 136, 2: 141, 3: 156, 4: 149, 5: 135, 6: 221, 7: 271, 8: 320, 9: 441, and 10: 739.</p></caption>
          <?xmltex \igopts{width=341.433071pt}?><graphic xlink:href="https://amt.copernicus.org/articles/14/6695/2021/amt-14-6695-2021-f04.png"/>

        </fig>

</sec>
</sec>
<sec id="Ch1.S4">
  <label>4</label><title>Results</title>
<sec id="Ch1.S4.SS1">
  <label>4.1</label><title>Training and validation results of machine learning methods</title>
      <p id="d1e2657">Figure 4 shows the cloud cover prediction results obtained using the training set for each machine learning method. The hyper-parameters were optimized using the training and validation sets. Each box in the figure denotes the ratio (%) of the number of observations for each cloud cover in the DROM and the number of predictions for each cloud cover in the SVR model. The higher the frequency in the diagonal one-to-one boxes, the better the agreement between the observed and predicted cloud cover. In other words, the closer the diagonal one-to-one boxes are to red (i.e., 100 %), the higher is the agreement. For the training set, the highest human-eye observation data frequency by cloud cover was 26.80 % at 0 tenths; this was followed by 19.97 % at 10 tenths and 3.65 %–11.92 % at 1–9 tenths. For the SVR model, the 0- and 10-tenth frequencies were 71.88 % and 92.15 %, respectively, with the agreement being greatest among the machine learning models. As detailed in Table 3, the SVR accuracy, recall, and precision for all cloud cover were 0.94, 0.70, and 0.76, respectively, indicating the best prediction performance. The accuracy was in the range of 0.91–0.98 for each cloud cover, whereas recall and precision were in the ranges of 0.42–0.92 and 0.24–0.99, exhibiting low predictive power in the partly cloudy case. The bias was 0.07 tenths, the RMSE was 1.05 tenths, and <inline-formula><mml:math id="M100" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula> was 0.96. In the case of the RF model, the 0- and 10-tenth frequencies were 61.79 % and 80.65 %, respectively, being lower than those of the SVR model; however, the prediction for 1–9 tenths exhibited high agreement to within <inline-formula><mml:math id="M101" display="inline"><mml:mrow><mml:mo>±</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow></mml:math></inline-formula> tenth. The accuracy, recall, and precision were 0.93, 0.67, and 0.76, respectively, which is lower than SVR model, but the bias and RMSE were the smallest at 0.02 and 0.71 tenths, respectively, and the <inline-formula><mml:math id="M102" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula> value was the highest at 0.98. However, for the validation set, the SVR model prediction performance (accuracy: 0.88, recall: 0.41, precision: 0.51, bias: 0.06 tenths, RMSE: 1.51 tenths, and <inline-formula><mml:math id="M103" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula>: 0.93) was better than that of the RF model. In other words, the RF model exhibited a tendency to overfit in this study. The accuracy of these results exceeds that of the classification machine learning method (0.60–0.85) presented by Dev et al. (2016) using day and night image data, and it is higher than or similar to the accuracy (0.91–0.94) achieved using the regression and deep learning machine learning methods proposed by Shi et al. (2019, 2021) for day and night image data. Apart from the SVR and RF methods, the machine learning methods exhibited similar frequency distributions; however, the accuracy, recall, precision, and <inline-formula><mml:math id="M104" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula> were lower and the RMSE values were higher in the order of GBM, kNN, ANN, and MLR. In particular, the MLR model had very poor predictive power (accuracy: 0.75, recall: 0.08, and precision: 0.78) for 0 tenths using the training set.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F5"><?xmltex \currentcnt{5}?><?xmltex \def\figurename{Figure}?><label>Figure 5</label><caption><p id="d1e2700">Variable relative importance of the SVR model on training set.</p></caption>
          <?xmltex \igopts{width=236.157874pt}?><graphic xlink:href="https://amt.copernicus.org/articles/14/6695/2021/amt-14-6695-2021-f05.png"/>

        </fig>

      <?pagebreak page6703?><p id="d1e2709">The relative importance of the input variable of the SVR method, which
exhibited the best predictive performance in this study, is shown in Fig. 5.
BRD<inline-formula><mml:math id="M105" display="inline"><mml:msub><mml:mi/><mml:mrow><mml:msub><mml:mi>Q</mml:mi><mml:mn mathvariant="normal">4</mml:mn></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> had the highest relative importance at 8.54 %, whereas
RBR<inline-formula><mml:math id="M106" display="inline"><mml:msub><mml:mi/><mml:mi mathvariant="normal">mode</mml:mi></mml:msub></mml:math></inline-formula> had the lowest importance at 0.55 %. Among the RBR data
features, RBR<inline-formula><mml:math id="M107" display="inline"><mml:msub><mml:mi/><mml:mrow><mml:msub><mml:mi>Q</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> had the highest importance at 7.06 %, and, among the
<inline-formula><mml:math id="M108" display="inline"><mml:mi>Y</mml:mi></mml:math></inline-formula> data features, <inline-formula><mml:math id="M109" display="inline"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:msub><mml:mi>Q</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> had the highest importance at 3.78 %. In terms of the cumulative relative importance, the <?xmltex \hack{\mbox\bgroup}?>BRD-<?xmltex \hack{\egroup}?>, RBR- and <inline-formula><mml:math id="M110" display="inline"><mml:mi>Y</mml:mi></mml:math></inline-formula>-related data features contributed 38.25 %, 31.44 %, and 26.20 % of the total (100 %), respectively, to the cloud cover prediction, and the remaining data features contributed 4.10 %. The relationship between input data features is complex to determine the optimal hyperplane of the SVR model, and the variable importance is determined so that the cloud cover can be calculated with the smallest error using the observed cloud cover (Singh et al., 2020). Even if the BRD-related data features have the same RBR characteristics, they contribute to machine learning more comprehensively by
day, night, and cloud presence depending on the BRD value; therefore, they
are critical to cloud cover calculation. By contrast, the <inline-formula><mml:math id="M111" display="inline"><mml:mi>Y</mml:mi></mml:math></inline-formula>-related data
feature is sensitive to the RGB brightness (especially the <inline-formula><mml:math id="M112" display="inline"><mml:mi>G</mml:mi></mml:math></inline-formula> brightness) in
the image, but the difference in the <inline-formula><mml:math id="M113" display="inline"><mml:mi>Y</mml:mi></mml:math></inline-formula> characteristics according to the
cloud cover during the day was not large; thus, their importance was relatively low. Although time information and SZA can provide information
such as daytime, nighttime, and sunset/sunrise images, they have the lowest
importance because they do not have statistical characteristics that can be
used to directly calculate cloud cover. The importance of these data features may vary depending on the camera's sensor (Kazantzidis et al., 2012).</p>
</sec>
<?pagebreak page6704?><sec id="Ch1.S4.SS2">
  <label>4.2</label><title>Test set results for SVR model</title>
      <p id="d1e2810">Figure 6 shows the total cases and seasonal scatter plots of the DROM cloud
cover and the ACOS cloud cover prediction calculated from the SVR model using the test set. In the Korean Peninsula, the winter cloud cover is sparse (<inline-formula><mml:math id="M114" display="inline"><mml:mrow><mml:mo>&lt;</mml:mo><mml:mn mathvariant="normal">5</mml:mn></mml:mrow></mml:math></inline-formula> tenths) as the weather is generally clear because of the Siberian air mass. In summer, the rainy season is concentrated under the influence of the Yangtze River and Pacific air masses, and the cloud cover is dense (<inline-formula><mml:math id="M115" display="inline"><mml:mrow><mml:mo>&gt;</mml:mo><mml:mn mathvariant="normal">5</mml:mn></mml:mrow></mml:math></inline-formula> tenths) until fall because of typhoons (Kim and Lee, 2018; Kim et al., 2020a). Furthermore, the Korean Peninsula experiences a westerly wind, cumulus heat generated in the western sea moves inland, and the cloud cover changes rapidly and continuously (Kim et al., 2021). The cloud cover distributions calculated for all test set cases exhibited good agreement with the observed cloud cover, with accuracy, recall, and precision of 0.88, 0.42, and 0.52, respectively. Further, the bias, RMSE, and <inline-formula><mml:math id="M116" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula> were 0.04 tenths, 1.45 tenths, and 0.93, respectively. In fall, the bias, RMSE, and <inline-formula><mml:math id="M117" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula> were <inline-formula><mml:math id="M118" display="inline"><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">0.12</mml:mn></mml:mrow></mml:math></inline-formula> tenths, 1.30 tenths, and 0.95, respectively, indicating that the difference between the observed and calculated cloud cover was small. In winter and summer, the RMSE was larger and <inline-formula><mml:math id="M119" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula> was lower than in the other seasons. This is because the cloud cover calculation error is large at sunrise and sunset (100<inline-formula><mml:math id="M120" display="inline"><mml:msup><mml:mi/><mml:mo>∘</mml:mo></mml:msup></mml:math></inline-formula> <inline-formula><mml:math id="M121" display="inline"><mml:mo>≥</mml:mo></mml:math></inline-formula> SZA <inline-formula><mml:math id="M122" display="inline"><mml:mo>&gt;</mml:mo></mml:math></inline-formula> 80<inline-formula><mml:math id="M123" display="inline"><mml:msup><mml:mi/><mml:mo>∘</mml:mo></mml:msup></mml:math></inline-formula>), i.e., where daytime (SZA <inline-formula><mml:math id="M124" display="inline"><mml:mo>≤</mml:mo></mml:math></inline-formula> 80<inline-formula><mml:math id="M125" display="inline"><mml:msup><mml:mi/><mml:mo>∘</mml:mo></mml:msup></mml:math></inline-formula>) and nighttime (SZA <inline-formula><mml:math id="M126" display="inline"><mml:mo>&gt;</mml:mo></mml:math></inline-formula> 100<inline-formula><mml:math id="M127" display="inline"><mml:msup><mml:mi/><mml:mo>∘</mml:mo></mml:msup></mml:math></inline-formula>) intersect (Lalonde et al., 2010; Alonso et al., 2014; Kim and Cha, 2020).</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F6" specific-use="star"><?xmltex \currentcnt{6}?><?xmltex \def\figurename{Figure}?><label>Figure 6</label><caption><p id="d1e2932">Scatter plots of total <bold>(a)</bold> and seasonal <bold>(b–e)</bold> cloud cover based on observed (DROM) and calculated (ACOS) cloud cover for the test set. Parentheses values are the number of observations for each cloud cover in DROM.</p></caption>
          <?xmltex \igopts{width=341.433071pt}?><graphic xlink:href="https://amt.copernicus.org/articles/14/6695/2021/amt-14-6695-2021-f06.png"/>

        </fig>

      <?xmltex \floatpos{h!}?><fig id="Ch1.F7"><?xmltex \currentcnt{7}?><?xmltex \def\figurename{Figure}?><label>Figure 7</label><caption><p id="d1e2949">Daily mean time series of observed (DROM) and calculated (ACOS) cloud cover for the test set.</p></caption>
          <?xmltex \igopts{width=236.157874pt}?><graphic xlink:href="https://amt.copernicus.org/articles/14/6695/2021/amt-14-6695-2021-f07.png"/>

          <?xmltex \hack{\vspace*{2mm}}?>
        </fig>

      <p id="d1e2961">For the test set daytime cases, the bias, RMSE, and <inline-formula><mml:math id="M128" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula> were 0.10 tenths, 1.20 tenths, and 0.95, respectively, and 0.08 tenths, 1.59 tenths, and 0.93,
respectively, for the night data. However, for sunrise and sunset, these
values were <inline-formula><mml:math id="M129" display="inline"><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">0.22</mml:mn></mml:mrow></mml:math></inline-formula> tenths, 1.71 tenths, and 0.90, respectively. Relatively,
the bias and RMSE were large and <inline-formula><mml:math id="M130" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula> was low. In spring and autumn, sunrise
and sunset images were learned at similar times (sunrise: 06:00–07:00 LST,
sunset: 18:00–19:00 LST); however, differences between the winter (sunrise:
07:00–08:00 LST, sunset: 17:00–18:00 LST) and summer (sunrise: 05:00–06:00 LST, sunset: 19:00–20:00 LST) results are apparent because sunrise and sunset occurred late or early and exhibited different features from the data features learned for those times (Liu et al., 2014; Li et al., 2019). That is, owing to the sunrise/sunset glow, high cloud cover calculation errors are obtained at sunrise/sunset, when it is difficult to distinguish between the sky and clouds because of the reddish sky on a clear day and the bluish cloud on a cloudy day (Kim et al., 2021). Therefore, for the test set, the bias, RMSE, and <inline-formula><mml:math id="M131" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula> for sunrise and sunset in spring and autumn were <inline-formula><mml:math id="M132" display="inline"><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">0.24</mml:mn></mml:mrow></mml:math></inline-formula> tenths, 1.46 tenths, and 0.93, respectively. However, in winter and summer, the bias, RMSE, and <inline-formula><mml:math id="M133" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula> were <inline-formula><mml:math id="M134" display="inline"><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">0.21</mml:mn></mml:mrow></mml:math></inline-formula> tenths, 1.93 tenths, and 0.86, respectively. Nevertheless, the results of this study surpass those of Kim et al. (2016) for daytime (08:00–17:00 LST; bias: <inline-formula><mml:math id="M135" display="inline"><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">0.36</mml:mn></mml:mrow></mml:math></inline-formula> tenths, RMSE: 2.12 tenths, <inline-formula><mml:math id="M136" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula>: 0.87) and Kim and Cha (2020) for nighttime (19:00–06:00 LST; bias: <inline-formula><mml:math id="M137" display="inline"><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">0.28</mml:mn></mml:mrow></mml:math></inline-formula> tenths, RMSE: 1.78 tenths, <inline-formula><mml:math id="M138" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula>: 0.91) cases. Shields et al. (2019) employed different day and night cloud cover calculation algorithms. In that approach, cloud cover calculation errors may occur at sunrise and sunset. Therefore, if a day and night continuous cloud cover calculation algorithm is considered, the calculation error for this discontinuous time period should be reduced (Huo and Lu, 2009; Li et al., 2019). Figure 7 shows the daily mean cloud cover results based on the observed and calculated cloud cover for the test set. For the observed and calculated cloud cover, a bias of 0.03 tenths, RMSE of 0.92 tenths, and <inline-formula><mml:math id="M139" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula> of 0.96 were obtained. The coefficient of determination (<inline-formula><mml:math id="M140" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula>) was 0.92, and the result calculated from the SVR model constructed in this study explained approximately 92 % of the observed data in the test set.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F8" specific-use="star"><?xmltex \currentcnt{8}?><?xmltex \def\figurename{Figure}?><label>Figure 8</label><caption><p id="d1e3078">Relative frequency distributions of differences between observed (DROM) and calculated (ACOS) cloud cover by season and time for the test set.</p></caption>
          <?xmltex \igopts{width=341.433071pt}?><graphic xlink:href="https://amt.copernicus.org/articles/14/6695/2021/amt-14-6695-2021-f08.png"/>

        </fig>

      <p id="d1e3087">Figure 8 shows the frequency distribution of the differences between ACOS and DROM by season and time. In this frequency distribution, the higher the 0-tenth frequency, the higher the agreement between the observed and calculated cloud cover. The highest 0-tenth frequency was obtained in winter
(46.05 %) and the lowest in spring (35.58 %), but 41.69 % agreement was obtained for all seasons. Conditioned on the time of day, high agreement of approximately 44 % was obtained for both daytime and nighttime, but the
lowest agreement (30.49 %) was obtained for sunrise and sunset. Previous
studies obtained a difference of approximately 2 tenths from the observed
cloud cover for the cloud cover calculated based on the ground-based imager
data (Kazantzidis et al., 2012; Kim et al., 2016; Kim and Cha, 2020; Wang et al., 2021). When a difference of up to 2 tenths was allowed between the observed and calculated cloud cover, the agreement was 90.95 %, as detailed in Table 4. When a difference up to 1 tenth between both cloud cover results was allowed for all cases, the agreement was 79.05 %. When the difference was within 2 tenths, high agreement of 86.59 % to 94.41 % by season and by time<?pagebreak page6706?> was obtained. These results reveal greater agreement than those obtained by Cazorla et al. (2008), Kreuter et al. (2009), Kazantzidis et al. (2012), Kim et al. (2016), Krinitskiy and Sinitsyn (2016), Fa et al. (2019), Kim and Cha (2020), Xie et al. (2020), and Wang et al. (2021). In those works, 80 %–94 % agreement was achieved when the allowed difference between the observed and calculated cloud cover was 2 oktas (2.5 tenths) or 2 tenths for day, night, and day and night cases.</p>

<?xmltex \floatpos{h!}?><table-wrap id="Ch1.T4" specific-use="star"><?xmltex \currentcnt{4}?><label>Table 4</label><caption><p id="d1e3093">Concordance frequency (%) according to the difference between the observed (DROM) and calculated (ACOS) cloud cover for the test
set.</p></caption><oasis:table frame="topbot"><oasis:tgroup cols="9">
     <oasis:colspec colnum="1" colname="col1" align="left"/>
     <oasis:colspec colnum="2" colname="col2" align="center"/>
     <oasis:colspec colnum="3" colname="col3" align="center"/>
     <oasis:colspec colnum="4" colname="col4" align="center"/>
     <oasis:colspec colnum="5" colname="col5" align="center"/>
     <oasis:colspec colnum="6" colname="col6" align="center"/>
     <oasis:colspec colnum="7" colname="col7" align="center"/>
     <oasis:colspec colnum="8" colname="col8" align="center"/>
     <oasis:colspec colnum="9" colname="col9" align="center"/>
     <oasis:thead>
       <oasis:row>
         <oasis:entry colname="col1">Difference</oasis:entry>
         <oasis:entry colname="col2">Winter</oasis:entry>
         <oasis:entry colname="col3">Spring</oasis:entry>
         <oasis:entry colname="col4">Summer</oasis:entry>
         <oasis:entry colname="col5">Fall</oasis:entry>
         <oasis:entry colname="col6">Annual</oasis:entry>
         <oasis:entry colname="col7">Daytime</oasis:entry>
         <oasis:entry colname="col8">Nighttime</oasis:entry>
         <oasis:entry colname="col9">Sunrise/</oasis:entry>
       </oasis:row>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2"/>
         <oasis:entry colname="col3"/>
         <oasis:entry colname="col4"/>
         <oasis:entry colname="col5"/>
         <oasis:entry colname="col6"/>
         <oasis:entry colname="col7"/>
         <oasis:entry colname="col8"/>
         <oasis:entry colname="col9">sunset</oasis:entry>
       </oasis:row>
     </oasis:thead>
     <oasis:tbody>
       <oasis:row>
         <oasis:entry colname="col1"><inline-formula><mml:math id="M141" display="inline"><mml:mrow><mml:mo>±</mml:mo><mml:mn mathvariant="normal">0</mml:mn></mml:mrow></mml:math></inline-formula> tenths</oasis:entry>
         <oasis:entry colname="col2">46.05</oasis:entry>
         <oasis:entry colname="col3">35.58</oasis:entry>
         <oasis:entry colname="col4">39.84</oasis:entry>
         <oasis:entry colname="col5">45.48</oasis:entry>
         <oasis:entry colname="col6">41.69</oasis:entry>
         <oasis:entry colname="col7">43.96</oasis:entry>
         <oasis:entry colname="col8">43.88</oasis:entry>
         <oasis:entry colname="col9">30.49</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"><inline-formula><mml:math id="M142" display="inline"><mml:mrow><mml:mo>±</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow></mml:math></inline-formula> tenth</oasis:entry>
         <oasis:entry colname="col2">78.53</oasis:entry>
         <oasis:entry colname="col3">76.01</oasis:entry>
         <oasis:entry colname="col4">78.36</oasis:entry>
         <oasis:entry colname="col5">83.24</oasis:entry>
         <oasis:entry colname="col6">79.05</oasis:entry>
         <oasis:entry colname="col7">84.29</oasis:entry>
         <oasis:entry colname="col8">77.45</oasis:entry>
         <oasis:entry colname="col9">68.70</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"><inline-formula><mml:math id="M143" display="inline"><mml:mrow><mml:mo>±</mml:mo><mml:mn mathvariant="normal">2</mml:mn></mml:mrow></mml:math></inline-formula> tenths</oasis:entry>
         <oasis:entry colname="col2">90.11</oasis:entry>
         <oasis:entry colname="col3">88.68</oasis:entry>
         <oasis:entry colname="col4">90.50</oasis:entry>
         <oasis:entry colname="col5">94.41</oasis:entry>
         <oasis:entry colname="col6">90.95</oasis:entry>
         <oasis:entry colname="col7">93.66</oasis:entry>
         <oasis:entry colname="col8">89.69</oasis:entry>
         <oasis:entry colname="col9">86.59</oasis:entry>
       </oasis:row>
     </oasis:tbody>
   </oasis:tgroup></oasis:table></table-wrap>

</sec>
</sec>
<sec id="Ch1.S5" sec-type="conclusions">
  <label>5</label><title>Conclusions</title>
      <p id="d1e3310">In this study, data features of images captured using ACOS, a camera-based
imager on the ground, were used in conjunction with machine learning methods to continuously calculate cloud cover for 24 h, at day and night. The data
features of the images used as the machine learning input data were the
mean, mode, frequency of mode, skewness, kurtosis, and quantile (<inline-formula><mml:math id="M144" display="inline"><mml:mrow><mml:msub><mml:mi>Q</mml:mi><mml:mn mathvariant="normal">0</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula>–<inline-formula><mml:math id="M145" display="inline"><mml:mrow><mml:msub><mml:mi>Q</mml:mi><mml:mn mathvariant="normal">4</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula>) of the RBR, BRD, and <inline-formula><mml:math id="M146" display="inline"><mml:mi>Y</mml:mi></mml:math></inline-formula> frequency distributions, respectively, along with the Julian day, hour, and SZA. The RBR, BRD, and <inline-formula><mml:math id="M147" display="inline"><mml:mi>Y</mml:mi></mml:math></inline-formula> data features were calculated through pre-processing using the methods described by Kim and Cha (2020) (masking removal and distortion correction). These features indicate the sky and cloud colors depending on the light scattering characteristics in the day and night, along with the presence or absence of clouds and the position of the sun (Heinle et al., 2010; Blazek and Pata, 2015; Li et al., 2019). The collected image data (100 %) were composed of training (50 %), validation (30 %), and test (20 %) sets and were used for optimization of the models produced by the machine learning methods, comparative analysis of the prediction results of each machine learning method, and verification of the predicted cloud cover. In this study, the SVR, RF, GBM, kNN, ANN, and MLR supervised machine learning methods were used. Among these methods, the SVR model exhibited the best prediction performance, with accuracy, recall, and precision of 0.94, 0.70, and 0.76, respectively. The cloud cover calculation results produced by the SVR on the test set had a bias of 0.04 tenths, RMSE of 1.45 tenths, and <inline-formula><mml:math id="M148" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula> of 0.93. With respect to this calculation result, when a difference of 2 tenths from the observed cloud cover was allowed, the agreement was 41.69 %, 79.05 %, and 90.95 % for 0-, 1-, and 2-tenth difference, respectively.</p>
      <p id="d1e3356">Using the image data features and machine learning methods (best: SVR, worst: MLR) considered in this study, high-accuracy cloud cover calculation can be expected; further, this approach is suitable for nowcasting. The cloud information obtained from such cloud detection and cloud cover calculation makes it possible to calculate the physical properties of various clouds (Wang et al., 2016; Román et al., 2017; Ye et al., 2017; J. Zhang et al., 2018). In other words, it is possible to calculate cloud-based height and cloud motion vector through the geometric and kinematic analysis of continuous images using single or multiple cameras (Nguyen and Kleissl, 2014), which can be used for cloud type classification according to cloud cover, cloud-based height, and cloud color feature (Heinle et al., 2010; Ghonima et al., 2012). Ground-based observation of clouds using a camera-based imager, accompanied by cloud characteristic calculation, is an
economical method that can replace manned observations at synoptic observatories with automated (unmanned) observations. In addition, objective
and low-uncertainty cloud observation is expected to be possible through
widespread distribution of instruments such as those used in this study to
unmanned as well as manned observatories. Therefore, active research and
development of imager-based cloud observation instruments is merited.</p>
</sec>

      
      </body>
    <back><notes notes-type="codeavailability"><title>Code availability</title>

      <p id="d1e3363">The code for this paper is available from the corresponding author.</p>
  </notes><notes notes-type="sampleavailability"><title>Sample availability</title>

      <p id="d1e3369">The samples for this paper are available from the corresponding author.</p>
  </notes><notes notes-type="authorcontribution"><title>Author contributions</title>

      <p id="d1e3375">BYK carried out this study and the analysis. The results were discussed with JWC and KHC. BYK developed the<?pagebreak page6707?> machine learning model code and performed the simulations and visualizations. The manuscript was mainly written by BYK with contributions by JWC and KHC.</p>
  </notes><notes notes-type="competinginterests"><title>Competing interests</title>

      <p id="d1e3381">The contact author has declared that neither they nor their co-authors have any competing interests.</p>
  </notes><notes notes-type="disclaimer"><title>Disclaimer</title>

      <p id="d1e3387">Publisher's note: Copernicus Publications remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.</p>
  </notes><ack><title>Acknowledgements</title><p id="d1e3393">This work was funded by the Korea Meteorological Administration Research and Development Program, Development of Application Technology on Atmospheric Research Aircraft, under grant (KMA2018-00222).</p></ack><notes notes-type="financialsupport"><title>Financial support</title>

      <p id="d1e3398">This research has been supported by the National Institute of Meteorological Sciences (grant no. KMA2018-00222).</p>
  </notes><notes notes-type="reviewstatement"><title>Review statement</title>

      <p id="d1e3405">This paper was edited by Dmitry Efremenko and reviewed by two anonymous referees.</p>
  </notes><ref-list>
    <title>References</title>

      <ref id="bib1.bib1"><label>1</label><?label 1?><mixed-citation>Al Banna, M. H., Taher, K. A., Kaiser, M. S., Mahmud, M., Rahman, M. S.,
Hosen, A. S., and Cho, G. H.: Application of artificial intelligence in
predicting earthquakes: state-of-the-art and future challenges, IEEE Access.,
8, 192880–192923, <ext-link xlink:href="https://doi.org/10.1109/ACCESS.2020.3029859" ext-link-type="DOI">10.1109/ACCESS.2020.3029859</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bib2"><label>2</label><?label 1?><mixed-citation>Al-lahham, A., Theeb, O., Elalem, K., Alshawi, T., and Alshebeili, S.: Sky
Imager-Based Forecast of Solar Irradiance Using Machine Learning, Electronics, 9, 1700, <ext-link xlink:href="https://doi.org/10.3390/electronics9101700" ext-link-type="DOI">10.3390/electronics9101700</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bib3"><label>3</label><?label 1?><mixed-citation>Alonso, J., Batlles, F. J., López, G., and Ternero, A.: Sky camera imagery processing based on a sky classification using radiometric data, Energy, 68, 599–608, <ext-link xlink:href="https://doi.org/10.1016/j.energy.2014.02.035" ext-link-type="DOI">10.1016/j.energy.2014.02.035</ext-link>, 2014.</mixed-citation></ref>
      <ref id="bib1.bib4"><label>4</label><?label 1?><mixed-citation>Alonso-Montesinos, J.: Real-Time Automatic Cloud Detection Using a Low-Cost
Sky Camera, Remote Sens., 12, 1382, <ext-link xlink:href="https://doi.org/10.3390/rs12091382" ext-link-type="DOI">10.3390/rs12091382</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bib5"><label>5</label><?label 1?><mixed-citation>Azhar, M. A. D. M., Hamid, N. S. A., Kamil, W. M. A. W. M., and Mohamad, N. S.: Daytime Cloud Detection Method Using the All-Sky Imager over PERMATApintar Observatory, Universe, 7, 41, <ext-link xlink:href="https://doi.org/10.3390/universe7020041" ext-link-type="DOI">10.3390/universe7020041</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bib6"><label>6</label><?label 1?><mixed-citation>
Bergstra, J. and Bengio, Y.: Random search for hyper-parameter optimization, J. Mach. Learn. Res., 13, 281–305, 2012.</mixed-citation></ref>
      <ref id="bib1.bib7"><label>7</label><?label 1?><mixed-citation>Blazek, M. and Pata, P.: Colour transformations and <inline-formula><mml:math id="M149" display="inline"><mml:mi>K</mml:mi></mml:math></inline-formula>-means segmentation for automatic cloud detection, Meteorol. Z., 24, 503–509, <ext-link xlink:href="https://doi.org/10.1127/metz/2015/0656" ext-link-type="DOI">10.1127/metz/2015/0656</ext-link>, 2015.</mixed-citation></ref>
      <ref id="bib1.bib8"><label>8</label><?label 1?><mixed-citation>Boers, R., De Haij, M. J., Wauben, W. M. F., Baltink, H. K., Van Ulft, L.
H., Savenije, M., and Long, C. N.: Optimized fractional cloudiness determination from five ground-based remote sensing techniques, J. Geophys. Res., 115, D24116, <ext-link xlink:href="https://doi.org/10.1029/2010JD014661" ext-link-type="DOI">10.1029/2010JD014661</ext-link>, 2010.</mixed-citation></ref>
      <ref id="bib1.bib9"><label>9</label><?label 1?><mixed-citation>Calbó, J., Long, C. N., González, J. A., Augustine, J., and McComiskey, A.: The thin border between cloud and aerosol: Sensitivity of several ground based observation techniques, Atmos. Res., 196, 248–260,
<ext-link xlink:href="https://doi.org/10.1016/j.atmosres.2017.06.010" ext-link-type="DOI">10.1016/j.atmosres.2017.06.010</ext-link>, 2017.</mixed-citation></ref>
      <ref id="bib1.bib10"><label>10</label><?label 1?><mixed-citation>Cazorla, A., Olmo, F. J., Alados-Arboledas, L.: Development of a sky imager for cloud cover assessment, J. Opt. Soc. Am. A, 25, 29–39, <ext-link xlink:href="https://doi.org/10.1364/JOSAA.25.000029" ext-link-type="DOI">10.1364/JOSAA.25.000029</ext-link>, 2008.</mixed-citation></ref>
      <ref id="bib1.bib11"><label>11</label><?label 1?><mixed-citation>Cazorla, A., Husillos, C., Antón, M., and Alados-Arboledas, L.: Multi-exposure adaptive threshold technique for cloud detection with sky imagers, Sol. Energy, 114, 268–277, <ext-link xlink:href="https://doi.org/10.1016/j.solener.2015.02.006" ext-link-type="DOI">10.1016/j.solener.2015.02.006</ext-link>,
2015.</mixed-citation></ref>
      <ref id="bib1.bib12"><label>12</label><?label 1?><mixed-citation>Chauvin, R., Nou, J., Thil, S., and Grieu, S.: Modelling the clear-sky intensity distribution using a sky imager, Sol. Energy, 119, 1–17,
<ext-link xlink:href="https://doi.org/10.1016/j.solener.2015.06.026" ext-link-type="DOI">10.1016/j.solener.2015.06.026</ext-link>, 2015.</mixed-citation></ref>
      <ref id="bib1.bib13"><label>13</label><?label 1?><mixed-citation>Çınar, Z. M., Abdussalam Nuhu, A., Zeeshan, Q., Korhan, O., Asmael, M., and Safaei, B.: Machine learning in predictive maintenance towards
sustainable smart manufacturing in industry 4.0, Sustainability, 12, 8211, <ext-link xlink:href="https://doi.org/10.3390/su12198211" ext-link-type="DOI">10.3390/su12198211</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bib14"><label>14</label><?label 1?><mixed-citation>Costa-Surós, M., Calbó, J., González, J. A., and Long, C. N.:
Comparing the cloud vertical structure derived from several methods based on
radiosonde profiles and ground-based remote sensing measurements, Atmos.
Meas. Tech., 7, 2757–2773, <ext-link xlink:href="https://doi.org/10.5194/amt-7-2757-2014" ext-link-type="DOI">10.5194/amt-7-2757-2014</ext-link>, 2014.</mixed-citation></ref>
      <ref id="bib1.bib15"><label>15</label><?label 1?><mixed-citation>
Dev, S., Savoy, F. M., Lee, Y. H., and Winkler, S.: Design of low-cost, compact and weather-proof whole sky imagers for High-Dynamic-Range captures, in: IGARSS 2015–2015 IEEE International Geoscience and Remote Sensing
Symposium, 26–31 July 2015, Milan, Italy, 5359–5362, 2015.</mixed-citation></ref>
      <ref id="bib1.bib16"><label>16</label><?label 1?><mixed-citation>Dev, S., Wen, B., Lee, Y. H., and Winkler, S.: Ground-based image analysis: A
tutorial on machine-learning techniques and applications, IEEE Geosci. Remote Sens. M., 4, 79–93, <ext-link xlink:href="https://doi.org/10.1109/MGRS.2015.2510448" ext-link-type="DOI">10.1109/MGRS.2015.2510448</ext-link>, 2016.</mixed-citation></ref>
      <?pagebreak page6708?><ref id="bib1.bib17"><label>17</label><?label 1?><mixed-citation>
Dev, S., Savoy, F. M., Lee, Y. H., and Winkler, S.: Nighttime sky/cloud image
segmentation, in: ICIP 2017–2017 IEEE International Conference on Image
Processing, 17–20 September 2017, Beijing, China, 345–349, 2017.</mixed-citation></ref>
      <ref id="bib1.bib18"><label>18</label><?label 1?><mixed-citation>Dev, S., Nautiyal, A., Lee, Y. H., and Winkler, S.: Cloudsegnet: A deep network for nychthemeron cloud image segmentation, IEEE Geosci. Remote Sens. Lett., 16, 1814–1818, <ext-link xlink:href="https://doi.org/10.1109/lgrs.2019.2912140" ext-link-type="DOI">10.1109/lgrs.2019.2912140</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bib19"><label>19</label><?label 1?><mixed-citation>Fa, T., Xie, W., Wang, Y., and Xia, Y.: Development of an all-sky imaging system for cloud cover assessment, Appl. Optics, 58, 5516–5524,
<ext-link xlink:href="https://doi.org/10.1364/AO.58.005516" ext-link-type="DOI">10.1364/AO.58.005516</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bib20"><label>20</label><?label 1?><mixed-citation>Friedman, J. H.: Greedy function approximation: A gradient boosting machine,
Ann. Stat., 29, 1189–1232, <ext-link xlink:href="https://doi.org/10.1214/aos/1013203451" ext-link-type="DOI">10.1214/aos/1013203451</ext-link>, 2001.</mixed-citation></ref>
      <ref id="bib1.bib21"><label>21</label><?label 1?><mixed-citation>Gani, W., Taleb, H., and Limam, M.: Support vector regression based residual
control charts, J. Appl. Stat., 37, 309–324, <ext-link xlink:href="https://doi.org/10.1080/02664760903002667" ext-link-type="DOI">10.1080/02664760903002667</ext-link>, 2010.</mixed-citation></ref>
      <ref id="bib1.bib22"><label>22</label><?label 1?><mixed-citation>Geyer, C. J.: Generalized linear models in R, R Reference Document, 1–23, available at: <uri>https://www.stat.umn.edu/geyer/5931/mle/glm.pdf</uri> (last access: 1 June 2021), 2003.</mixed-citation></ref>
      <ref id="bib1.bib23"><label>23</label><?label 1?><mixed-citation>Ghonima, M. S., Urquhart, B., Chow, C. W., Shields, J. E., Cazorla, A., and
Kleissl, J.: A method for cloud detection and opacity classification based
on ground based sky imagery, Atmos. Meas. Tech., 5, 2881–2892,
<ext-link xlink:href="https://doi.org/10.5194/amt-5-2881-2012" ext-link-type="DOI">10.5194/amt-5-2881-2012</ext-link>, 2012.</mixed-citation></ref>
      <ref id="bib1.bib24"><label>24</label><?label 1?><mixed-citation>Greenwell, B., Boehmke, B., and Cunningham, J.: GBM Developers: Package `gbm', R Reference Document, 1–39, available at:
<uri>https://cran.r-project.org/web/packages/gbm/gbm.pdf</uri> (last access: 1 June 2021), 2020.</mixed-citation></ref>
      <ref id="bib1.bib25"><label>25</label><?label 1?><mixed-citation>Heinle, A., Macke, A., and Srivastav, A.: Automatic cloud classification of
whole sky images, Atmos. Meas. Tech., 3, 557–567, <ext-link xlink:href="https://doi.org/10.5194/amt-3-557-2010" ext-link-type="DOI">10.5194/amt-3-557-2010</ext-link>, 2010.</mixed-citation></ref>
      <ref id="bib1.bib26"><label>26</label><?label 1?><mixed-citation>Huo, J. and Lu, D.: Cloud determination of all-sky images under low-visibility conditions, J. Atmos. Ocean. Tech., 26, 2172–2181,
<ext-link xlink:href="https://doi.org/10.1175/2009JTECHA1324.1" ext-link-type="DOI">10.1175/2009JTECHA1324.1</ext-link>, 2009.</mixed-citation></ref>
      <ref id="bib1.bib27"><label>27</label><?label 1?><mixed-citation>Kazantzidis, A., Tzoumanikas, P., Bais, A. F., Fotopoulos, S., and Economou, G.: Cloud detection and classification with the use of whole-sky ground-based
images, Atmos. Res., 113, 80–88, <ext-link xlink:href="https://doi.org/10.1016/j.atmosres.2012.05.005" ext-link-type="DOI">10.1016/j.atmosres.2012.05.005</ext-link>, 2012.</mixed-citation></ref>
      <ref id="bib1.bib28"><label>28</label><?label 1?><mixed-citation>Kim, B. Y. and Cha, J. W.: Cloud Observation and Cloud Cover Calculation at
Nighttime Using the Automatic Cloud Observation System (ACOS) Package, Remote Sens., 12, 2314, <ext-link xlink:href="https://doi.org/10.3390/rs12142314" ext-link-type="DOI">10.3390/rs12142314</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bib29"><label>29</label><?label 1?><mixed-citation>Kim, B. Y. and Lee, K. T.: Radiation component calculation and energy budget
analysis for the Korean Peninsula region, Remote Sens., 10, 1147,
<ext-link xlink:href="https://doi.org/10.3390/rs10071147" ext-link-type="DOI">10.3390/rs10071147</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bib30"><label>30</label><?label 1?><mixed-citation>Kim, B. Y. and Lee, K. T.: Using the himawari-8 ahi multi-channel to improve
the calculation accuracy of outgoing longwave radiation at the top of the
atmosphere. Remote Sens., 11, 589, <ext-link xlink:href="https://doi.org/10.3390/rs11050589" ext-link-type="DOI">10.3390/rs11050589</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bib31"><label>31</label><?label 1?><mixed-citation>Kim, B. Y., Jee, J. B., Zo, I. S., and Lee, K. T.: Cloud cover retrieved from
skyviewer: A validation with human observations. Asia-Pac, J. Atmos. Sci., 52, 1–10, <ext-link xlink:href="https://doi.org/10.1007/s13143-015-0083-4" ext-link-type="DOI">10.1007/s13143-015-0083-4</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bib32"><label>32</label><?label 1?><mixed-citation>Kim, B. Y., Lee, K. T., Jee, J. B., and Zo, I. S.: Retrieval of outgoing longwave radiation at top-of-atmosphere using Himawari-8 AHI data, Remote
Sens. Environ., 204, 498–508, <ext-link xlink:href="https://doi.org/10.1016/j.rse.2017.10.006" ext-link-type="DOI">10.1016/j.rse.2017.10.006</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bib33"><label>33</label><?label 1?><mixed-citation>Kim, B. Y., Cha, J. W., Ko, A. R., Jung, W., and Ha, J. C.: Analysis of the
occurrence frequency of seedable clouds on the Korean Peninsula for precipitation enhancement experiments, Remote Sens., 12, 1487,
<ext-link xlink:href="https://doi.org/10.3390/rs12091487" ext-link-type="DOI">10.3390/rs12091487</ext-link>, 2020a.</mixed-citation></ref>
      <ref id="bib1.bib34"><label>34</label><?label 1?><mixed-citation>Kim, B. Y., Cha, J. W., Jung, W., and Ko, A. R.: Precipitation Enhancement
Experiments in Catchment Areas of Dams: Evaluation of Water Resource Augmentation and Economic Benefits, Remote Sens., 12, 3730, <ext-link xlink:href="https://doi.org/10.3390/rs12223730" ext-link-type="DOI">10.3390/rs12223730</ext-link>, 2020b.</mixed-citation></ref>
      <ref id="bib1.bib35"><label>35</label><?label 1?><mixed-citation>Kim, B. Y., Cha, J. W., Chang, K. H., and Lee, C.: Visibility Prediction over
South Korea Based on Random Forest, Atmosphere, 12, 552, <ext-link xlink:href="https://doi.org/10.3390/atmos12050552" ext-link-type="DOI">10.3390/atmos12050552</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bib36"><label>36</label><?label 1?><mixed-citation>Kreuter, A., Zangerl, M., Schwarzmann, M., and Blumthaler, M.: All-sky imaging: a simple, versatile system for atmospheric research, Appl. Optics, 48, 1091–1097, <ext-link xlink:href="https://doi.org/10.1364/AO.48.001091" ext-link-type="DOI">10.1364/AO.48.001091</ext-link>, 2009.</mixed-citation></ref>
      <ref id="bib1.bib37"><label>37</label><?label 1?><mixed-citation>Krinitskiy, M. A. and Sinitsyn, A. V.: Adaptive algorithm for cloud cover
estimation from all-sky images over the sea, Oceanology, 56, 315–319,
<ext-link xlink:href="https://doi.org/10.1134/S0001437016020132" ext-link-type="DOI">10.1134/S0001437016020132</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bib38"><label>38</label><?label 1?><mixed-citation>Kyba, C. C., Ruhtz, T., Fischer, J., and Hölker, F.: Red is the new black: how the colour of urban skyglow varies with cloud cover, Mon. Notic. Roy. Astron. Soc., 425, 701–708, <ext-link xlink:href="https://doi.org/10.1111/j.1365-2966.2012.21559.x" ext-link-type="DOI">10.1111/j.1365-2966.2012.21559.x</ext-link>, 2012.</mixed-citation></ref>
      <ref id="bib1.bib39"><label>39</label><?label 1?><mixed-citation>Lalonde, J. F., Narasimhan, S. G., and Efros, A. A.: What do the sun and the sky tell us about the camera?, Int. J. Comput. Vis., 88, 24–51,
<ext-link xlink:href="https://doi.org/10.1007/s11263-009-0291-4" ext-link-type="DOI">10.1007/s11263-009-0291-4</ext-link>, 2010.</mixed-citation></ref>
      <ref id="bib1.bib40"><label>40</label><?label 1?><mixed-citation>Lee, S. H., Kim, B. Y., Lee, K. T., Zo, I. S., Jung, H. S., and Rim, S. H.:
Retrieval of reflected shortwave radiation at the top of the atmosphere
using Himawari-8/AHI data, Remote Sens., 10, 213, <ext-link xlink:href="https://doi.org/10.3390/rs10020213" ext-link-type="DOI">10.3390/rs10020213</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bib41"><label>41</label><?label 1?><mixed-citation>Li, Q., Lu, W., and Yang, J.: A hybrid thresholding algorithm for cloud
detection on ground-based color images, J. Atmos. Ocean. Tech., 28, 1286–1296, <ext-link xlink:href="https://doi.org/10.1175/JTECH-D-11-00009.1" ext-link-type="DOI">10.1175/JTECH-D-11-00009.1</ext-link>, 2011.</mixed-citation></ref>
      <ref id="bib1.bib42"><label>42</label><?label 1?><mixed-citation>Li, X., Lu, Z., Zhou, Q., and Xu, Z.: A Cloud Detection Algorithm with Reduction of Sunlight Interference in Ground-Based Sky Images, Atmosphere, 10, 640, <ext-link xlink:href="https://doi.org/10.3390/atmos10110640" ext-link-type="DOI">10.3390/atmos10110640</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bib43"><label>43</label><?label 1?><mixed-citation>Liu, S., Zhang, L., Zhang, Z., Wang, C., and Xiao, B.: Automatic cloud detection for all-sky images using superpixel segmentation, IEEE Geosci. Remote Sens. Lett., 12, 354–358, <ext-link xlink:href="https://doi.org/10.1109/LGRS.2014.2341291" ext-link-type="DOI">10.1109/LGRS.2014.2341291</ext-link>, 2014.</mixed-citation></ref>
      <ref id="bib1.bib44"><label>44</label><?label 1?><mixed-citation>Long, C. N., Sabburg, J. M., Calbó, J., and Pagès, D.: Retrieving cloud characteristics from ground-based daytime color all-sky images, J. Atmos. Ocean. Tech., 23, 633–652, <ext-link xlink:href="https://doi.org/10.1175/JTECH1875.1" ext-link-type="DOI">10.1175/JTECH1875.1</ext-link>, 2006.</mixed-citation></ref>
      <ref id="bib1.bib45"><label>45</label><?label 1?><mixed-citation>Lothon, M., Barnéoud, P., Gabella, O., Lohou, F., Derrien, S., Rondi, S., Chiriaco, M., Bastin, S., Dupont, J. C., Haeffelin, M., Badosa, J., Pascal, N., and Montoux, N.: ELIFAN, an algorithm for the estimation of cloud cover from sky imagers, Atmos. Meas. Tech., 12, 5519–5534, <ext-link xlink:href="https://doi.org/10.5194/amt-12-5519-2019" ext-link-type="DOI">10.5194/amt-12-5519-2019</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bib46"><label>46</label><?label 1?><mixed-citation>Mantelli Neto, S. L., von Wangenheim, A., Pereira, E. B., and Comunello, E.: The use of Euclidean geometric distance on RGB color space for the
classification of sky and cloud patterns, J. Atmos. Ocean. Tech., 27, 1504–1517, <ext-link xlink:href="https://doi.org/10.1175/2010JTECHA1353.1" ext-link-type="DOI">10.1175/2010JTECHA1353.1</ext-link>, 2010.</mixed-citation></ref>
      <?pagebreak page6709?><ref id="bib1.bib47"><label>47</label><?label 1?><mixed-citation>Martínez, F., Frías, M. P., Charte, F., and Rivera, A. J.: Time Series Forecasting with KNN in R: the tsfknn Package, R J., 11, 229–242, <ext-link xlink:href="https://doi.org/10.32614/RJ-2019-004" ext-link-type="DOI">10.32614/RJ-2019-004</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bib48"><label>48</label><?label 1?><mixed-citation>Meyer, D.: Support vector machines. The Interface to libsvm in package e1071, R News, 23–26, available at: <uri>https://cran.r-project.org/web/packages/e1071/vignettes/svmdoc.pdf</uri> (last access: 1 June 2021), 2001.</mixed-citation></ref>
      <ref id="bib1.bib49"><label>49</label><?label 1?><mixed-citation>Meyer, D., Dimitriadou, E., Hornik, K., Weingessel, A., Leisch, F., Chang, C. C., and Lin, C. C.: Package `e1071', R Reference Document, 1–66, available at: <uri>https://cran.r-project.org/web/packages/e1071/e1071.pdf</uri>, last access: 1 June 2021.</mixed-citation></ref>
      <ref id="bib1.bib50"><label>50</label><?label 1?><mixed-citation>Nguyen, D. A. and Kleissl, J.: Stereographic methods for cloud base height
determination using two sky imagers, Sol. Energy, 107, 495–509,
<ext-link xlink:href="https://doi.org/10.1016/j.solener.2014.05.005" ext-link-type="DOI">10.1016/j.solener.2014.05.005</ext-link>, 2014.</mixed-citation></ref>
      <ref id="bib1.bib51"><label>51</label><?label 1?><mixed-citation>
Olive, D. J.: Multiple linear regression, in: Linear regression. Springer,
Cham, 17–83, 2017.</mixed-citation></ref>
      <ref id="bib1.bib52"><label>52</label><?label 1?><mixed-citation>Peng, Z., Yu, D., Huang, D., Heiser, J., Yoo, S., and Kalb, P.: 3D cloud
detection and tracking system for solar forecast using multiple sky imagers,
Sol. Energy, 118, 496–519, <ext-link xlink:href="https://doi.org/10.1016/j.solener.2015.05.037" ext-link-type="DOI">10.1016/j.solener.2015.05.037</ext-link>, 2015.</mixed-citation></ref>
      <ref id="bib1.bib53"><label>53</label><?label 1?><mixed-citation>
Ridgeway, G.: Generalized Boosted Models: A guide to the gbm package, Update,
1, 1–15, 2020.</mixed-citation></ref>
      <ref id="bib1.bib54"><label>54</label><?label 1?><mixed-citation>Ripley, B. and Venables, W.: Package `class', R Reference Document, 1–19,
available at: <uri>https://cran.r-project.org/web/packages/class/class.pdf</uri> (last access: 1 June 2021), 2021a.</mixed-citation></ref>
      <ref id="bib1.bib55"><label>55</label><?label 1?><mixed-citation>Ripley, B. and Venables, W.: Package `nnet', R Reference Document, 1–11,
available at: <uri>https://cran.r-project.org/web/packages/nnet/nnet.pdf</uri> (last access: 1 June 2021), 2021b.</mixed-citation></ref>
      <ref id="bib1.bib56"><label>56</label><?label 1?><mixed-citation>Román, R., Cazorla, A., Toledano, C., Olmo, F. J., Cachorro, V. E., de Frutos, A., and Alados-Arboledas, L.: Cloud cover detection combining high
dynamic range sky images and ceilometer measurements, Atmos. Res., 196,
224–236, <ext-link xlink:href="https://doi.org/10.1016/j.atmosres.2017.06.006" ext-link-type="DOI">10.1016/j.atmosres.2017.06.006</ext-link>, 2017.</mixed-citation></ref>
      <ref id="bib1.bib57"><label>57</label><?label 1?><mixed-citation>
Rosa, J. P., Guerra, D. J., Horta, N. C., Martins, R. M., and Lourenço, N. C.: Overview of Artificial Neural Networks, in: Using Artificial Neural
Networks for Analog Integrated Circuit Design Automation, Springer, Cham, 21–44, 2020.</mixed-citation></ref>
      <ref id="bib1.bib58"><label>58</label><?label 1?><mixed-citation>
Sazzad, T. S., Islam, S., Mamun, M. M. R. K., and Hasan, M. Z.: Establishment of an efficient color model from existing models for better gamma encoding in image processing, Int. J. Image Process., 7, 90–100, 2013.</mixed-citation></ref>
      <ref id="bib1.bib59"><label>59</label><?label 1?><mixed-citation>Shi, C., Zhou, Y., Qiu, B., He, J., Ding, M., and Wei, S.: Diurnal and nocturnal cloud segmentation of all-sky imager (ASI) images using enhancement fully convolutional networks. Atmos. Meas. Tech., 12, 4713–4724,
<ext-link xlink:href="https://doi.org/10.5194/amt-12-4713-2019" ext-link-type="DOI">10.5194/amt-12-4713-2019</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bib60"><label>60</label><?label 1?><mixed-citation>Shi, C., Zhou, Y., and Qiu, B.: CloudU-Netv2: A Cloud Segmentation Method for
Ground-Based Cloud Images Based on Deep Learning, Neural Process. Lett., 53, 2715–2728, <ext-link xlink:href="https://doi.org/10.1007/s11063-021-10457-2" ext-link-type="DOI">10.1007/s11063-021-10457-2</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bib61"><label>61</label><?label 1?><mixed-citation>Shields, J. E., Karr, M. E., Johnson, R. W., and Burden, A. R.: Day/night whole sky imagers for 24-h cloud and sky assessment: history and overview, Appl. Optics, 52, 1605–1616, <ext-link xlink:href="https://doi.org/10.1364/AO.52.001605" ext-link-type="DOI">10.1364/AO.52.001605</ext-link>, 2013.</mixed-citation></ref>
      <ref id="bib1.bib62"><label>62</label><?label 1?><mixed-citation>Shields, J. E., Burden, A. R., and Karr, M. E.: Atmospheric cloud algorithms for day/night whole sky imagers, Appl. Optics, 58, 7050–7062,
<ext-link xlink:href="https://doi.org/10.1364/AO.58.007050" ext-link-type="DOI">10.1364/AO.58.007050</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bib63"><label>63</label><?label 1?><mixed-citation>Shimoji, N., Aoyama, R., and Hasegawa, W.: Spatial variability of correlated
color temperature of lightning channels, Results Phys., 6, 161–162,
<ext-link xlink:href="https://doi.org/10.1016/j.rinp.2016.03.004" ext-link-type="DOI">10.1016/j.rinp.2016.03.004</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bib64"><label>64</label><?label 1?><mixed-citation>Shin, J. Y., Kim, B. Y., Park, J., Kim, K. R., and Cha, J. W.: Prediction of
Leaf Wetness Duration Using Geostationary Satellite Observations and Machine
Learning Algorithms, Remote Sens., 12, 3076, <ext-link xlink:href="https://doi.org/10.3390/rs12183076" ext-link-type="DOI">10.3390/rs12183076</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bib65"><label>65</label><?label 1?><mixed-citation>Singh, A., Kotiyal, V., Sharma, S., Nagar, J., and Lee, C. C.: A machine
learning approach to predict the average localization error with applications to wireless sensor networks, IEEE Access., 8, 208253–208263,
<ext-link xlink:href="https://doi.org/10.1109/ACCESS.2020.3038645" ext-link-type="DOI">10.1109/ACCESS.2020.3038645</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bib66"><label>66</label><?label 1?><mixed-citation>Sunil, S., Padmakumari, B., Pandithurai, G., Patil, R. D., and Naidu, C. V.:
Diurnal (24 h) cycle and seasonal variability of cloud fraction retrieved
from a Whole Sky Imager over a complex terrain in the Western Ghats and
comparison with MODIS, Atmos. Res., 248, 105180, <ext-link xlink:href="https://doi.org/10.1016/j.atmosres.2020.105180" ext-link-type="DOI">10.1016/j.atmosres.2020.105180</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bib67"><label>67</label><?label 1?><mixed-citation>Taghizadeh-Mehrjardi, R., Neupane, R., Sood, K., and Kumar, S.: Artificial bee colony feature selection algorithm combined with machine learning algorithms to predict vertical and lateral distribution of soil organic matter in South Dakota, USA, Carbon Manage., 8, 277–291,
<ext-link xlink:href="https://doi.org/10.1080/17583004.2017.1330593" ext-link-type="DOI">10.1080/17583004.2017.1330593</ext-link>, 2017.</mixed-citation></ref>
      <ref id="bib1.bib68"><label>68</label><?label 1?><mixed-citation>Valentín, L., Peña-Cruz, M. I., Moctezuma, D., Peña-Martínez, C. M., Pineda-Arellano, C. A., and Díaz-Ponce, A.: Towards the Development of a Low-Cost Irradiance Nowcasting Sky Imager, Appl. Sci., 9, 1131, <ext-link xlink:href="https://doi.org/10.3390/app9061131" ext-link-type="DOI">10.3390/app9061131</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bib69"><label>69</label><?label 1?><mixed-citation>Wang, G., Kurtz, B., and Kleissl, J.: Cloud base height from sky imager and
cloud speed sensor, Sol. Energy, 131, 208–221, <ext-link xlink:href="https://doi.org/10.1016/j.solener.2016.02.027" ext-link-type="DOI">10.1016/j.solener.2016.02.027</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bib70"><label>70</label><?label 1?><mixed-citation>Wang, Y., Liu, D., Xie, W., Yang, M., Gao, Z., Ling, X., Huang, Y., Li, C.,
Liu, Y., and Xia, Y.: Day and Night Clouds Detection Using a Thermal-Infrared
All-Sky-View Camera, Remote Sens., 13, 1852, <ext-link xlink:href="https://doi.org/10.3390/rs13091852" ext-link-type="DOI">10.3390/rs13091852</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bib71"><label>71</label><?label 1?><mixed-citation>Wright, M. N. and Ziegler, A.: Ranger: A fast implementation of random forests for high dimensional data in C<inline-formula><mml:math id="M150" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mo>+</mml:mo></mml:mrow></mml:math></inline-formula> and R, J. Stat. Softw., 77, 1–17,
<ext-link xlink:href="https://doi.org/10.18637/jss.v077.i01" ext-link-type="DOI">10.18637/jss.v077.i01</ext-link>, 2017.</mixed-citation></ref>
      <ref id="bib1.bib72"><label>72</label><?label 1?><mixed-citation>Wright, M. N., Wager, S., and Probst, P.: Package `ranger', R Reference
Document, 1–25, available at: <uri>https://cran.r-project.org/web/packages/ranger/ranger.pdf</uri> (last access: 1 June 2021), 2020.</mixed-citation></ref>
      <ref id="bib1.bib73"><label>73</label><?label 1?><mixed-citation>Xie, W., Liu, D., Yang, M., Chen, S., Wang, B., Wang, Z., Xia, Y., Liu, Y.,
Wang, Y., and Zhang, C.: SegCloud: a novel cloud image segmentation model using a deep convolutional neural network for ground-based all-sky-view camera observation, Atmos. Meas. Tech., 13, 1953–1961,
<ext-link xlink:href="https://doi.org/10.5194/amt-13-1953-2020" ext-link-type="DOI">10.5194/amt-13-1953-2020</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bib74"><label>74</label><?label 1?><mixed-citation>Xiong, Z., Cui, Y., Liu, Z., Zhao, Y., Hu, M., and Hu, J.: Evaluating
explorative prediction power of machine learning algorithms for materials
discovery using <inline-formula><mml:math id="M151" display="inline"><mml:mi>k</mml:mi></mml:math></inline-formula>-fold forward cross-validation, Comput. Mater. Sci., 171,
109203, <ext-link xlink:href="https://doi.org/10.1016/j.commatsci.2019.109203" ext-link-type="DOI">10.1016/j.commatsci.2019.109203</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bib75"><label>75</label><?label 1?><mixed-citation>Yabuki, M., Shiobara, M., Nishinaka, K., and Kuji, M.: Development of a cloud
detection method from whole-sky color images, Polar Sci., 8, 315–326,
<ext-link xlink:href="https://doi.org/10.1016/j.polar.2014.07.004" ext-link-type="DOI">10.1016/j.polar.2014.07.004</ext-link>, 2014.</mixed-citation></ref>
      <ref id="bib1.bib76"><label>76</label><?label 1?><mixed-citation>Yang, J., Min, Q., Lu, W., Yao, W., Ma, Y., Du, J., Lu, T., and Liu, G.: An
automated cloud detection method based on th<?pagebreak page6710?>e green channel of total-sky
visible images, Atmos. Meas. Tech., 8, 4671–4679, <ext-link xlink:href="https://doi.org/10.5194/amt-8-4671-2015" ext-link-type="DOI">10.5194/amt-8-4671-2015</ext-link>, 2015.</mixed-citation></ref>
      <ref id="bib1.bib77"><label>77</label><?label 1?><mixed-citation>Yang, J., Min, Q., Lu, W., Ma, Y., Yao, W., Lu, T., Du, J., and Liu, G.: A total sky cloud detection method using real clear sky background, Atmos. Meas. Tech., 9, 587–597, <ext-link xlink:href="https://doi.org/10.5194/amt-9-587-2016" ext-link-type="DOI">10.5194/amt-9-587-2016</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bib78"><label>78</label><?label 1?><mixed-citation>Ye, L., Cao, Z., and Xiao, Y.: DeepCloud: Ground-based cloud image categorization using deep convolutional features, IEEE T. Geosci. Remote, 55, 5729–5740, <ext-link xlink:href="https://doi.org/10.1109/TGRS.2017.2712809" ext-link-type="DOI">10.1109/TGRS.2017.2712809</ext-link>, 2017.</mixed-citation></ref>
      <ref id="bib1.bib79"><label>79</label><?label 1?><mixed-citation>Ying, X.: An overview of overfitting and its solutions, J. Phys.: Conf. Ser., 1168, 022022, <ext-link xlink:href="https://doi.org/10.1088/1742-6596/1168/2/022022" ext-link-type="DOI">10.1088/1742-6596/1168/2/022022</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bib80"><label>80</label><?label 1?><mixed-citation>Yoshida, S., Misumi, R., and Maesaka, T.: Early Detection of Convective Echoes and Their Development Using a Ka-Band Radar Network. Weather Forecast., 36, 253–264, <ext-link xlink:href="https://doi.org/10.1175/WAF-D-19-0233.1" ext-link-type="DOI">10.1175/WAF-D-19-0233.1</ext-link>, 2021.
</mixed-citation></ref><?xmltex \hack{\newpage}?>
      <ref id="bib1.bib81"><label>81</label><?label 1?><mixed-citation>Yun, H. K. and Whang, S. M.: Development of a cloud cover reader from whole sky images, Int. J. Eng. Technol., 7, 252–256, <ext-link xlink:href="https://doi.org/10.14419/ijet.v7i3.33.21023" ext-link-type="DOI">10.14419/ijet.v7i3.33.21023</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bib82"><label>82</label><?label 1?><mixed-citation>Zhang, J., Liu, P., Zhang, F., and Song, Q.: CloudNet: Ground-based cloud
classification with deep convolutional neural network, Geophys. Res. Lett.,
45, 8665–8672, <ext-link xlink:href="https://doi.org/10.1029/2018GL077787" ext-link-type="DOI">10.1029/2018GL077787</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bib83"><label>83</label><?label 1?><mixed-citation>Zhang, S., Cheng, D., Deng, Z., Zong, M., and Deng, X.: A novel kNN algorithm
with data-drive k parameter computation, Pattern Recognit. Lett., 109, 44–54, <ext-link xlink:href="https://doi.org/10.1016/j.patrec.2017.09.036" ext-link-type="DOI">10.1016/j.patrec.2017.09.036</ext-link>, 2018.</mixed-citation></ref>

  </ref-list></back>
    <!--<article-title-html>Twenty-four-hour cloud cover calculation using  a ground-based imager with machine learning</article-title-html>
<abstract-html><p>In this study, image data features and machine learning methods were used to calculate 24&thinsp;h continuous cloud cover from image data obtained by a camera-based imager on the ground. The image data features were the time (Julian day and hour), solar zenith angle, and statistical characteristics of the red–blue ratio, blue–red difference, and luminance. These features were determined from the red, green, and blue brightness of images subjected to a pre-processing process involving masking removal and distortion correction. The collected image data were divided into training,
validation, and test sets and were used to optimize and evaluate the
accuracy of each machine learning method. The cloud cover calculated by each machine learning method was verified with human-eye observation data from a manned observatory. Supervised machine learning models suitable for
nowcasting, namely, support vector regression, random forest, gradient
boosting machine, <i>k</i>-nearest neighbor, artificial neural network, and multiple linear regression methods, were employed and their results were compared. The best learning results were obtained by the support vector regression model, which had an accuracy, recall, and precision of 0.94, 0.70, and 0.76, respectively. Further, bias, root mean square error, and correlation coefficient values of 0.04 tenths, 1.45 tenths, and 0.93, respectively, were obtained for the cloud cover calculated using the test set. When the difference between the calculated and observed cloud cover was allowed to range between 0, 1, and 2 tenths, high agreements of approximately 42&thinsp;%, 79&thinsp;%, and 91&thinsp;%, respectively, were obtained. The proposed system involving a ground-based imager and machine learning methods is expected to be suitable for application as an automated system to replace human-eye observations.</p></abstract-html>
<ref-html id="bib1.bib1"><label>1</label><mixed-citation>
Al Banna, M. H., Taher, K. A., Kaiser, M. S., Mahmud, M., Rahman, M. S.,
Hosen, A. S., and Cho, G. H.: Application of artificial intelligence in
predicting earthquakes: state-of-the-art and future challenges, IEEE Access.,
8, 192880–192923, <a href="https://doi.org/10.1109/ACCESS.2020.3029859" target="_blank">https://doi.org/10.1109/ACCESS.2020.3029859</a>, 2020.
</mixed-citation></ref-html>
<ref-html id="bib1.bib2"><label>2</label><mixed-citation>
Al-lahham, A., Theeb, O., Elalem, K., Alshawi, T., and Alshebeili, S.: Sky
Imager-Based Forecast of Solar Irradiance Using Machine Learning, Electronics, 9, 1700, <a href="https://doi.org/10.3390/electronics9101700" target="_blank">https://doi.org/10.3390/electronics9101700</a>, 2020.
</mixed-citation></ref-html>
<ref-html id="bib1.bib3"><label>3</label><mixed-citation>
Alonso, J., Batlles, F. J., López, G., and Ternero, A.: Sky camera imagery processing based on a sky classification using radiometric data, Energy, 68, 599–608, <a href="https://doi.org/10.1016/j.energy.2014.02.035" target="_blank">https://doi.org/10.1016/j.energy.2014.02.035</a>, 2014.
</mixed-citation></ref-html>
<ref-html id="bib1.bib4"><label>4</label><mixed-citation>
Alonso-Montesinos, J.: Real-Time Automatic Cloud Detection Using a Low-Cost
Sky Camera, Remote Sens., 12, 1382, <a href="https://doi.org/10.3390/rs12091382" target="_blank">https://doi.org/10.3390/rs12091382</a>, 2020.
</mixed-citation></ref-html>
<ref-html id="bib1.bib5"><label>5</label><mixed-citation>
Azhar, M. A. D. M., Hamid, N. S. A., Kamil, W. M. A. W. M., and Mohamad, N. S.: Daytime Cloud Detection Method Using the All-Sky Imager over PERMATApintar Observatory, Universe, 7, 41, <a href="https://doi.org/10.3390/universe7020041" target="_blank">https://doi.org/10.3390/universe7020041</a>, 2021.
</mixed-citation></ref-html>
<ref-html id="bib1.bib6"><label>6</label><mixed-citation>
Bergstra, J. and Bengio, Y.: Random search for hyper-parameter optimization, J. Mach. Learn. Res., 13, 281–305, 2012.
</mixed-citation></ref-html>
<ref-html id="bib1.bib7"><label>7</label><mixed-citation>
Blazek, M. and Pata, P.: Colour transformations and <i>K</i>-means segmentation for automatic cloud detection, Meteorol. Z., 24, 503–509, <a href="https://doi.org/10.1127/metz/2015/0656" target="_blank">https://doi.org/10.1127/metz/2015/0656</a>, 2015.
</mixed-citation></ref-html>
<ref-html id="bib1.bib8"><label>8</label><mixed-citation>
Boers, R., De Haij, M. J., Wauben, W. M. F., Baltink, H. K., Van Ulft, L.
H., Savenije, M., and Long, C. N.: Optimized fractional cloudiness determination from five ground-based remote sensing techniques, J. Geophys. Res., 115, D24116, <a href="https://doi.org/10.1029/2010JD014661" target="_blank">https://doi.org/10.1029/2010JD014661</a>, 2010.
</mixed-citation></ref-html>
<ref-html id="bib1.bib9"><label>9</label><mixed-citation>
Calbó, J., Long, C. N., González, J. A., Augustine, J., and McComiskey, A.: The thin border between cloud and aerosol: Sensitivity of several ground based observation techniques, Atmos. Res., 196, 248–260,
<a href="https://doi.org/10.1016/j.atmosres.2017.06.010" target="_blank">https://doi.org/10.1016/j.atmosres.2017.06.010</a>, 2017.
</mixed-citation></ref-html>
<ref-html id="bib1.bib10"><label>10</label><mixed-citation>
Cazorla, A., Olmo, F. J., Alados-Arboledas, L.: Development of a sky imager for cloud cover assessment, J. Opt. Soc. Am. A, 25, 29–39, <a href="https://doi.org/10.1364/JOSAA.25.000029" target="_blank">https://doi.org/10.1364/JOSAA.25.000029</a>, 2008.
</mixed-citation></ref-html>
<ref-html id="bib1.bib11"><label>11</label><mixed-citation>
Cazorla, A., Husillos, C., Antón, M., and Alados-Arboledas, L.: Multi-exposure adaptive threshold technique for cloud detection with sky imagers, Sol. Energy, 114, 268–277, <a href="https://doi.org/10.1016/j.solener.2015.02.006" target="_blank">https://doi.org/10.1016/j.solener.2015.02.006</a>,
2015.
</mixed-citation></ref-html>
<ref-html id="bib1.bib12"><label>12</label><mixed-citation>
Chauvin, R., Nou, J., Thil, S., and Grieu, S.: Modelling the clear-sky intensity distribution using a sky imager, Sol. Energy, 119, 1–17,
<a href="https://doi.org/10.1016/j.solener.2015.06.026" target="_blank">https://doi.org/10.1016/j.solener.2015.06.026</a>, 2015.
</mixed-citation></ref-html>
<ref-html id="bib1.bib13"><label>13</label><mixed-citation>
Çınar, Z. M., Abdussalam Nuhu, A., Zeeshan, Q., Korhan, O., Asmael, M., and Safaei, B.: Machine learning in predictive maintenance towards
sustainable smart manufacturing in industry 4.0, Sustainability, 12, 8211, <a href="https://doi.org/10.3390/su12198211" target="_blank">https://doi.org/10.3390/su12198211</a>, 2020.
</mixed-citation></ref-html>
<ref-html id="bib1.bib14"><label>14</label><mixed-citation>
Costa-Surós, M., Calbó, J., González, J. A., and Long, C. N.:
Comparing the cloud vertical structure derived from several methods based on
radiosonde profiles and ground-based remote sensing measurements, Atmos.
Meas. Tech., 7, 2757–2773, <a href="https://doi.org/10.5194/amt-7-2757-2014" target="_blank">https://doi.org/10.5194/amt-7-2757-2014</a>, 2014.
</mixed-citation></ref-html>
<ref-html id="bib1.bib15"><label>15</label><mixed-citation>
Dev, S., Savoy, F. M., Lee, Y. H., and Winkler, S.: Design of low-cost, compact and weather-proof whole sky imagers for High-Dynamic-Range captures, in: IGARSS 2015–2015 IEEE International Geoscience and Remote Sensing
Symposium, 26–31 July 2015, Milan, Italy, 5359–5362, 2015.
</mixed-citation></ref-html>
<ref-html id="bib1.bib16"><label>16</label><mixed-citation>
Dev, S., Wen, B., Lee, Y. H., and Winkler, S.: Ground-based image analysis: A
tutorial on machine-learning techniques and applications, IEEE Geosci. Remote Sens. M., 4, 79–93, <a href="https://doi.org/10.1109/MGRS.2015.2510448" target="_blank">https://doi.org/10.1109/MGRS.2015.2510448</a>, 2016.
</mixed-citation></ref-html>
<ref-html id="bib1.bib17"><label>17</label><mixed-citation>
Dev, S., Savoy, F. M., Lee, Y. H., and Winkler, S.: Nighttime sky/cloud image
segmentation, in: ICIP 2017–2017 IEEE International Conference on Image
Processing, 17–20 September 2017, Beijing, China, 345–349, 2017.
</mixed-citation></ref-html>
<ref-html id="bib1.bib18"><label>18</label><mixed-citation>
Dev, S., Nautiyal, A., Lee, Y. H., and Winkler, S.: Cloudsegnet: A deep network for nychthemeron cloud image segmentation, IEEE Geosci. Remote Sens. Lett., 16, 1814–1818, <a href="https://doi.org/10.1109/lgrs.2019.2912140" target="_blank">https://doi.org/10.1109/lgrs.2019.2912140</a>, 2019.
</mixed-citation></ref-html>
<ref-html id="bib1.bib19"><label>19</label><mixed-citation>
Fa, T., Xie, W., Wang, Y., and Xia, Y.: Development of an all-sky imaging system for cloud cover assessment, Appl. Optics, 58, 5516–5524,
<a href="https://doi.org/10.1364/AO.58.005516" target="_blank">https://doi.org/10.1364/AO.58.005516</a>, 2019.
</mixed-citation></ref-html>
<ref-html id="bib1.bib20"><label>20</label><mixed-citation>
Friedman, J. H.: Greedy function approximation: A gradient boosting machine,
Ann. Stat., 29, 1189–1232, <a href="https://doi.org/10.1214/aos/1013203451" target="_blank">https://doi.org/10.1214/aos/1013203451</a>, 2001.
</mixed-citation></ref-html>
<ref-html id="bib1.bib21"><label>21</label><mixed-citation>
Gani, W., Taleb, H., and Limam, M.: Support vector regression based residual
control charts, J. Appl. Stat., 37, 309–324, <a href="https://doi.org/10.1080/02664760903002667" target="_blank">https://doi.org/10.1080/02664760903002667</a>, 2010.
</mixed-citation></ref-html>
<ref-html id="bib1.bib22"><label>22</label><mixed-citation>
Geyer, C. J.: Generalized linear models in R, R Reference Document, 1–23, available at: <a href="https://www.stat.umn.edu/geyer/5931/mle/glm.pdf" target="_blank"/> (last access: 1 June 2021), 2003.
</mixed-citation></ref-html>
<ref-html id="bib1.bib23"><label>23</label><mixed-citation>
Ghonima, M. S., Urquhart, B., Chow, C. W., Shields, J. E., Cazorla, A., and
Kleissl, J.: A method for cloud detection and opacity classification based
on ground based sky imagery, Atmos. Meas. Tech., 5, 2881–2892,
<a href="https://doi.org/10.5194/amt-5-2881-2012" target="_blank">https://doi.org/10.5194/amt-5-2881-2012</a>, 2012.
</mixed-citation></ref-html>
<ref-html id="bib1.bib24"><label>24</label><mixed-citation>
Greenwell, B., Boehmke, B., and Cunningham, J.: GBM Developers: Package `gbm', R Reference Document, 1–39, available at:
<a href="https://cran.r-project.org/web/packages/gbm/gbm.pdf" target="_blank"/> (last access: 1 June 2021), 2020.
</mixed-citation></ref-html>
<ref-html id="bib1.bib25"><label>25</label><mixed-citation>
Heinle, A., Macke, A., and Srivastav, A.: Automatic cloud classification of
whole sky images, Atmos. Meas. Tech., 3, 557–567, <a href="https://doi.org/10.5194/amt-3-557-2010" target="_blank">https://doi.org/10.5194/amt-3-557-2010</a>, 2010.
</mixed-citation></ref-html>
<ref-html id="bib1.bib26"><label>26</label><mixed-citation>
Huo, J. and Lu, D.: Cloud determination of all-sky images under low-visibility conditions, J. Atmos. Ocean. Tech., 26, 2172–2181,
<a href="https://doi.org/10.1175/2009JTECHA1324.1" target="_blank">https://doi.org/10.1175/2009JTECHA1324.1</a>, 2009.
</mixed-citation></ref-html>
<ref-html id="bib1.bib27"><label>27</label><mixed-citation>
Kazantzidis, A., Tzoumanikas, P., Bais, A. F., Fotopoulos, S., and Economou, G.: Cloud detection and classification with the use of whole-sky ground-based
images, Atmos. Res., 113, 80–88, <a href="https://doi.org/10.1016/j.atmosres.2012.05.005" target="_blank">https://doi.org/10.1016/j.atmosres.2012.05.005</a>, 2012.
</mixed-citation></ref-html>
<ref-html id="bib1.bib28"><label>28</label><mixed-citation>
Kim, B. Y. and Cha, J. W.: Cloud Observation and Cloud Cover Calculation at
Nighttime Using the Automatic Cloud Observation System (ACOS) Package, Remote Sens., 12, 2314, <a href="https://doi.org/10.3390/rs12142314" target="_blank">https://doi.org/10.3390/rs12142314</a>, 2020.
</mixed-citation></ref-html>
<ref-html id="bib1.bib29"><label>29</label><mixed-citation>
Kim, B. Y. and Lee, K. T.: Radiation component calculation and energy budget
analysis for the Korean Peninsula region, Remote Sens., 10, 1147,
<a href="https://doi.org/10.3390/rs10071147" target="_blank">https://doi.org/10.3390/rs10071147</a>, 2018.
</mixed-citation></ref-html>
<ref-html id="bib1.bib30"><label>30</label><mixed-citation>
Kim, B. Y. and Lee, K. T.: Using the himawari-8 ahi multi-channel to improve
the calculation accuracy of outgoing longwave radiation at the top of the
atmosphere. Remote Sens., 11, 589, <a href="https://doi.org/10.3390/rs11050589" target="_blank">https://doi.org/10.3390/rs11050589</a>, 2019.
</mixed-citation></ref-html>
<ref-html id="bib1.bib31"><label>31</label><mixed-citation>
Kim, B. Y., Jee, J. B., Zo, I. S., and Lee, K. T.: Cloud cover retrieved from
skyviewer: A validation with human observations. Asia-Pac, J. Atmos. Sci., 52, 1–10, <a href="https://doi.org/10.1007/s13143-015-0083-4" target="_blank">https://doi.org/10.1007/s13143-015-0083-4</a>, 2016.
</mixed-citation></ref-html>
<ref-html id="bib1.bib32"><label>32</label><mixed-citation>
Kim, B. Y., Lee, K. T., Jee, J. B., and Zo, I. S.: Retrieval of outgoing longwave radiation at top-of-atmosphere using Himawari-8 AHI data, Remote
Sens. Environ., 204, 498–508, <a href="https://doi.org/10.1016/j.rse.2017.10.006" target="_blank">https://doi.org/10.1016/j.rse.2017.10.006</a>, 2018.
</mixed-citation></ref-html>
<ref-html id="bib1.bib33"><label>33</label><mixed-citation>
Kim, B. Y., Cha, J. W., Ko, A. R., Jung, W., and Ha, J. C.: Analysis of the
occurrence frequency of seedable clouds on the Korean Peninsula for precipitation enhancement experiments, Remote Sens., 12, 1487,
<a href="https://doi.org/10.3390/rs12091487" target="_blank">https://doi.org/10.3390/rs12091487</a>, 2020a.
</mixed-citation></ref-html>
<ref-html id="bib1.bib34"><label>34</label><mixed-citation>
Kim, B. Y., Cha, J. W., Jung, W., and Ko, A. R.: Precipitation Enhancement
Experiments in Catchment Areas of Dams: Evaluation of Water Resource Augmentation and Economic Benefits, Remote Sens., 12, 3730, <a href="https://doi.org/10.3390/rs12223730" target="_blank">https://doi.org/10.3390/rs12223730</a>, 2020b.
</mixed-citation></ref-html>
<ref-html id="bib1.bib35"><label>35</label><mixed-citation>
Kim, B. Y., Cha, J. W., Chang, K. H., and Lee, C.: Visibility Prediction over
South Korea Based on Random Forest, Atmosphere, 12, 552, <a href="https://doi.org/10.3390/atmos12050552" target="_blank">https://doi.org/10.3390/atmos12050552</a>, 2021.
</mixed-citation></ref-html>
<ref-html id="bib1.bib36"><label>36</label><mixed-citation>
Kreuter, A., Zangerl, M., Schwarzmann, M., and Blumthaler, M.: All-sky imaging: a simple, versatile system for atmospheric research, Appl. Optics, 48, 1091–1097, <a href="https://doi.org/10.1364/AO.48.001091" target="_blank">https://doi.org/10.1364/AO.48.001091</a>, 2009.
</mixed-citation></ref-html>
<ref-html id="bib1.bib37"><label>37</label><mixed-citation>
Krinitskiy, M. A. and Sinitsyn, A. V.: Adaptive algorithm for cloud cover
estimation from all-sky images over the sea, Oceanology, 56, 315–319,
<a href="https://doi.org/10.1134/S0001437016020132" target="_blank">https://doi.org/10.1134/S0001437016020132</a>, 2016.
</mixed-citation></ref-html>
<ref-html id="bib1.bib38"><label>38</label><mixed-citation>
Kyba, C. C., Ruhtz, T., Fischer, J., and Hölker, F.: Red is the new black: how the colour of urban skyglow varies with cloud cover, Mon. Notic. Roy. Astron. Soc., 425, 701–708, <a href="https://doi.org/10.1111/j.1365-2966.2012.21559.x" target="_blank">https://doi.org/10.1111/j.1365-2966.2012.21559.x</a>, 2012.
</mixed-citation></ref-html>
<ref-html id="bib1.bib39"><label>39</label><mixed-citation>
Lalonde, J. F., Narasimhan, S. G., and Efros, A. A.: What do the sun and the sky tell us about the camera?, Int. J. Comput. Vis., 88, 24–51,
<a href="https://doi.org/10.1007/s11263-009-0291-4" target="_blank">https://doi.org/10.1007/s11263-009-0291-4</a>, 2010.
</mixed-citation></ref-html>
<ref-html id="bib1.bib40"><label>40</label><mixed-citation>
Lee, S. H., Kim, B. Y., Lee, K. T., Zo, I. S., Jung, H. S., and Rim, S. H.:
Retrieval of reflected shortwave radiation at the top of the atmosphere
using Himawari-8/AHI data, Remote Sens., 10, 213, <a href="https://doi.org/10.3390/rs10020213" target="_blank">https://doi.org/10.3390/rs10020213</a>, 2018.
</mixed-citation></ref-html>
<ref-html id="bib1.bib41"><label>41</label><mixed-citation>
Li, Q., Lu, W., and Yang, J.: A hybrid thresholding algorithm for cloud
detection on ground-based color images, J. Atmos. Ocean. Tech., 28, 1286–1296, <a href="https://doi.org/10.1175/JTECH-D-11-00009.1" target="_blank">https://doi.org/10.1175/JTECH-D-11-00009.1</a>, 2011.
</mixed-citation></ref-html>
<ref-html id="bib1.bib42"><label>42</label><mixed-citation>
Li, X., Lu, Z., Zhou, Q., and Xu, Z.: A Cloud Detection Algorithm with Reduction of Sunlight Interference in Ground-Based Sky Images, Atmosphere, 10, 640, <a href="https://doi.org/10.3390/atmos10110640" target="_blank">https://doi.org/10.3390/atmos10110640</a>, 2019.
</mixed-citation></ref-html>
<ref-html id="bib1.bib43"><label>43</label><mixed-citation>
Liu, S., Zhang, L., Zhang, Z., Wang, C., and Xiao, B.: Automatic cloud detection for all-sky images using superpixel segmentation, IEEE Geosci. Remote Sens. Lett., 12, 354–358, <a href="https://doi.org/10.1109/LGRS.2014.2341291" target="_blank">https://doi.org/10.1109/LGRS.2014.2341291</a>, 2014.
</mixed-citation></ref-html>
<ref-html id="bib1.bib44"><label>44</label><mixed-citation>
Long, C. N., Sabburg, J. M., Calbó, J., and Pagès, D.: Retrieving cloud characteristics from ground-based daytime color all-sky images, J. Atmos. Ocean. Tech., 23, 633–652, <a href="https://doi.org/10.1175/JTECH1875.1" target="_blank">https://doi.org/10.1175/JTECH1875.1</a>, 2006.
</mixed-citation></ref-html>
<ref-html id="bib1.bib45"><label>45</label><mixed-citation>
Lothon, M., Barnéoud, P., Gabella, O., Lohou, F., Derrien, S., Rondi, S., Chiriaco, M., Bastin, S., Dupont, J. C., Haeffelin, M., Badosa, J., Pascal, N., and Montoux, N.: ELIFAN, an algorithm for the estimation of cloud cover from sky imagers, Atmos. Meas. Tech., 12, 5519–5534, <a href="https://doi.org/10.5194/amt-12-5519-2019" target="_blank">https://doi.org/10.5194/amt-12-5519-2019</a>, 2019.
</mixed-citation></ref-html>
<ref-html id="bib1.bib46"><label>46</label><mixed-citation>
Mantelli Neto, S. L., von Wangenheim, A., Pereira, E. B., and Comunello, E.: The use of Euclidean geometric distance on RGB color space for the
classification of sky and cloud patterns, J. Atmos. Ocean. Tech., 27, 1504–1517, <a href="https://doi.org/10.1175/2010JTECHA1353.1" target="_blank">https://doi.org/10.1175/2010JTECHA1353.1</a>, 2010.
</mixed-citation></ref-html>
<ref-html id="bib1.bib47"><label>47</label><mixed-citation>
Martínez, F., Frías, M. P., Charte, F., and Rivera, A. J.: Time Series Forecasting with KNN in R: the tsfknn Package, R J., 11, 229–242, <a href="https://doi.org/10.32614/RJ-2019-004" target="_blank">https://doi.org/10.32614/RJ-2019-004</a>, 2019.
</mixed-citation></ref-html>
<ref-html id="bib1.bib48"><label>48</label><mixed-citation>
Meyer, D.: Support vector machines. The Interface to libsvm in package e1071, R News, 23–26, available at: <a href="https://cran.r-project.org/web/packages/e1071/vignettes/svmdoc.pdf" target="_blank"/> (last access: 1 June 2021), 2001.
</mixed-citation></ref-html>
<ref-html id="bib1.bib49"><label>49</label><mixed-citation>
Meyer, D., Dimitriadou, E., Hornik, K., Weingessel, A., Leisch, F., Chang, C. C., and Lin, C. C.: Package `e1071', R Reference Document, 1–66, available at: <a href="https://cran.r-project.org/web/packages/e1071/e1071.pdf" target="_blank"/>, last access: 1 June 2021.
</mixed-citation></ref-html>
<ref-html id="bib1.bib50"><label>50</label><mixed-citation>
Nguyen, D. A. and Kleissl, J.: Stereographic methods for cloud base height
determination using two sky imagers, Sol. Energy, 107, 495–509,
<a href="https://doi.org/10.1016/j.solener.2014.05.005" target="_blank">https://doi.org/10.1016/j.solener.2014.05.005</a>, 2014.
</mixed-citation></ref-html>
<ref-html id="bib1.bib51"><label>51</label><mixed-citation>
Olive, D. J.: Multiple linear regression, in: Linear regression. Springer,
Cham, 17–83, 2017.
</mixed-citation></ref-html>
<ref-html id="bib1.bib52"><label>52</label><mixed-citation>
Peng, Z., Yu, D., Huang, D., Heiser, J., Yoo, S., and Kalb, P.: 3D cloud
detection and tracking system for solar forecast using multiple sky imagers,
Sol. Energy, 118, 496–519, <a href="https://doi.org/10.1016/j.solener.2015.05.037" target="_blank">https://doi.org/10.1016/j.solener.2015.05.037</a>, 2015.
</mixed-citation></ref-html>
<ref-html id="bib1.bib53"><label>53</label><mixed-citation>
Ridgeway, G.: Generalized Boosted Models: A guide to the gbm package, Update,
1, 1–15, 2020.
</mixed-citation></ref-html>
<ref-html id="bib1.bib54"><label>54</label><mixed-citation>
Ripley, B. and Venables, W.: Package `class', R Reference Document, 1–19,
available at: <a href="https://cran.r-project.org/web/packages/class/class.pdf" target="_blank"/> (last access: 1 June 2021), 2021a.
</mixed-citation></ref-html>
<ref-html id="bib1.bib55"><label>55</label><mixed-citation>
Ripley, B. and Venables, W.: Package `nnet', R Reference Document, 1–11,
available at: <a href="https://cran.r-project.org/web/packages/nnet/nnet.pdf" target="_blank"/> (last access: 1 June 2021), 2021b.
</mixed-citation></ref-html>
<ref-html id="bib1.bib56"><label>56</label><mixed-citation>
Román, R., Cazorla, A., Toledano, C., Olmo, F. J., Cachorro, V. E., de Frutos, A., and Alados-Arboledas, L.: Cloud cover detection combining high
dynamic range sky images and ceilometer measurements, Atmos. Res., 196,
224–236, <a href="https://doi.org/10.1016/j.atmosres.2017.06.006" target="_blank">https://doi.org/10.1016/j.atmosres.2017.06.006</a>, 2017.
</mixed-citation></ref-html>
<ref-html id="bib1.bib57"><label>57</label><mixed-citation>
Rosa, J. P., Guerra, D. J., Horta, N. C., Martins, R. M., and Lourenço, N. C.: Overview of Artificial Neural Networks, in: Using Artificial Neural
Networks for Analog Integrated Circuit Design Automation, Springer, Cham, 21–44, 2020.
</mixed-citation></ref-html>
<ref-html id="bib1.bib58"><label>58</label><mixed-citation>
Sazzad, T. S., Islam, S., Mamun, M. M. R. K., and Hasan, M. Z.: Establishment of an efficient color model from existing models for better gamma encoding in image processing, Int. J. Image Process., 7, 90–100, 2013.
</mixed-citation></ref-html>
<ref-html id="bib1.bib59"><label>59</label><mixed-citation>
Shi, C., Zhou, Y., Qiu, B., He, J., Ding, M., and Wei, S.: Diurnal and nocturnal cloud segmentation of all-sky imager (ASI) images using enhancement fully convolutional networks. Atmos. Meas. Tech., 12, 4713–4724,
<a href="https://doi.org/10.5194/amt-12-4713-2019" target="_blank">https://doi.org/10.5194/amt-12-4713-2019</a>, 2019.
</mixed-citation></ref-html>
<ref-html id="bib1.bib60"><label>60</label><mixed-citation>
Shi, C., Zhou, Y., and Qiu, B.: CloudU-Netv2: A Cloud Segmentation Method for
Ground-Based Cloud Images Based on Deep Learning, Neural Process. Lett., 53, 2715–2728, <a href="https://doi.org/10.1007/s11063-021-10457-2" target="_blank">https://doi.org/10.1007/s11063-021-10457-2</a>, 2021.
</mixed-citation></ref-html>
<ref-html id="bib1.bib61"><label>61</label><mixed-citation>
Shields, J. E., Karr, M. E., Johnson, R. W., and Burden, A. R.: Day/night whole sky imagers for 24-h cloud and sky assessment: history and overview, Appl. Optics, 52, 1605–1616, <a href="https://doi.org/10.1364/AO.52.001605" target="_blank">https://doi.org/10.1364/AO.52.001605</a>, 2013.
</mixed-citation></ref-html>
<ref-html id="bib1.bib62"><label>62</label><mixed-citation>
Shields, J. E., Burden, A. R., and Karr, M. E.: Atmospheric cloud algorithms for day/night whole sky imagers, Appl. Optics, 58, 7050–7062,
<a href="https://doi.org/10.1364/AO.58.007050" target="_blank">https://doi.org/10.1364/AO.58.007050</a>, 2019.
</mixed-citation></ref-html>
<ref-html id="bib1.bib63"><label>63</label><mixed-citation>
Shimoji, N., Aoyama, R., and Hasegawa, W.: Spatial variability of correlated
color temperature of lightning channels, Results Phys., 6, 161–162,
<a href="https://doi.org/10.1016/j.rinp.2016.03.004" target="_blank">https://doi.org/10.1016/j.rinp.2016.03.004</a>, 2016.
</mixed-citation></ref-html>
<ref-html id="bib1.bib64"><label>64</label><mixed-citation>
Shin, J. Y., Kim, B. Y., Park, J., Kim, K. R., and Cha, J. W.: Prediction of
Leaf Wetness Duration Using Geostationary Satellite Observations and Machine
Learning Algorithms, Remote Sens., 12, 3076, <a href="https://doi.org/10.3390/rs12183076" target="_blank">https://doi.org/10.3390/rs12183076</a>, 2020.
</mixed-citation></ref-html>
<ref-html id="bib1.bib65"><label>65</label><mixed-citation>
Singh, A., Kotiyal, V., Sharma, S., Nagar, J., and Lee, C. C.: A machine
learning approach to predict the average localization error with applications to wireless sensor networks, IEEE Access., 8, 208253–208263,
<a href="https://doi.org/10.1109/ACCESS.2020.3038645" target="_blank">https://doi.org/10.1109/ACCESS.2020.3038645</a>, 2020.
</mixed-citation></ref-html>
<ref-html id="bib1.bib66"><label>66</label><mixed-citation>
Sunil, S., Padmakumari, B., Pandithurai, G., Patil, R. D., and Naidu, C. V.:
Diurnal (24&thinsp;h) cycle and seasonal variability of cloud fraction retrieved
from a Whole Sky Imager over a complex terrain in the Western Ghats and
comparison with MODIS, Atmos. Res., 248, 105180, <a href="https://doi.org/10.1016/j.atmosres.2020.105180" target="_blank">https://doi.org/10.1016/j.atmosres.2020.105180</a>, 2021.
</mixed-citation></ref-html>
<ref-html id="bib1.bib67"><label>67</label><mixed-citation>
Taghizadeh-Mehrjardi, R., Neupane, R., Sood, K., and Kumar, S.: Artificial bee colony feature selection algorithm combined with machine learning algorithms to predict vertical and lateral distribution of soil organic matter in South Dakota, USA, Carbon Manage., 8, 277–291,
<a href="https://doi.org/10.1080/17583004.2017.1330593" target="_blank">https://doi.org/10.1080/17583004.2017.1330593</a>, 2017.
</mixed-citation></ref-html>
<ref-html id="bib1.bib68"><label>68</label><mixed-citation>
Valentín, L., Peña-Cruz, M. I., Moctezuma, D., Peña-Martínez, C. M., Pineda-Arellano, C. A., and Díaz-Ponce, A.: Towards the Development of a Low-Cost Irradiance Nowcasting Sky Imager, Appl. Sci., 9, 1131, <a href="https://doi.org/10.3390/app9061131" target="_blank">https://doi.org/10.3390/app9061131</a>, 2019.
</mixed-citation></ref-html>
<ref-html id="bib1.bib69"><label>69</label><mixed-citation>
Wang, G., Kurtz, B., and Kleissl, J.: Cloud base height from sky imager and
cloud speed sensor, Sol. Energy, 131, 208–221, <a href="https://doi.org/10.1016/j.solener.2016.02.027" target="_blank">https://doi.org/10.1016/j.solener.2016.02.027</a>, 2016.
</mixed-citation></ref-html>
<ref-html id="bib1.bib70"><label>70</label><mixed-citation>
Wang, Y., Liu, D., Xie, W., Yang, M., Gao, Z., Ling, X., Huang, Y., Li, C.,
Liu, Y., and Xia, Y.: Day and Night Clouds Detection Using a Thermal-Infrared
All-Sky-View Camera, Remote Sens., 13, 1852, <a href="https://doi.org/10.3390/rs13091852" target="_blank">https://doi.org/10.3390/rs13091852</a>, 2021.
</mixed-citation></ref-html>
<ref-html id="bib1.bib71"><label>71</label><mixed-citation>
Wright, M. N. and Ziegler, A.: Ranger: A fast implementation of random forests for high dimensional data in C+ +  and R, J. Stat. Softw., 77, 1–17,
<a href="https://doi.org/10.18637/jss.v077.i01" target="_blank">https://doi.org/10.18637/jss.v077.i01</a>, 2017.
</mixed-citation></ref-html>
<ref-html id="bib1.bib72"><label>72</label><mixed-citation>
Wright, M. N., Wager, S., and Probst, P.: Package `ranger', R Reference
Document, 1–25, available at: <a href="https://cran.r-project.org/web/packages/ranger/ranger.pdf" target="_blank"/> (last access: 1 June 2021), 2020.
</mixed-citation></ref-html>
<ref-html id="bib1.bib73"><label>73</label><mixed-citation>
Xie, W., Liu, D., Yang, M., Chen, S., Wang, B., Wang, Z., Xia, Y., Liu, Y.,
Wang, Y., and Zhang, C.: SegCloud: a novel cloud image segmentation model using a deep convolutional neural network for ground-based all-sky-view camera observation, Atmos. Meas. Tech., 13, 1953–1961,
<a href="https://doi.org/10.5194/amt-13-1953-2020" target="_blank">https://doi.org/10.5194/amt-13-1953-2020</a>, 2020.
</mixed-citation></ref-html>
<ref-html id="bib1.bib74"><label>74</label><mixed-citation>
Xiong, Z., Cui, Y., Liu, Z., Zhao, Y., Hu, M., and Hu, J.: Evaluating
explorative prediction power of machine learning algorithms for materials
discovery using <i>k</i>-fold forward cross-validation, Comput. Mater. Sci., 171,
109203, <a href="https://doi.org/10.1016/j.commatsci.2019.109203" target="_blank">https://doi.org/10.1016/j.commatsci.2019.109203</a>, 2020.
</mixed-citation></ref-html>
<ref-html id="bib1.bib75"><label>75</label><mixed-citation>
Yabuki, M., Shiobara, M., Nishinaka, K., and Kuji, M.: Development of a cloud
detection method from whole-sky color images, Polar Sci., 8, 315–326,
<a href="https://doi.org/10.1016/j.polar.2014.07.004" target="_blank">https://doi.org/10.1016/j.polar.2014.07.004</a>, 2014.
</mixed-citation></ref-html>
<ref-html id="bib1.bib76"><label>76</label><mixed-citation>
Yang, J., Min, Q., Lu, W., Yao, W., Ma, Y., Du, J., Lu, T., and Liu, G.: An
automated cloud detection method based on the green channel of total-sky
visible images, Atmos. Meas. Tech., 8, 4671–4679, <a href="https://doi.org/10.5194/amt-8-4671-2015" target="_blank">https://doi.org/10.5194/amt-8-4671-2015</a>, 2015.
</mixed-citation></ref-html>
<ref-html id="bib1.bib77"><label>77</label><mixed-citation>
Yang, J., Min, Q., Lu, W., Ma, Y., Yao, W., Lu, T., Du, J., and Liu, G.: A total sky cloud detection method using real clear sky background, Atmos. Meas. Tech., 9, 587–597, <a href="https://doi.org/10.5194/amt-9-587-2016" target="_blank">https://doi.org/10.5194/amt-9-587-2016</a>, 2016.
</mixed-citation></ref-html>
<ref-html id="bib1.bib78"><label>78</label><mixed-citation>
Ye, L., Cao, Z., and Xiao, Y.: DeepCloud: Ground-based cloud image categorization using deep convolutional features, IEEE T. Geosci. Remote, 55, 5729–5740, <a href="https://doi.org/10.1109/TGRS.2017.2712809" target="_blank">https://doi.org/10.1109/TGRS.2017.2712809</a>, 2017.
</mixed-citation></ref-html>
<ref-html id="bib1.bib79"><label>79</label><mixed-citation>
Ying, X.: An overview of overfitting and its solutions, J. Phys.: Conf. Ser., 1168, 022022, <a href="https://doi.org/10.1088/1742-6596/1168/2/022022" target="_blank">https://doi.org/10.1088/1742-6596/1168/2/022022</a>, 2019.
</mixed-citation></ref-html>
<ref-html id="bib1.bib80"><label>80</label><mixed-citation>
Yoshida, S., Misumi, R., and Maesaka, T.: Early Detection of Convective Echoes and Their Development Using a Ka-Band Radar Network. Weather Forecast., 36, 253–264, <a href="https://doi.org/10.1175/WAF-D-19-0233.1" target="_blank">https://doi.org/10.1175/WAF-D-19-0233.1</a>, 2021.

</mixed-citation></ref-html>
<ref-html id="bib1.bib81"><label>81</label><mixed-citation>
Yun, H. K. and Whang, S. M.: Development of a cloud cover reader from whole sky images, Int. J. Eng. Technol., 7, 252–256, <a href="https://doi.org/10.14419/ijet.v7i3.33.21023" target="_blank">https://doi.org/10.14419/ijet.v7i3.33.21023</a>, 2018.
</mixed-citation></ref-html>
<ref-html id="bib1.bib82"><label>82</label><mixed-citation>
Zhang, J., Liu, P., Zhang, F., and Song, Q.: CloudNet: Ground-based cloud
classification with deep convolutional neural network, Geophys. Res. Lett.,
45, 8665–8672, <a href="https://doi.org/10.1029/2018GL077787" target="_blank">https://doi.org/10.1029/2018GL077787</a>, 2018.
</mixed-citation></ref-html>
<ref-html id="bib1.bib83"><label>83</label><mixed-citation>
Zhang, S., Cheng, D., Deng, Z., Zong, M., and Deng, X.: A novel kNN algorithm
with data-drive k parameter computation, Pattern Recognit. Lett., 109, 44–54, <a href="https://doi.org/10.1016/j.patrec.2017.09.036" target="_blank">https://doi.org/10.1016/j.patrec.2017.09.036</a>, 2018.
</mixed-citation></ref-html>--></article>
