1,283
Views
4
CrossRef citations to date
0
Altmetric
Research Article

Atmosphere air temperature forecasting using the honey badger optimization algorithm: on the warmest and coldest areas of the world

, , , &
Article: 2174189 | Received 05 Dec 2022, Accepted 24 Jan 2023, Published online: 24 Feb 2023

Abstract

Precisely forecasting air temperature as a significant meteorological parameter has a critical role in environment quality management. Hence, this study employs a hybrid intelligent model for accurately monthly temperature forecasting for one to three times ahead in the hottest and coldest regions of the world. The hybrid model contains the artificial neural network (ANN) hybridized with the powerful hetaeristic Honey Badger Algorithm (HBA-ANN). The average mutual information (AMI) technique is employed to find the optimal time delay values for the temperature variable for different time horizons. Finally, the performance of the developed hybrid model is compared with the classical ANN and the Gene Expression Programming (GEP) using some statistical criteria, and the Taylor and scatter diagrams. Results indicated that in each time horizon, the HBA-ANN model with the lowest distance from observation points based on Taylor diagram, high values for NSE and R2, and low values for RMSE, MAE, and RSR outperformed the ANN and GEP models in both training and testing phases. Hence, using the Honey Badger Algorithm could increase the accuracy of the model. This model's precise performance supports the case for it to be employed to forecast other environmental parameters.

1. Introduction

Air temperature among the main factors is in the discussion of hydrology, agriculture, irrigation, and the environment. It also is a critical meteorological parameter, which can describe global warming and climate change (Alomar et al., Citation2022). Many processes on the earth's surface, such as photosynthesis, respiration, evaporation, are regulated by air temperature. Air temperature significantly influenced human life, numerous ecological and meteorological events, and crops in agricultural areas. In this regard, knowledge of spatial changes in air temperature on a wide scale is necessary to carry out climatic, meteorological and hydrological studies and investigations. Accurate air temperature forecasting is necessary for planning agricultural operations, recreational activities, tourism, transportation, and energy generation, as well as developing measures to deal with temperature fluctuations (Yakut & Süzülmüş, Citation2020). Forecasting air temperature is an effective parameter in providing early warnings for both heat stress and dangerous frost. A critical aspect of weather prediction is forecasting air temperature because protecting human lives and property is its purpose. Having an unhealthy air temperature can cause serious health problems (Lan et al., Citation2010; Schulte et al., Citation2016).

Based on literature, General circulation or Physical models and statistical models are two basic approaches of forecasting air temperature (Attar et al., Citation2018; Curceac et al., Citation2019; Naganna et al., Citation2019; Salcedo-Sanz et al., Citation2016; Smith et al., Citation2005). Physical model is a traditional method that named the numerical weather prediction (NWP) model and has been done by computer models based on physical equations (Marchuk, Citation2012; Rasnopolsky & Fox-Rabinovitz, Citation2006; Richardson, Citation2007). Needing to a large computational capacity and well-defined prior knowledge are two limitations of the physical models. Statistical models involving time series analysis such as ARIMA, simple regression based on ANN (Chen & Lai, Citation2011; Horenko et al., Citation2008; Kuligowski & Barros, Citation1998; Rasnopolsky & Fox-Rabinovitz, Citation2006; Voyant et al., Citation2012), support vector-based regressors (Radkiha and Shashi, Citation2009), and Bayesian networks (Abramson et al., Citation1996; Cofino et al., Citation2002) are the other approaches that are used to forecast air temperature. Statistical models attempt to minimize implementing computer models. Compared to physical models, they have less computationally intensive and more straightforward. A bulk of researches indicated that statistical models had consistent results with physical models. Cointgration approaches and regression approaches are two categories of statistical analysis (Alomar et al., Citation2022).

New technologies are essential to reducing simulation error and uncertainty of model variables (Moravej et al., Citation2020). The use of data-driven methodologies in air temperature simulation and forecasting has expanded as computational capabilities have increased (Cifuentes et al., Citation2020; Deng et al., Citation2022; Roy, Citation2020; Zhu et al., Citation2019). Machine-learning techniques such as Support Vector Model (SVM) (Mirarabi et al. Citation2019; S. band et al. Citation2021), Random Forest (RF) (Handler et al., Citation2020; Mcgovern et al., Citation2010; McGovern et al., Citation2011), and temporal modelling via Restricted Boltzmann Machines (RBM) (Mittelman et al., Citation2014; Sutskever et al., Citation2009) are successfully implemented in air temperature forecasting.

High accuracy of the deep learning methods such as Stacked Auto-Encoder (Liu et al., Citation2015), Stacked Denoising Auto-Encoder (Hossain et al., Citation2015), recurrent neural network (RNN), convolutional neural network (CNN), Conditional Restricted Boltzmann Machine (CRBM) and (Xingjian et al., Citation2015), and long short-term memory (LSTM) model (Roy, Citation2020; Yang et al., Citation2020) is also encouraged researchers to used them in air temperature forecasting. Furthermore, the capability of artificial neural networks (ANNs) to capture nonlinear relationships between inputs and outputs has attracted the attention of scientists in various disciplines, including meteorology, hydrology, and water resources (Anushka et al., Citation2020; Chevalier et al., Citation2011; Rajendra et al., Citation2019).

There are various studies, which used different models to predict air temperature. Mohsenzadeh Karimi, Kisi, Porrajabali, Rouhani-Nia, and Shiri (Citation2018) to predict the long-term monthly temperature, investigates the ability of random forest (RF), support vector machine (SVM), and Geostatistical Methods (GS) with 30-point data in Iran. Based on the results, the SVM and RF models outperformed the NF and NN models. Hanoon et al. (Citation2021) implemented random forest (RF), gradient boosting tree (GBT), different artificial neural network (ANN) architectures (multi-layered perceptron, radial basis function), and Linear regression (LR) models to forecast relative humidity and air temperature in Malaysia. Results indicated the high performance of the ANN-MLP in forecasting daily air temperature.

In the study by Lin et al. (Citation2021), a state-of-the-art hybrid of Multi-dimensional Complementary Ensemble Empirical Mode Decomposition (MCEEMD) outperformed the Radial Basis Function Neural Network (RBFNN) in forecasting maximum air temperature over next 7 days in Taipei. Very recently, Alomar et al. (Citation2022) to forecast air temperature in North America at daily and weekly horizons have implemented ARIMA, Random Forest (RF), Regression Tree (RT), Support Vector Regression (SVR), and Gradient Boosting Regression (GBR), and Quantile Regression Tree (QRT) models. Results showed that the SVR model outperformed the other implemented models. The study by Sari et al. (Citation2022) indicated the high accuracy of the GRU-LSTM Hybrid Model in predicting daily air temperature in Indonesia. In the study by Hou et al. (Citation2022) for forecasting air temperature in China, the CNN-LSTM hybrid model showed better performance than CNN and LSTM models. Li and Yang (Citation2022) showed that in forecasting daily air temperature, the SARIMA-LSTM hybrid model had high accuracy than ARIMA, SARIMA, and LSTM models. The study of Khan and Maity (Citation2022) indicated that the long short-term memory (LSTM) neural network (hereinafter hybrid Conv1D-LSTM) and one-dimensional convolutional neural network (Conv1D) hybrid model has high ability in daily air temperature prediction. Singh et al. (Citation2022) showed the superior performance of the Genetic Algorithm (GA) based hybrid machine learning-pedotransfer Function (ML-PTF) model in forecasting spatial pattern of saturated hydraulic conductivity. In the study by Zhang et al. (Citation2022), the Adaptive Neuro-Fuzzy Inference System (ANFIS) model outperformed the bilayered NN model in forecasting the daily dew point temperature.

As it is impossible to detect accurate air temperature forecasting results with limited data in some circumstances, Artificial Intelligence (AI) models would be a better choice for the purpose of understanding and illuminating the underlying mechanisms regulating such variations in temperature patterns. As a result, data-driven simulation methods in the form of both standalone and hybrid models are rapidly evolving in a variety of scientific domains, including hydrology. Hybrid models can outperform standalone models. Since there is usually a trade-off between model performance and training time, as a result, a hybrid model that can deliver both high performance and quick training time is required.

Research significance and novelty of the work:

In large-scale optimization issues finding a best possible solution in the search space is a complicated matter. Furthermore, the algorithm convergence has not been much influenced with changing algorithm variables. Hence, for high complexity and massive dataset, even if the correct initial parameters have been determined, the algorithm could not accomplish adequate exploitation and exploration. Consequently, to perform the best exploitation and exploration for earning the general local and global searches, applying a powerful operator is needed. In this work, we explore and compare the abilities of the Honey Badger Algorithm (HBA), which has been successfully applied to atmosphere air temperature forecasting. To the best of our knowledge, no research has been done so far on the use of hybridization of the HBA with the artificial neural network (HBA-ANN) to forecast air temperatures in specific climatic regions, such as what is known as the hottest and coldest places on earth. Forecasting of atmosphere air temperature changes in such places on earth is very important due to the importance of water crises in many arid and semi-arid regions of the world. Besides, accurate air temperature forecasting helps to accurate energy saving and climate change policies. Therefore, the aim of this study is forecasting air temperature using the HBA-ANN hybrid model in Furnace Creek (hottest region) and Vostok stations (coldest region); and compare its efficiency with the standalone ANN model, and Gene Expression Programming (GEP).

The rest of the paper is organized as continuing to follow: Methodology is described in Section 2, outlining the research field and the data utilized. Section 3 displays and discusses the acquired results. Finally, the conclusions are provided to show the efficacy of the models created for the air temperature forecasting.

2. Materials and methods

2.1. Artificial neural network

Artificial neural network (ANN) consists of a black box model. In the ANN model, the simulation process is done according to the NNs of the human brain (Haykin, Citation1999; Nayak et al., Citation2006). Each ANN model has a set of connections between nodes, weights, and action functions. Moreover, there are an input layer, one or more hidden layers, and an output layer. the number of hidden layers and choosing an appropriate algorithm can have a significant impact on the efficiency of ANN when modelling problems. ANN models contain various algorithms, including: Levenberg-Marquardt (LM), Bayesian regularization (BR), adaptive learning rate (GDX), scaled conjugate gradient (SCG), and gradient descent with momentum. Artificial neural networks are typically designed through feed-forward neural networks (FFNN), which is one of the most popular, which is widely used for simulating and predicting hydrological problems. In the ANN, the relationship between input (x) and output (Y) is as follows: (1) Y=f(W1X1+W2X2++WnXn+b)(1) where f is the action function; b is the bias; Wi is the weight of link. The structure of a common ANN model is shown in Figure .

Figure 1. Common structure of the artificial neural network (ANN).

Figure 1. Common structure of the artificial neural network (ANN).

2.2. Honey badger algorithm (HBA)

Honey badgers are a type of mammals that mostly live in Southwest Asia, Africa, and the Indian subcontinent. Their prey includes more than 60 species of animals, including deadly snakes. In addition to being clever mammals, they enjoy honey and can utilize tools. Subspecies of honey badgers exist in 12 different groups. In the event of a predator being unable to escape, they are strong animals that are not afraid to attack them. Furthermore, they feed themselves by attacking bird nests and honey beehives by climbing trees. (Begg et al., Citation2003; Begg et al., Citation2005). A mathematical model based on the honey badger's behaviour is presented in the following sections.

2.2.1. Inspiration

HBA follows honey badger explore behaviour. Food sources are found by the HB either by scenting and digging or by tracking the honeyguide bird. Digging mode refers to the first stage, then another stage is known as honey mode. In the previous phase, to estimate its prey's location, it uses its sniffing skills, Honey badgers use the honey guide bird as a guide to locate beehives directly in the final mode.

2.2.2. Model based on mathematics

The following subsections provide an explanation of the mathematical models used in the HBA. In terms of mathematics, in the HBA, the population of candidate solutions (X) is as follows: (2) X=[X11X12X13X1dX21X22X23X2dXn1Xn2Xn3Xnd](2) j-th honey badger position (3) Xj=[xj1xj2xj3xj4](3) The first phase of the process is initialization:

In order to determine the relative positions of honey badgers with n populations, use the following expression: (4) Xj=LBi+r1(UBiLBi),r1[0,1](4) where LBi, and Ubi represent the lower and upper bounds of the search space, respectively; while xj presents the j-th honey badger position and in a population with size n refers to a nominee solution.

The second step is to define the intensity:

There are several factors that affect the intensity (I), including the efficacy of concentration of the prey and the distance among it and the j-th honey badger. Ij it shows the intensity of the smell of the prey; slow motion occurs when the scent is low and fast motion occurs when it is high. It is furnished by inverse square law (ISL) (Kapner et al., Citation2007), as represented in Figure  and described by (5) as: (5) Ij=r2S4πdj2,r2[0,1](5) (6) S=(XjXj+1)2(6) (7) dj=XpreyXj(7) where S, an indicator of the concentration or intensity of a source (prey position), and dj, distances between prey and j-th badgers are measured.

Figure 2. Flowchart of the HBA-ANN model.

Figure 2. Flowchart of the HBA-ANN model.

Updating the density factor is done in the third step:

As a result of the density factor (α) governing time-varying randomness from exploration to exploitation, exploration to exploitation can take place seamlessly. Using EquationEq. (8), the updated decreasing factor (α) with each iteration, the randomization decreases: (8) α=C×exp(iterMaxiter)(8) where (C) shows a fixed number that is more than 1 (the value is 2, by default), and Maxiter it indicates the maximum number of repeats.

The fourth step is to fleeing the local solution:

To flee from the local solution area, the current and two next steps are used in the HBA. According to this scenario, it uses a flag (F) to change the search direction and provides favourable conditions for accurate scanning of the search field.

The fifth step is to update the positions of the operators:

Update of HBA positions (Xnew) is involved into two modes: ‘digging mode’ and ‘honey mode’.

Detailed descriptions are provided below:

  • Digging mode: There is a cardioid shape to a honey badger's dig (Elseify, Kamel, Abdel-Mawgoud, & Elattar, Citation2022) in the process of digging. EquationEq. (9) simulates the approximate cardioids motion as: (9) Xnew=Xprey+FβIXprey+Fr3αdj|cos(2πr4)[1cos(2πr5)]|(9) where Xprey describes the best position the prey has achieved to date, in other words, the global optimal position. β 1 indicate Honey badgers are capable of finding food (default = 6). The r3, r4, and r5 are random numbers generated within an interval of three [0,1]. F shows a flag that substitutes the direction of the search, and it is specified by (10): (10) F={1if r61/1221otherwiser6[0,1](10) In the digging mode, three factors greatly influence the behaviour of honey badger: A measure of the intensity of the scent (I) of the prey (Xprey), the distance among prey and HB (dj), and the decreasing operator (α). In addition, while digging, a badger may feel any outage (F), letting it to discover even the prime prey location (Figure ).

  • Honey mode: Eq. (11) simulation of the honeyguide bird exploring a beehive with the honeyguide bird. (11) Xnew=Xprey+αdjr7F,r7[0,1](11)

Eq. (8) and (10) calculate the value of (α) and (F), respectively. Xnew, and Xprey describe the new position and location of the prey, respectively. this can be seen from (11) it is assumed that the HB searches near Xprey, which is the optimal prey position, in accordance with the distance information (dj). Also in this step, the evolution of search behavior (α) affects search results. it is also possible for honey badgers to experience perturbations (F).

Figure 3. Pseudocode of the HBA algorithm adopted by (Akopyan, 2015).

Figure 3. Pseudocode of the HBA algorithm adopted by (Akopyan, 2015).

According to theory, the HBA is noticed a global optimization method as the stages of exploration and exploitation. This algorithm is designed to be simple to implement and comprehend by reducing the number of agents that need to be improved. the HBA is mainly determined by three variables, The number of variables in a state (d), the number of populations (n), and the maximum number of iterations (Maxiter). By using honey attraction to attract soles in the population towards the optimal soles, the HBA optimization technique ensures strong local search capability. in addition, the density factor guarantees that the local optimal solutions will not be generated by the algorithm, as well as providing the algorithm with global search capabilities.

Figure , shows the overall flowchart of hybrid algorithm (HBA-ANN). Figure , showed the pseudo-code related to algorithm of Honey Badger Optimization. Table  shows the parameters used in the modelling stage of ANN and HBA-ANN models.

Table 1. Parameters used for ANN and HBA-ANN.

The main characteristics of Honey badger algorithm are the following: using randomization for initialization, two phases that are randomly determined, and Cardioid motion in the exploration phase. There are several advantages of using HBA for optimization, namely: Hybridization: it can be merged with more traditional optimization methods. HBA algorithm can be compounded with others and detect the best solution for problems by utilizing the operators and advantages of other algorithms. Broad applicability: it can be used to each problem that can be formulated as function optimization. The proposed algorithm, denoted as HBA-ANN, is an enhanced version of the newly developed honey Badger algorithm. The enhancement is achieved by adding training function and Transfer function from ANN. The HBA-ANN improved the diversity of the population and the ability to escape from local minimums.

2.3. Gene expression programming (GEP)

Genetic Programming (GP) is truly a ‘bottom-up’ process, can be identify for any defined time series. GEP does not have any assumption on the structure of the relation between dependent and independent variables. The structure of the relation is produced feasible by two efficient and details emulation of evolutionary procedures become feasible only when these detail functions hand-in-hand. A parse tree (a functional set of mathematical functions) and the terminal set (including parts of functions and their parameters) are two components of two GEP, which those chosen in this study are: (12) {+,,×,x,x2,x3,x}(12) The identified relation in a specific GEP modelling is never constant and continually evolving. The assessment begins from n primary chosen random population of models. The values of the dependent and independent variables are used to evaluate the fitness value of each model. As the population doffs from one descendant to another, old models were replaced with new models by having verifiably better efficiency (Khatibi et al., Citation2011).

The method implemented in the present research is adverted to as the Gene Expression Programming (GEP) according to erupting computer programs of various shapes. The shapes also codified in linear chromosomes of fixed lengths. It has been stated that GEP is 10 thousand times more effective than Genetic Programming (GP) systems (Ferreira, Citation2001a, Citation2001b) for some proofs, including: (i) the chromosomes are easy entities: compact, easy to manipulate genetically, linear, and relatively small; (ii) the parse tree is exclusively the statement of its relative chromosomes.

Exerting operators such as mutation and crossover to the winners, ‘offspring’ or ‘children’ are built, that crossovers are accountable for maintaining equal flair from a descendant to the other but mutation motives a random variation in the parse tree, however data mutation is feasible, too. This perfects the actions at the primary descendant and until termination, the procedure is repeated. For using GEP model, the GenXpro software application (Ferreira, Citation2001a), (Ferreira, Citation2001b) was implemented in this research.

2.4. Study area and datasets

Death Valley National Park in California, USA, with an area of 1300 km2, is famous for its very hot and dry climate. High temperature, high evaporation, low humidity, and low rainfall in summer are the main features of this valley. The place Furnace Creek is designated as a census-designated place (CDP) in Inyo County, California, United States. as reported by the US Census Bureau, Furnace Creek has a total area of 82 km2. Daytime temperatures range from roughly 18 °C in December to 47 °C in July, while overnight lows typically oscillate from 4 to 32 °C. During that time period, during July, the temperatures were the highest, with an average daily high temperature of 46.9 °C, and in June, the driest month was recorded, with an average monthly precipitation of 1.3 mm. As the world's highest temperature recorded, Furnace Creek holds the record, reaching 134 °F (56.7 °C) on 10 July 1913. The lowest temperatures on Death Valley summer nights can remain above 38 °C and this has not been recorded anywhere else in the USA. With a temperature of 56.7 °C on 10 July 1913 at the Furnace Station, this point is known as the hottest spot-on Earth. There are likely areas in the world with higher temperatures than Furnace Creek, but they are undocumented.

Antarctica is a remote continent with an area of 14 million km2, of which only 0.18% of its surface is ice-free. Antarctica, as one of the less understood parts of the earth, has an undeniable intrinsic value. The scientific information that is produced every year about the various vital and physical aspects of Antarctica shows the remarkable scientific value of Antarctica. Located in the highlands of the Antarctic Plateau, the Vostok region has a very cold and continental polar climate. There is a weak catabatic southerly south-westerly wind regime at Vostok station, with mean annual speeds above 5.0 m/s. In March and September, wind speed variations are pronounced, and in January, they are sharply Minimum. September has the highest wind speed (5.5 m/s) of all the months. January is the calmest month of the year. wind speed during the multiyear period is 4.2 m/s. The average temperature in January and December is around −32 °C as the highest temperature, but as a result of ‘coreless winter’, the temperature drops rapidly to about −65 °C by April. Vostok is the only deep drilling station in Antarctica where the thickness of the snow isotope composition can be directly compared with relatively long meteorological observational data (45 years). With a temperature of −89.2 °C on 21 July 1983 at Vostok Station, this point is known as the coldest point on Earth.

The geographical location of the two stations corresponding to the hottest and coldest point on Earth is shown in Figure . Therefore, in this study, first, daily temperature (T) data related to Furnace Creek and Vostok stations in the period 1979–2020 were collected. The statistical characteristics of T (°C) and coordinates of the studied stations are also in accordance with Table . Data segmentation in both stations were 70% (1979–2008) for the training and 30% (2009–2020) for the testing phase.

Figure 4. Geographical location of meteorological stations at the hottest (A) and coldest points (B) on Earth.

Figure 4. Geographical location of meteorological stations at the hottest (A) and coldest points (B) on Earth.

Table 2. Daily statistical characteristics of temperature variable and coordinates of the studied stations.

2.5. Error analysis

In order to evaluate the performance of the models, three metrics are used in this study and defined in Table . In this Table, p(i) and o(i) are the forecasted and observed temperature values, respectively; p¯ and o¯ are the averaged forecasted and observed temperature values, respectively; and N is the numbers of data (Shabani, Hayati, Pishbahar, Ghorbani, & Ghahremanzadeh, Citation2021).

Table 3. The evaluation metrics of the models.

A model will have high accuracy, if it has R2 close to 1, and the RMSE and MAE close to 0. The various ranges of the NSE and RSR performance metrics are presented in Table  (Moriasi et al., Citation2015).

Table 4. RSR and NSE performance ranges.

A variety of metrics have been used by researchers in the different fields to analyze their models, and there is no standard for a unified parameter. The NSE, RMSE, MAE, RSR, and R2 criteria are used in this study to evaluate the performance of the models. The lower the values of the RMSE and NSE the more accurate the models’ results are; conversely, the higher the values of the R2, the better the model’s performance. Taylor diagrams in order to evaluate the studied models, visual analysis was used. using the Taylor diagram offers the advantage of applying two common statistics, namely the standard deviation, and the correlation coefficient. The closer the forecasted values to the observed ones in terms of standard deviation and correlation coefficient, will be the most accurate model (Shabani et al., Citation2021).

3. Results and discussion

This section is slot into three subsections. The first subsection provides the findings of the average mutual information for the data collected from the warmest and the coldest place. The second subsection presents the suggested HBA-ANN model performance for the coldest and warmest places of the earth. The results are analyzed and discussed in the subsequent section.

3.1. Average mutual information (AMI)

In this study, the monthly temperatures are forecasted at three times horizons T + 1, T + 2, and T + 3 for both the warmest and coldest places. For this case, the standalone ANN and hybrid HBA-ANN models are used as modelling approaches. In order to run such models, determining optimal set of inputs is required. The average mutual information (AMI) technique is employed in this study to find the optimal time delay values for the temperature input variable in different time horizons at the selected warmest and coldest places. The results of the AMI technique are shown in Figure  for the coldest and warmest places. It is found that up to 4 (T, T-1, T-2, T-3, T-4) and 3 (T, T-1, T-2, T-3) time delays can be used as the optimal time delays for the monthly temperature variable in the coldest and warmest places, respectively (Table ).

Figure 5. AMI results for selecting optimal time lags at the coldest and warmest places.

Figure 5. AMI results for selecting optimal time lags at the coldest and warmest places.

3.2. Models’ performance in the coldest region (Vostok station)

Five different input combinations were considered for developing the ANN, GEP, and HBA-ANN models for forecasting air temperature at three times horizons (T + 1, T + 2, and T + 3) (Table ). The performance of the hybrid HBA-ANN models was assessed using the error metrics (R2, RMSE, MAE, RSR, and NSE) and compared with the standalone ANN and GEP in Table . According to this Table, in most cases, it is observed that the accuracy of the temperature forecasting increases with the number of input variables (time lags) for all models and all times ahead. In other words, often, the last input combination including lag times up to 4 (T, T-1, T-2, T-3, and T-4) is the best input combination for temperature forecasting in 1 month to 3 months ahead for all models.

Furthermore, it is seen that the hybrid HBA-ANN model outperforms the ANN and GEP models based on the low values of the RMSE, MAE, and RSR and high values of the R2 and NSE e.g. the R2, RMSE, and NSE values of the standalone ANN model for T + 1 time are 0.891, 4.519, 0.889 in the training period, and 0.909, 3.925, 0.906 in the testing period, respectively, while these values improves into 0.937, 3.399, 0.937 in the training period, and 0.970, 2.204, 0.970 in the testing period for the HBA-ANN hybrid model. Similar results can be seen for the other two times ahead (T + 2 and T + 3) for both of the ANN and HBA-ANN models. Furthermore, in both training and testing phases, compared to GEP model, the HBA-ANN model has high values for R2, NSE, and low values for RMSE, MAE, and RSR at each time horizons. In general, the HBA-ANN model shows high capability than the classical ANN and GEP models in air temperature forecasting based on all statistical metrics especially in the testing phase. As a last finding from this Table, the accuracy of the air temperature forecasted values is decreased by increasing times ahead even in proposed model. The best results of the ANN, GEP, and the HBA-ANN models for T + 1, T + 2, and T + 3 times were considered for the further graphical analysis.

Scatter plots of the observed and forecasted temperature along with R2, RMSE, 95% confidence interval, and error histograms were provided for the best ANN, GEP, and HBA-ANN models in order to judge the model’s performance in a more comprehensive manner for all times ahead (Figure ). The quantification of the uncertainty of the developed models is more ascertained by the results of the prediction intervals. These plots suggest superior forecasting capability of the HBA-ANN models, with R2 values ranging from 0.939 to 0.970, which are higher than those of the ANN model ranging from 0.833 to 0.908; and with RMSE values ranging from 2.204 to 3.212, which are lower than those of the ANN ranging from 3.925 to 5.126 for all three times ahead. This point is clearly visible in the scatter plots drawn between the observed and modelled temperature data in terms of having lesser dispersions around the perfect line (i.e. y = x) for all three times ahead. The shapes of the error histograms are approximately similar to the normal distribution especially for the HBA-ANN models, which it is another evidence for high accuracy of the HBA-ANN model in forecasting air temperature.

Figure 6. Scatter plots of the observed and forecasted temperature along with R2, RMSE, 95% confidence interval, and the error histograms for the ANN (a), GEP (b), and HBA-ANN (b) models for three times ahead at the coldest place.

Figure 6. Scatter plots of the observed and forecasted temperature along with R2, RMSE, 95% confidence interval, and the error histograms for the ANN (a), GEP (b), and HBA-ANN (b) models for three times ahead at the coldest place.

In addition to the scatter plots, the ANN, GEP and HBA-ANN model’s performances were also evaluated using the Taylor diagrams (Figure ). This is an effective and popular way of comparing the models’ rankings. It gives a better picture of the resemblance of model results to that of observed data (Taylor, Citation2001). The similarity between the observed and forecasted results for three times ahead was quantified in terms of their RMSD, standard deviations, and correlation coefficient in the Taylor diagrams. The performances were analyzed by observing the positions of the models’ points on the respective plots. It is observed that the HBA-ANN models show better performance than the ANN and GEP models for all the times ahead based on near distance of its points to the observed points.

Figure 7. Taylor diagram of the models for T + 1, T + 2, and T + 3 times horizons for the coldest region.

Figure 7. Taylor diagram of the models for T + 1, T + 2, and T + 3 times horizons for the coldest region.

For visual comparison of the temporal variation of the temperatures, the time series analysis was applied (Figure ). According to Figure , it is worth mentioning that the forecasting values of the HBA-ANN models especially the extreme values are fitted well with the observed values in the testing period rather than the training period for all the times ahead. This may be due to the better generalization capability of the HBA algorithm, which is being utilized for optimizing the internal parameters of the ANN model development. Moreover, by increasing time ahead, the efficiency of all three models decreased.

Figure 8. Time series plots of the observed and forecasted temperature of the ANN (a), GEP (b), and HBA-ANN (c) models for all times ahead in the training and testing periods at the coldest region.

Figure 8. Time series plots of the observed and forecasted temperature of the ANN (a), GEP (b), and HBA-ANN (c) models for all times ahead in the training and testing periods at the coldest region.

Following the development of ANN, GEP, and HBA-ANN models, errors were calculated and stacked histogram of the errors were plotted (Figure ). The error values show negative, if the forecasted values are larger than observed and vice versa. The y-axis reflects the number of samples in each dataset and the zero-error line is equivalent to no error value. For T + 1 time horizon, it is interesting to note that there are smaller errors in the testing period than the training period for all three models. This is more evidently seen in the hybrid HBA-ANN model. Moreover, the total error range for the hybrid model (about [−4, 10]) is smaller than the ANN and GEP model. In addition, all the error plots show an approximately asymmetrical histogram with respect to the zero error. Similar observations can be seen for other times ahead as well. The results confirm that all three models are statistically significant, however better and superior performances can be seen for the HBA-ANN model for forecasting air temperature data at the coldest place. Because the forecasted values with HBA-ANN model is almost near to the actual values and there is very low error. Hence, the HBA-ANN model could forecast the future values of the air temperature with high accuracy, and so it can increase the efficiency of the climatic policies.

Figure 9. Error histogram plots of the ANN (a), GEP (b), and HBA-ANN (c) models for all times ahead in the training and testing periods at the coldest region.

Figure 9. Error histogram plots of the ANN (a), GEP (b), and HBA-ANN (c) models for all times ahead in the training and testing periods at the coldest region.

3.3. Models’ performance in the warmest region (Furnace Creek station)

Similar to the coldest region, different input combinations were considered for developing the ANN, GEP, and HBA-ANN models for forecasting air temperature at three times horizons (T + 1, T + 2, and T + 3) for the warmest region (Table ). The performance of the hybrid HBA-ANN models was assessed using the same error metrics (R2, RMSE, MAE, RSR, and NSE) and compared with the standalone ANN and GEP in Table . According to this Table, it is observed that, in most cases, the accuracy of the temperature forecasting increases with the number of input variables (time lags) for all models and all times ahead. In other words, the last input combination including lag times up to 3 (T, T-1, T-2, and T-3) is the best input combination for temperature forecasting in 1 month to 3 months ahead for the models.

Table 5. Error metrics of the ANN and HBA-ANN models at the warmest place.

According to Table , it can be seen that the hybrid HBA-ANN model outperforms the ANN and GEP models based on the low values of the RMSE, MAE, and RSR and high values of the R2 and NSE e.g. the R2, RMSE, and NSE values of the standalone ANN model for T + 1 time are 0.914, 2.724, 0.911 in the training period, and 0.940, 2.432, 0.934 in the testing period, respectively, while these values improves into 0.956, 1.954, 0.954 in the training period, and 0.976, 1.496, 0.975 in the testing period for the HBA-ANN hybrid model. Similar results can be seen for the other two times ahead (T + 2 and T + 3) for both of the ANN and HBA-ANN models. The performance of the HBA-ANN model is also better than the GEP model in both training and testing phases at all three horizons. High value for R2 and low values for RMSE in hybrid model compared to other ones indicated that the forecasted values with HBA-ANN model is almost near to the actual values which causes in less errors and more reliable results. Hence, the HBA-ANN model shows high capability than the classical ANN model and GEP model in air temperature forecasting based on the all statistical metrics especially in the testing phase. Based on Table, it also can be seen that the accuracy of temperature forecasting values is decreased by increasing times ahead for all models. It may be mentioned as one of the limitations of the proposed models. The best results of the ANN and the HBA-ANN models for T + 1, T + 2 and T + 3 times were considered for the further graphical analysis.

Table 6. Error metrics of the ANN, GEP and HBA-ANN models at the coldest place.

in order to evaluate the model’s performance in a more comprehensive manner for all times ahead, Figure  presents the scatter plots of the observed and forecasted temperature along with R2, RMSE, 95% confidence interval, and error histograms were provided for the best ANN, GEP and HBA-ANN models. The quantification of the uncertainty of the developed models is more ascertained by the results of the prediction intervals. These plots suggest superior forecasting capability of the HBA-ANN models, with R2 values ranging from 0.967 to 0.976, which are higher than those of the ANN model ranging from 0.869 to 0.939; and higher than GEP model ranging from 0.927 to 0.944 and with RMSE values ranging from 1.725 to 1.496, which are lower than those of the ANN and GEP models for all three times ahead. This point is clearly visible in the scatter plots drawn between the observed and modelled temperature data in terms of having lesser dispersions around the perfect line (i.e. y = x) for all three times ahead. The shapes of the error histograms are approximately similar to the normal distribution especially for the HBA-ANN models.

Figure 10. Scatter plots of the observed and forecasted temperature along with R2, RMSE, 95% confidence interval, and the error histograms for the ANN (a), GEP (b), and HBA-ANN (b) models for three times ahead at the warmest place.

Figure 10. Scatter plots of the observed and forecasted temperature along with R2, RMSE, 95% confidence interval, and the error histograms for the ANN (a), GEP (b), and HBA-ANN (b) models for three times ahead at the warmest place.

We also utilized the Taylor diagram to evaluate the ANN, GEP and HBA-ANN model’s performances (Figure ). Taylor diagram identified the similarity between the observed and forecasted results for three times ahead was quantified in terms of their RMSD, standard deviations, and correlation coefficient. The performances were analyzed by observing the positions of the models’ points on the respective plots. The model with low distance from observation point will have superior performance. It is observed that the HBA-ANN models show better performance than the ANN and GEP for all the times ahead based on near distance of their points to the observed points.

Figure 11. Taylor diagram of the models for T + 1, T + 2, and T + 3 times horizons for the warmest region.

Figure 11. Taylor diagram of the models for T + 1, T + 2, and T + 3 times horizons for the warmest region.

Furthermore, the time series analysis was applied for visual comparison of the temporal variation of the air temperatures at Furnace Creek station (Figure ). According to Figure 14, it is worth mentioning that the estimations of the HBA-ANN models especially the extreme values are fitted well with the observed values in the testing period rather than the training period for all the times ahead. This indicated the better generalization capability of the HBA algorithm, which is being used for optimizing the internal parameters of the ANN model development. Moreover, by increasing time ahead, the efficiency of the models (even HBA-ANN hybrid model) was be decreased. Therefore, although the hybrid model increases the performance of the neural network, its efficiency decreases in long-term predictions like the neural network model and the gene program.

Figure 12. Time series plots of the observed and forecasted temperature of the ANN (a), GEP (b), and HBA-ANN (c) models for all times ahead in the training and testing periods at the warmest region.

Figure 12. Time series plots of the observed and forecasted temperature of the ANN (a), GEP (b), and HBA-ANN (c) models for all times ahead in the training and testing periods at the warmest region.

Following the development of ANN, GEP and HBA-ANN models, errors were also calculated and stacked histogram of the errors were plotted (Figure ). For T + 1 time horizon, it is interesting to note that there are smaller errors in the testing period than the training period for all three models. This is more evidently seen in the hybrid HBA-ANN model. Moreover, the total error range for the hybrid model (about [−3, 6]) is smaller than the standalone ANN model and GEP. In addition, all the error plots show an approximately asymmetrical histogram with respect to the zero error. Similar observations cab been seen for other times ahead as well. Based on results, ANN, GEP, and HBA-ANN models are statistically significant, however better and superior performances are seen for the HBA-ANN model for forecasting air temperature time series at the Furnace Creek station.

Figure 13. Error histogram plots of the ANN (a), GEP (b), and HBA-ANN (c) models for all times ahead in the training and testing periods at the warmest region.

Figure 13. Error histogram plots of the ANN (a), GEP (b), and HBA-ANN (c) models for all times ahead in the training and testing periods at the warmest region.

Accurate air temperature forecasting is a critical key in human life, and agriculture, climatic and energy-saving policies. In general, the results indicated more precise performance of the HBA-ANN model in air temperature forecasting in both of the coldest and warmest regions compared to the ANN and GEP models. In other words, the HBA optimization algorithm has a high efficiency in improving the results of the standalone ANN model in both regions with different climate conditions. Our result is similar to the studies that indicated the superior performance of the hybrid model (with R2 > 0.95) compared to stand alone models (such as Kazemi et al., Citation2021; Khan & Maity, Citation2022; Li & Yang, Citation2022; Lin et al., Citation2021; Sari et al., Citation2022). The computational cost of the HBA-ANN model is similar to the ANN and is less than the GEP. Moreover, comparing the obtained results in both regions showed that error values of implemented models were less in the warmest region than those in the coldest region. It seems that in the cold regions, efficiency of the hydrological models may decrease in simulating phenomena.

4. Conclusion

In this study, a hybrid intelligent model contains the artificial neural network (ANN) hybridized with the powerful hetaeristic Honey Badger Algorithm (HBA-ANN) was proposed for monthly temperature forecasting for one to three times ahead in two different regions of the world with the hottest and coldest climate conditions. The average mutual information (AMI) technique was employed to find the optimal time delay values for the temperature variable for three-time horizons. An analysis of the performance of the hybrid model developed was compared with the classical ANN and GEP models using some statistical criteria (e.g. Nash–Sutcliffe Coefficient, Root-Mean Square Error, and Coefficient of Determination) and plots (e.g. Taylor and scatter diagrams). The important results obtained are as follows:

  1. The HBA-ANN model outperforms the ANN and GEP model in both the coldest and warmest regions of the world based on all the statistical criteria and plots.

  2. The HBA optimization algorithm performed well in optimizing the ANN’s parameters and so improving its accuracy in both regions.

  3. The ANN, GEP and HBA-ANN models’ efficiency was better in the hottest region than the coldest one.

  4. By increasing number of the input variables (i.e. lag times of the temperature variable), the precise of both AI models approximately increased.

  5. By increasing the times ahead values, the forecasting accuracy of standalone ANN, GEP, and hybrid models were decreased in both regions.

It is recommended to evaluate the HBA-ANN model capability in predicting temperature with other climatologically input variables such as wind speed in the same/other regions of the world. the capability of the model could be evaluated for hourly and daily data if it is available. Moreover, it is recommended to hybrid the HBA with other intelligent standalone models (e.g. SVM model) and to examine its performance in other hydrological and environmental parameters, such as air pollutant.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Funding

This work was supported by the National Natural Science Foundation of China (No.61862051), the Science and Technology Foundation of Guizhou Province (No. ZK[2022]549), the Natural Science Foundation of Education of Guizhou province (No. [2019]203, No. KY[2019]067), and the Funds of Qiannan Normal University for Nationalities (No.qnsy2019rc09).

References

  • Abramson, B., Brown, J., Edwards, W., Murphy, A., & Winkler, R. L. (1996). Hailfinder: A bayesian system for forecasting severe weather. International Journal of Forecasting, 12(1), 57–71. https://doi.org/10.1016/0169-2070(95)00664-8
  • Alomar, M. K., Khaleel, F., Aljumaily, M. M., Masood, A., Razali, S. F. M., AlSaadi, M. A., Al-Ansari, N., & Hamed, M. M. (2022). Data-driven models for atmospheric air temperature forecasting at a continental climate region. PLoS ONE, 17(11), e0277079. https://doi.org/10.1371/journal.pone.0277079
  • Anushka, P., Hazi, MD,A, & Upaka, R. (2020). Comparison of different artificial neural network (ANN) training algorithms to predict the atmospheric temperature in Tabuk, Saudi Arabia. MAUSAM, 71(2), 233–244. https://doi.org/10.54302/mausam.v71i2.22
  • Attar, N. F., Khalili, K., Behmanesh, J., & Khan Mohammadi, N. (2018). On the reliability of soft computing methods in the estimation of dew point temperature: The case of arid regions of Iran. Computers and Electronics in Agriculture, 153, 334–346. https://doi.org/10.1016/j.compag.2018.08.029
  • Band, S. S., Hegi, E., Bateni, S., Karimi, H., Rabiee, M., Samadianfard, S., Chau, K.W., & Mosavi, A. (2021). Groundwater level prediction in arid areas using wavelet analysis and Gaussian process regression. Engineering Applications of Computational Fluid Mechanics, 15(1), 1147–1158.
  • Begg, C. M., Begg, K. S., Du Toit, J. T., & Mills, M. G. L. (2003). Scent-marking behavior of the honey badger, mellivora capensis (mustelidae), in the southern Kalahari. Animal Behaviour, 66(5), 917–929. https://doi.org/10.1006/anbe.2003.2223
  • Begg, C. M., Begg, K. S., Du Toit, J. T., & Mills, M. G. L. (2005). Life-history variables of an atypical mustelid, the honey badger Mellivora capensis. Journal of Zoology, 265(1), 17–22. https://doi.org/10.1017/S0952836904005990
  • Chen, L., & Lai, X. (2011). Comparison between arima and ann models used in short-term wind speed forecasting, in: 2011 Asia-Pacific Power and Energy Engineering Conference, IEEE. pp. 1–4.
  • Chevalier, R. F., Hoogenboom, G., McClendon, R. W., & Paz, A. J. (2011). Support vector regression with reduced training sets for air temperature prediction: A comparison with artificial neural networks. Neural Computing and Applications, 20(1), 151–159. https://doi.org/10.1007/s00521-010-0363-y
  • Cifuentes, J., Marulanda, G., Bello, A., & Reneses, J. (2020). Air temperature forecasting using machine learning techniques: A review. Energies, 13(16), 4215. https://doi.org/10.3390/en13164215
  • Cofino, A. S., Cano Trueba, R., Sordo, C. M., & Guti 'errez Llorente, J. M. (2002). Bayesian networks for probabilistic weather prediction.
  • Curceac, S., Ternynck, C., Ouarda, T. B., Chebana, F., & Niang, S. D. (2019). Short-term air temperature forecasting using nonparametric functional data analysis and SARMA models. Environmental Modelling & Software, 111, 394–408. https://doi.org/10.1016/j.envsoft.2018.09.017
  • Deng, T., Duan, H. F., & Keramat, A. (2022). Spatiotemporal characterization and forecasting of coastal water quality in the semi-enclosed Tolo Harbour based on machine learning and EKC analysis. Engineering Applications of Computational Fluid Mechanics, 16(1), 694–712. https://doi.org/10.1080/19942060.2022.2035257
  • Elseify, M. A., Kamel, S., Abdel-Mawgoud, H., & Elattar, E. (2022). A novel approach based on honey badger algorithm for optimal allocation of multiple DG and capacitor in radial distribution networks considering power loss sensitivity. Mathematics, 10(12), 2081. https://doi.org/10.3390/math10122081
  • Ferreira, C. (2001a). Gene expression programming in problem solving. In: 6th Online World Conference on Soft computing in Industrial Applications (invited tutorial).
  • Ferreira, C. (2001b). Gene expression programming: A new adaptive algorithm for solving problems. Complex Systems, 13(2), 87–129. https://doi.org/10.48550/arXiv.cs/0102027
  • Handler, S. L., Reeves, H. D., & McGovern, A. (2020). Development of a probabilistic subfreezing road temperature nowcast and forecast using machine learning. Weather and Forecasting, 1845–1863. https://doi.org/10.1175/WAF-D-19-0159.1
  • Hanoon, M. S., Najah Ahmed, A., Zaini, N., Kumar, P., Sherif, M., Sefelnasr, A., & El-Shafie, A. (2021). Developing machine learning algorithms for meteorological temperature and humidity forecasting at terengganu state in Malaysia. Scientific Reports, 11(1), 18935. https://doi.org/10.1038/s41598-021-96872-w
  • Haykin, S. S. (1999). Neural networks: A comprehensive foundation. Prentice Hall.
  • Horenko, I., Klein, R., Dolaptchiev, S., & Schütte, C. (2008). Automated generation of reduced stochastic weather models i: Simultaneous dimension and model reduction for time series analysis. Multiscale Modeling & Simulation, 6(4), 1125–1145. https://doi.org/10.1137/060670535
  • Hossain, M., Rekabdar, B., Louis, S. J., & Dascalu, S. (2015). Forecasting the weather of nevada: A deep learning approach, in: 2015 international joint conference on neural networks (IJCNN), IEEE. pp. 1–6.
  • Hou, J., Wang, Y., Zhou, J., & Tian, Q. (2022). Prediction of hourly air temperature based on CNN–LSTM. Geomatics, Natural Hazards and Risk, 13(1), 1962–1986. https://doi.org/10.1080/19475705.2022.2102942
  • Kapner, D. J., Cook, T. S., Adelberger, E. G., Gundlach, J. H., Heckel, B. R., Hoyle, C. D., & Swanson, H. E. (2007). Tests of the gravitational inverse-square law below the dark-energy length scale. Physical Review Letters, 98(2), 021101. https://doi.org/10.1103/PhysRevLett.98.021101
  • Kazemi, S. M., Saffarian, M., & Babaiyan, V. (2021). Time series forecasting of air temperature using an intelligent hybrid model of genetic algorithm and neural network. Journal of Industrial and Systems Engineering, 13(3), 1–15. 20.1001.1.17358272.2021.13.3.1.3
  • Khan, M. I., & Maity, R. (2022). Hybrid deep learning approach for multi-step-ahead prediction for daily maximum temperature and heatwaves. Theoretical and Applied Climatology, 149(3-4), 945–963. https://doi.org/10.1007/s00704-022-04103-7
  • Khatibi, R., Ghorbani, M. A., Kashani, M. H., & Kisi, O. (2011). Comparison of three artificial intelligence techniques for discharge routing. Journal of Hydrology. https://doi.org/10.1016/j.jhydrol.2011.03.007.
  • Kuligowski, R. J., & Barros, A. P. (1998). Localized precipitation forecasts from a numerical weather prediction model using artificial neural networks. Weather and Forecasting, 13(4), 1194–1204. https://doi.org/10.1175/1520-0434(1998)013<1194:LPFFAN>2.0.CO;2
  • Lan, L., Lian, Z., & Pan, L. (2010). The effects of air temperature on office workers’ well-being, workload and productivity-evaluated with subjective ratings. Applied Ergonomics, 42(1), 29–36. https://doi.org/10.1016/j.apergo.2010.04.003
  • Li, G., & Yang, N. (2022). A hybrid SARIMA-LSTM model for Air temperature forecasting. Advanced Theory and Simulations. A Hybrid SARIMA-LSTM Model for Air Temperature Forecasting, 2200502. https://doi.org/10.1002/adts.202200502
  • Lin, M. L., Tsai, C. H. W., & Chen, C. K. (2021). Daily maximum temperature forecasting in changing climate using a hybrid of multi-dimensional complementary ensemble empirical mode decomposition and radial basis function neural network. Journal of Hydrology: Regional Studies, 38(2021), 100923. https://doi.org/10.1016/j.ejrh.2021.100923
  • Liu, J. N., Hu, Y., He, Y., Chan, P. W., & Lai, L. (2015). Deep neural network modeling for big data weather forecasting, in: Information granularity, Big data, and computational intelligence (pp. 389–408). Springer.
  • Marchuk, G. (2012). Numerical methods in weather prediction. Elsevier.
  • McGovern, A., John Gagne, D., Troutman, N., Brown, R. A., Basara, J., & Williams, J. K. (2011). Using spatiotemporal relational random forests to improve our understanding of severe weather processes. Statistical Analysis and Data Mining: The ASA Data Science Journal, 4(4), 407–429. https://doi.org/10.1002/sam.10128
  • Mcgovern, A., Supinie, T., Gagne, I., Collier, M., Brown, R., Basara, J., & Williams, J. (2010). Understanding severe weather processes through spatiotemporal relational random forests, in: 2010 NASA conference on intelligent data understanding (to appear).
  • Mirarabi, A., Nassery, H. R., Nakhaei, M., Adamowski, J., Akbarzadeh, A. H., & Alijani, F. (2019). Evaluation of datadriven models (SVR and ANN) for groundwater-level prediction in confined and unconfined systems. Environmental Earth Sciences, 78(15), 489. https://doi.org/10.1007/s12665-019-8474-y
  • Mittelman, R., Kuipers, B., Savarese, S., & Lee, H. (2014). Structured recurrent temporal restricted boltzmann machines (pp. 1647–1655). International Conference on Machine Learning.
  • Mohsenzadeh Karimi, N., Kisi, O. Porrajabali, M., Rouhani-Nia, M., and Shiri, J. (2018). Evaluation of the support vector machine, random forest and geo-statistical methodologies for predicting long-term air temperature. ISH Journal of Hydraulic Engineering, 26(4), 376–386. https://doi.org/10.1080/09715010.2018.1495583
  • Moriasi, D. A., Arnold, J. G., Van Liew, M. W., Bingner, R. L., Harmel, R. D., & Veith, T. L. (2015). Model evaluation guidelines for systematic quantification of accuracy in watershed simulations. American Society of Agricultural and Biological Engineers, 50(3), 885–900.
  • Moravej, M., Amani, P., & Hosseini-Moghari, S. M. (2020). Groundwater level simulation and forecasting using interior search algorithm-least square support vector regression (ISA-LSSVR). Groundw Sustain Dev, 11, 100447. https://doi.org/10.1016/j.gsd.2020.100447
  • Naganna, S. R., Deka, P. C., Ghorbani, M. A., Biazar, S. M., Al-Ansari, N., & Yaseen, Z. M. (2019). Dew point temperature estimation: Application of artificial intelligence model integrated with nature-inspired optimization algorithms. Water, 11(4), 742. https://doi.org/10.3390/w11040742
  • Nayak, P. C., Rao, V. R. S., & Sudheer, K. P. (2006). Groundwater level forecasting in shallow aquifer using artificial neural network approach. Water Resources Management, 20(1), 77–90. https://doi.org/10.1007/s11269-006-4007-z
  • Radhika, Y., & Shashi, M. (2009). Atmospheric temperature prediction using support vector machines. International Journal of Computer Theory and Engineering, 1, 55. https://doi.org/10.7763/IJCTE.2009.V1.9
  • Rajendra, P., Murthy, K. V. N., Subbarao, A., & Boadeh, R. (2019). Use of ANN models in the prediction of meteorological data. Modeling Earth Systems and Environment, 5(3), 1051–1058. https://doi.org/10.1007/s40808-019-00590-2
  • Rasnopolsky, V. M., & Fox-Rabinovitz, M. S. (2006). Complex hybrid models combining deterministic and machine learning components for numerical climate modeling and weather prediction. Neural Networks, 19(2), 122–134. https://doi.org/10.1016/j.neunet.2006.01.002
  • Richardson, L. F. (2007). Weather prediction by numerical process. Cambridge university press.
  • Roy, D. S. (2020). Forecasting the Air temperature at a weather station using deep neural networks. Procedia Computer Science, 178, 38–46. https://doi.org/10.1016/j.procs.2020.11.005
  • Salcedo-Sanz, S., Deo, R., Carro-Calvo, L., & Saavedra-Moreno, B. (2016). Monthly prediction of air temperature in Australia and New Zealand with machine learning algorithms. Theoretical and Applied Climatology, 125(1-2), 13–25. https://doi.org/10.1007/s00704-015-1480-4
  • Sari, Y., Arifin, Y. F., Novitasari, N., & Faisal, M. L. (2022). Deep learning approach using the GRU-LSTM hybrid model for Air temperature prediction on daily basis. International Journal of Intelligent Systems and Applications in Engineering, 10(3), 430–436.
  • Schulte, P. A., Bhattacharya, A., Butler, C. R., Chun, H. K., Jacklitsch, B., Jacobs, T., Kiefer, M., Lincoln, J., Pendergrass, S., Shire, J., et al. (2016). Advancing the framework for considering the effects of climate change on worker safety and health. Journal of Occupational and Environmental Hygiene, 13(11), 847–865. https://doi.org/10.1080/15459624.2016.1179388
  • Shabani, E., Hayati, B., Pishbahar, E., Ghorbani, M. A., & Ghahremanzadeh, M. (2021). A novel approach to predict CO2 emission in the agriculture sector of Iran based on Inclusive Multiple Model. Journal of Cleaner Production, 279(2021), 123708. https://doi.org/10.1016/j.jclepro.2020.123708
  • Singh, V. K., Panda, K., Sagar, A., Al-Ansari, N., Duan, H. F., Paramaguru, P. K., Vishwakarma, D. K., Kumar, A., Kumar, D., Kashya, P. S., Singh, R. M., & Elbeltagi, A. (2022). Novel genetic algorithm (GA) based hybrid machine learning-pedotransfer function (ML-PTF) for prediction of spatial pattern of saturated hydraulic conductivity. Engineering Applications of Computational Fluid Mechanics, 16(1), 1082–1099. doi: 10.1080/19942060.2022.2071994
  • Smith, B. A., McClendon, R. W., & Hoogenboom, G. (2005). An enhanced artificial neural network for air temperature prediction IEC (Prague) 1:7–12.
  • Sutskever, I., Hinton, G. E., & Taylor, G. W. (2009). The recurrent temporal restricted boltzmann machine (pp. 1601–1608). Advances in neural information processing systems.
  • Taylor, K. E. (2001). Summarizing multiple aspects of model performance in a single diagram. Journal of Geophysical Research: Atmospheres, 106(D7), 7183–7192. https://doi.org/10.1029/2000JD900719
  • Voyant, C., Muselli, M., Paoli, C., & Nivet, M. L. (2012). Numerical weather prediction (nwp) and hybrid arma/ann model to predict global radiation. Energy, 39(1), 341–355. https://doi.org/10.1016/j.energy.2012.01.006
  • Xingjian, S., Chen, Z., Wang, H., Yeung, D. Y., Wong, W. K., & Woo, W. C. (2015). Convolutional lstm network: A machine learning approach for precipitation now casting, in: Advances in neural information processing systems, pp. 802–810.
  • Yakut, E., & Süzülmüş, S. (2020). Modelling monthly mean air temperature using artificial neural network, adaptive neuro-fuzzy inference system and support vector regression methods: A case of study for Turkey. Network: Computation in Neural Systems, 31(1-4), 1–36. https://doi.org/10.1080/0954898X.2020.1759833
  • Yang, Q., Lee, C. Y., & Tippett, M. K. (2020). A long short-term memory model for global rapid intensification prediction. Weather and Forecasting, 35(4), 1203–1220. https://doi.org/10.1175/WAF-D-19-0199.1
  • Zhang, G., Band, S. S., Ardabili, S., Chau, K. W., & Mosavi, A. (2022). Integration of neural network and fuzzy logic decision making compared with bilayered neural network in the simulation of daily dew point temperature. Engineering Applications of Computational Fluid Mechanics, 16(1), 713–723. https://doi.org/10.1080/19942060.2022.2043187
  • Zhu, S., Hadzima-Nyarko, M., Gao, A., Wang, F., Wu, J., & Wu, S. (2019). Two hybrid data-driven models for modeling water-air temperature relationship in rivers. Environmental Science and Pollution Research, 26(12), 12622–12630. https://doi.org/10.1007/s11356-019-04716-y