Figures
Abstract
This paper proposes a model called X-LSTM-EO, which integrates explainable artificial intelligence (XAI), long short-term memory (LSTM), and equilibrium optimizer (EO) to reliably forecast solar power generation. The LSTM component forecasts power generation rates based on environmental conditions, while the EO component optimizes the LSTM model’s hyper-parameters through training. The XAI-based Local Interpretable and Model-independent Explanation (LIME) is adapted to identify the critical factors that influence the accuracy of the power generation forecasts model in smart solar systems. The effectiveness of the proposed X-LSTM-EO model is evaluated through the use of five metrics; R-squared (R2), root mean square error (RMSE), coefficient of variation (COV), mean absolute error (MAE), and efficiency coefficient (EC). The proposed model gains values 0.99, 0.46, 0.35, 0.229, and 0.95, for R2, RMSE, COV, MAE, and EC respectively. The results of this paper improve the performance of the original model’s conventional LSTM, where the improvement rate is; 148%, 21%, 27%, 20%, 134% for R2, RMSE, COV, MAE, and EC respectively. The performance of LSTM is compared with other machine learning algorithm such as Decision tree (DT), Linear regression (LR) and Gradient Boosting. It was shown that the LSTM model worked better than DT and LR when the results were compared. Additionally, the PSO optimizer was employed instead of the EO optimizer to validate the outcomes, which further demonstrated the efficacy of the EO optimizer. The experimental results and simulations demonstrate that the proposed model can accurately estimate PV power generation in response to abrupt changes in power generation patterns. Moreover, the proposed model might assist in optimizing the operations of photovoltaic power units. The proposed model is implemented utilizing TensorFlow and Keras within the Google Collab environment.
Citation: Rizk-Allah RM, Abouelmagd LM, Darwish A, Snasel V, Hassanien AE (2024) Explainable AI and optimized solar power generation forecasting model based on environmental conditions. PLoS ONE 19(10): e0308002. https://doi.org/10.1371/journal.pone.0308002
Editor: Upaka Rathnayake, Atlantic Technological University, IRELAND
Received: March 12, 2024; Accepted: July 16, 2024; Published: October 2, 2024
Copyright: © 2024 Rizk-Allah et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: benchmark [36] https://www.kaggle.com/datasets/anikannal/solar-power-generation-data?select=Plant_1_Generation_Data.csv.
Funding: The author(s) received no specific funding for this work.
Competing interests: The authors have declared that no competing interests exist.
List of abbreviations: Definitions, Acronyms; AI, Artificial Intelligence; ANNs, Artificial Neural Networks; ARMA, Autoregressive Moving Average; LSTM, Long Short-Term Memory; BiLSTM, Bidirectional LSTM; BPNN, Back-propagation Neural Network; CNN, Convolutional Neural Network; COV, Coefficient of Variation; DL, Deep Learning; DT, Decision tree; EC, Efficiency Coefficient; EO, Equilibrium Optimizer; KNN, k-nearest neighbor; LIME, Local Interpretable and Model-independent Explanation; LR, Linear regression; MAE, Mean Absolute Error; MPPT, Maximum Power Point Tracking; PCC, Pearson Correlation Coefficient; PSO, Particle Swarm Optimization; PV, Photovoltaic; R2, R-squared; RMSE, Root Mean Square Error; RNN, Recurrent Neural Network; SVM, Support Vector Machine; XAI, Explainable Artificial Intelligence; GWO, Grey Wolf Optimization; nRMSE, Normalized RMSE; nMAE, Normalized MAE; RE, Relative Error
1. Introduction
The worldwide development of different energy resources and increasing energy demand due to industrialization and the growing global population have raised the world’s need for electrical power generated [1]. Photovoltaic (PV) power units represent the mainstream of renewable energy technologies due to the characteristics of solar energy, such as being inexhaustible, clean, free-pollution, and environment-friendly. Therefore, high-tech countries worldwide have concentrated on spending on research and development while providing incentives to promote solar PV systems [2]. PV power unit entails the direct conversion of solar energy into electrical energy. When a semiconductor is exposed to sun radiation (n-and p-type silicon), electricity is produced as electrons flow between electrodes. Although the PV power plant is simpler to construct than a fossil fuel power plant, the PV power plant can be affected by the construction site, timing, size, and panel capability [3]. In addition, the electricity generated by the PV plant can fluctuate sporadically due to Unforeseeable and unmanageable meteorological factors which include solar radiation, temperature, humidity, wind speed, and cloud cover. Significant fluctuations in temperature and solar radiation can have a substantial effect on energy production [4]. Due to of the nature of these variables, PV power generation may become unstable with causing a reduction in PV output power or a sudden surplus. Moreover, this might lead to an imbalance between generating power and load demand, affecting the power grid’s ability to operate and control [5]. If electricity generation is precisely forecasted, operation optimization techniques, like peak trimming and reducing the system’s uncertainty for power generation, can be effectively adopted [6]. Therefore, a method for precisely forecasting the amount of produced energy is vital for industrial power system applications [7]. Precise forecasting is vital for improving the level of electricity delivered to a grid and reducing the costs associated with the general variability [8]. Additionally, it can be employed for a variety of operation and control tasks such as power scheduling in transmission and distribution grids [9].
Over the past few decades, researchers and engineers have been promoting the advantages of recent innovations in data science, machine learning, and artificial neural networks (ANNs) for predicting the power generated from photovoltaics. In this regard, the forecasting approaches can be categorized as physical methods, artificial intelligence-based methods, statistical methods, and ensemble methods [10]. Artificial intelligence (AI) approaches have the potential to be valuable tools for predicting solar power generation. This is because they can address the complex relationship between input and output data, which is nonlinear in nature. The primary techniques for short-term predictions include linear regression, autoregressive moving average (ARMA), support vector machine (SVM), time series modelling, and back-propagation neural network (BPNN), among others. Linear regression requires a substantial dataset, and the accuracy of the fitting results might be influenced by pathological data [11]. Auto-regressive integrated moving average (ARIMA) models rely only on past power outputs, which may lead to significant inaccuracies in predictions [12]. The SVM approach is not capable of efficiently handling huge volumes of data in terms of both training time and predicting accuracy [13]. Furthermore, the procedure of selecting the kernel functions is challenging due to its greater suitability for categorization [14]. In order to get a higher convergence rate, it is necessary to enhance the algorithm of a conventional Backpropagation Neural Network (BPNN) [15]. In addition, the Markov chain relies on a large dataset, yet it may still perform well even when there is missing data [16]. The solar power forecasting task has previously used the k-nearest neighbor (KNN) machine learning technique [17]. Boosting, bagging, and regression trees are other machine learning algorithms that have shown high accuracy and effectiveness.
The field of deep learning has gained significant attention due to its relevance in renewable power forecasting, specifically in wind power forecasts. However, it has been noticed that many ensemble models employed in previous studies do not incorporate deep learning (DL) techniques such as long short-term memory (LSTM) or gated recurrent unit (GRU) networks [18]. Moreover, Furthermore, these models may suffer from lower accuracy as a result of the limitations of traditional optimization techniques included into them to acquire the optimal internal parameters, such as being stuck in local minima and subsequently acquiring suboptimal parameters. Thus, this paper overcomes these issues by integrating the LSTM with the EO algorithm into the proposed model, which is then applied to accurately depict the relationship between solar output power and environmental factors.
Recently, there has been a growing interest in using deep learning models for data mining, regression, and feature extraction due to their capabilities [19]. The prevalent deep learning models utilized for predicting solar power generation comprise the deep neural network (DNN), Boltzmann machines, recurrent neural network (RNN), and deep belief network (DBN). RNN has emerged as the favored alternative for performing predictions in smart grids [20]. LSTM, a specialized form of RNN, has been utilized in research studies to enhance predicting accuracy when compared with standard ANN models [21]. Authors in [22] proposed a deep LSTM-RNN model for precise prediction of solar power output. While LSTM exhibits a significant level of predictive accuracy, it is characterized by a lengthy training duration. In [23], Authors suggested an integrated framework utilizing convolutional neural network (CNN) and bidirectional LSTM (BiLSTM) to precisely predict the energy output of a short-term photovoltaic system. After evaluating the model’s accuracy, they concluded that the suggested CNN-BiLSTM model exhibits a much higher predictive influence compared to both the CNN and BiLSTM models. Nevertheless, the model is still subpar in terms of prediction accuracy, which may be caused by the smaller characteristics of the input data. Authors in [24] suggested a hybrid model using the full wavelet packet decomposition (FWPD) and the BiLSTM, named FWPD-BiLSTM, to estimate a day ahead solar irradiance. The FWPD-BiLSTM model has been demonstrated to be a highly effective forecasting model for improving the performance of solar irradiance predictions. Nevertheless, the optimization of hyper-parameters and the increased duration of execution represent significant hurdles in the implement of the suggested model. Study [25] examined eleven distinct forecasting models for point and interval forecasting of solar global horizontal irradiance (GHI) on an hourly basis, specifically for two locations in India. After investigating the model’s accuracy, they observed that the BiLSTM model surpasses all individual models in terms of getting lower values for RMSE and MAE. Nevertheless, the study was challenged by the intricate hyper-parameter selection method and the significant amount of time required for execution. In [26], Authors offered a hybrid deep learning approach based on a robust local mean decomposition (RLMD) algorithm and the BiLSTM, named RLMD-BiLSTM, for accurate forecasting of solar GHI. The proposed hybrid model showed good accuracy in terms of RMSE and MAE over various contrast models, but hyper-parameter adjustment was selected by a grid search method, which is a time-consuming process. Moreover, it lacks the impact of combining other hyper-parameters like lag size, batch size, and drop period rate. Study [27] proposed a novel deep learning model for predicting solar power generation. The model includes data preprocessing, kernel principal component analysis, feature engineering, calculation, GRU model with time-of-day clustering, and error correction post processing. The findings of the experiments have shown that the suggested model exhibits superior forecasting accuracy compared to other conventional models and can produce outstanding prediction outcomes. Authors in [28] proposed a deep learning-based approach and a pre-processing algorithm to predict solar power. The reported results of the LSTM approach with adaptive moment estimation (ADAM) and root mean square propagation (RMSP) show a good fit compared to other approaches. In [29], Authors provided a novel method for predicting global horizontal irradiance that is based on the LSTM and back propagation (BP), named LSTM-BP model, and the multi-physical process of atmospheric optics. The suggested model was compared to the LSTM model using comparable time scales and meteorological conditions. The suggested approach outperforms the LSTM model in clear, cloudy, and partly cloudy circumstances. The suggested approach improves prediction accuracy and expands its applicability. Authors in [30] presented a hybrid model based on deep learning techniques incorporating CNN and LSTM to forecast the short-term PV power generation at different times ahead. The proposed hybrid CNN-LSTM auto encoder approach surpasses the existing models in the literature in terms of the RMSE and MAE metrics. The suggested hybrid model achieves much lower values, varying from 40% to 80%, in comparison to other models documented in the literature. The extent of the reduction depends on the predicting interval. Authors in [31] suggested a LSTM model that is more effective at extracting temporal information compared to other deep learning models. This model is specifically designed to predict solar radiation data. The newly introduced model is referred to as the Read-first LSTM (RLSTM) model. The primary novelty of this study is the development of an enhanced LSTM model for predicting solar radiation data and the establishment of a collaborative procedure amongst gates. The provided findings indicate that the RLSTM model decreased the centralized RMSE of the BiLSTM, LSTM, RNN, and radial basis function neural network (RBFNN) models by 30%, 60%, 67%, and 70% correspondingly. The RLSTM, BiLSTM, LSTM, RNN, and RBFNN models had correlation coefficients of 0.99, 0.98, 0.96, 0.95, and 0.93, respectively. However, the RLSTM model necessitates the optimizer to tweak its hyper-parameters in order to further increase its accuracy. Authors in [32] proposed an innovative approach for predicting solar GHI 24 hours in advance by utilizing information from nearby geographic areas. The suggested methodology encompasses feature selection, data pre-processing, the utilization of Convolutional Long Short-term Memory (ConvLSTM) for feature extraction, and the implementation of a fully connected neural network regression model. The proposed method surpasses all other examined methods in terms of correlation coefficient and RMSE. Furthermore, the suggested model demonstrates superior performance compared to existing approaches. This confirms the effective attainment of the research objectives in forecasting solar GHI. However, the structural parameters of ConvLSTM lack optimization using evolutionary algorithms or other optimization techniques, which could enhance the accuracy of predictions. Authors in [33] introduced a hybrid model that incorporates an attention mechanism with the CNN and BiLSTM, named CNN-BiLSTM-Attention, for a short-term photovoltaic power prediction. This approach seeks to reduce the negative effects of weather variability on the precision of PV power prediction by efficiently extracting important characteristics from multidimensional time series data. The results confirm that the CNN-BiLSTM-Attention model provides outstanding performance compared to other models, but its performance is reliant upon a significant volume of training data, and the intricate nature of the model demands significant computational resources. Authors in [34] introduced a novel approach for PV power forecasting, combining federated learning (FL) and transfer learning (TL) in a hybrid deep learning model called Federated Transfer Learning Convolutional Neural Network with Stacked Gated Recurrent Unit (FL-TL-Conv-SGRU). This model addresses data privacy and security concerns while optimizing forecasting performance. Using a bio-inspired Orchard Algorithm (OA) for hyperparameter tuning and eight diverse PV datasets, the FL-TL-Conv-SGRU model trains in a federated manner, enhancing generalization and predictive capabilities. Empirical results show the model outperforms traditional methods, offering accurate forecasts and efficient, sustainable energy management while adhering to data protection regulations. Authors in [35] proposed a composite model for short-term wind and PV power prediction, integrating LSTM and swarm intelligence algorithms to improve forecasting accuracy. This model leverages the Coati optimization algorithm (COA) to enhance hyperparameters of CNN-LSTM, leading to improved learning rates and performance. The results show a significant reduction in RMSE for day-ahead and hour-ahead predictions by 0.5% and 5.8%, respectively. The proposed COA-CNN-LSTM model outperforms existing models such as GWO-CNN-LSTM, LSTM, CNN, and PSO-CNN-LSTM, achieving nMAE of 4.6%, RE of 27%, and nRMSE of 6.2%. It also excels in the Nash-Sutcliffe metric analysis and Granger causality test, with scores of 0.98, and 0.0992, respectively. Experimental outcomes demonstrate the model’s effectiveness in providing accurate wind power predictions, aiding in the efficient management of renewable energy systems and contributing to the advancement of clean energy technology. Authors in [36] propose a Hybrid Deep Learning Model (DLM) for enhancing PV power output forecasting under dynamic environmental conditions. This model combines CNN, LSTM, and Bi-LSTM to capture spatial and temporal dependencies in weather data. Using the Kepler Optimization Algorithm (KOA) for hyperparameter tuning and Transductive Transfer Learning (TTL) for resource efficiency, the model is trained on diverse PV site datasets. Evaluations show the hybrid DLM outperforms individual models in short-term PV power forecasting, demonstrating superior accuracy and resilience, making it effective for PV power plant management. However, there is room for improving the prediction accuracy of solar PV power while ensuring the stability of micro grid operation by investigating more robust deep learning models. Moreover, these studies highlight the potential of integrating the LSTM with different AI architectures to enhance solar power forecasting. However, further research is needed to explore the explanation aspect of the deep learning models as well as optimize the intricate hyper-parameters of the model to improve prediction performance. Table 1 summarizes some of previous works for solar prediction systems based on AI tools.
Responding to the issues raised in the studies in order to boost solar power prediction accuracy and guarantee micro grid operation reliability. This paper proposes a solar power prediction model based on LSTM architecture and EO algorithm, called X-LSTM-EO. The proposed X-LSTM-EO model operates in two stages. The first one employs the LSTM to learn power generation trends based on the environmental conditions and then predict the generating energy, while the second stage which is using the EO algorithm that aims to optimize hyper- parameters for the deep learning model, including the number of LSTM cells, the choice of activation function (such as sigmoid, SoftMax, tanh, etc.), and the type of optimizer function (such as Adam, RMSprop, etc.), are all important components used in training a neural network. The proposed X-LSTM-EO scheme is trained and tested with the help of the power plant’s PV power output. Because the solar palettes could have lot of issues, Local Interpretable and Model-agnostic Explanation (LIME), an approach for explainable artificial intelligence (XAI), is used to identify the critical conditions for predicting power generation in a smart solar system. The accuracy of the proposed deep learning model was compared and verified with other models in terms of several metrics, including R2, RMSE, COV, MAE, and EC. Results indicate that this approach enhances forecasting accuracy and outperforms the compared models in forecasting efficacy.
The main contribution and the novelty of this paper is summarized as follows:
- Deep learning models might not be as accurate because they use traditional optimization methods to find the best internal parameters. These techniques can get stuck in local minima, which leads to finding parameters that aren’t as good as they could be. So, this paper solves these problems by adding the LSTM and the EO algorithm to the proposed model. This model is then used to show correctly how solar output power is related to external factors.
- Applying the EO algorithm for tuning the hyper-parameters of the LSTM to enhance the performance of the forecasting.
- Applying PSO optimizer for comparing its results with EO optimizer.
- The utilization of LSTM for effective exploration of the search space without being trapped in local optima areas.
- To understand the forecasting results. XAI’s approach called LIME has been applied to explain the obtained results and performance of the proposed deep learning model.
- The XAI explained the most important environmental condition that affects the model’s forecasting results.
- The propped X-LSTM-EO model proposes a common, accurate model that predicts well under many environmental scenarios. It mitigates PV power generation unpredictability and safely integrates large-scale PV power generation into micro grids, lowering operational costs and boosting efficiency and safety.
The following sections outline the rest of the paper. Section 2 provides an overview of the materials and methods used. The dataset description and analysis are illustrated in Section 3. The proposed X-LSTM-EO model is presented in Section 4. Section 5 presents and analyzes the experimental results. Finally, Section 6 summarizes the conclusions and presents the future works.
2. Preliminaries
This section provides the basic concepts regarding the LSTM, EO, and Locally Interpretable Model Agnostic Explanations (LIME)
2.1 LSTM (Long short-term memory)
The LSTM is categorized as a type of RNN, which is a potent type of artificial neural network that has the capacity to store input data in memory internally. Because of this characteristic, RNNs are particularly effective in addressing problems that involve sequential data, such as time series. However, a major problem that RNNs frequently experience is known as vanishing gradient, which causes the learning process of the model to become extremely slow or even stop altogether [29]. In order to anticipate a time series’ future patterns, its previous data is crucial. The time series’ historical data is encoded using the LSTM. Long-term memory functionality in the LSTM model may help with long-term sequence modeling’s gradient vanishing and exploding issues. We feed the feature vector xt into the LSTM model at time step t. formally; the calculations are carried out by the LSTM model as follows Eqs (1)–(6):
(1)
(2)
(3)
(4)
(5)
(6)
The architecture of the LSTM network includes various components such as input gates, forget gates, output gates, and unit states. A depiction of the network’s fundamental structure is presented in Fig 1 [30]. The hidden state in this case, ht−1, contains all the data up to the (t -1)th time step. Concatenation of ht-1 and xt produces the forget gate ft, input gate it, and output gate ot, respectively. To create a candidate cell state that symbolizes the newly added information, ht−1 and xt are also employed. Then, ct is created by combining ct−1 and
, with ft acting as the procedure’s balance factor. To output the current hidden state, ht, ot is finally multiplied by ct. Wf, Wi, Wo, and Wc are the parameters to be learned, ʘ represents the Hadamard product, σ(.) and tanh(.) are the sigmoid and tanh activation functions, respectively.
2.2 Basics of the EO
EO is a new meta-heuristic method that was suggested by Faramarzi [37] based on physics concept for handling engineering optimization problems. EO simulates the dynamic and equilibrium states that achieve the control volume mass balance models. EO is mathematically composed of search agents to define the particles (solutions) associated with their concentrations (positions). In this regard, the search agent renewed its concentration by choosing one of the best-so-far solutions at random (i.e., equilibrium candidate) to ultimately acquire the optimal outcome (i.e., equilibrium state). Furthermore, EO utilizes the “generation rate” step to boost the search skills in terms of explorative and exploitative trends while avoiding the stuck-in local optima dilemma. The mathematical optimization framework of EO is expressed by the following steps and its algorithm is shown in Algorithm (1).
- Create an initial population of N particles at random as follows:
(7)
Where N denotes population size, Δlower and Δupper denote respectively the lower and upper bounds of the search region, ri stands for a uniform random vector generated inside the interval [0, 1].Defines the initial position of the ith particle.
- Construct the equilibrium candidates (
) as in Eq (8) by adding the best four particles along with their average to
at aiming to enhance the exploration and exploitation capabilities of EO.
(8)
- Renew the concentration of the particle through following one of the equilibrium candidates chosen at random from
as follows.
(9)
Wheredefines an exponential equation that controls the balance among the explorative and exploitative features, and it is formulated as follows.
(10)
Wheredefines turnover vector which is composed of random numbers inside the interval [0, 1], t denotes the time which decreases gradually with iterations, and it is expressed as follows.
(11)
Moreover, the initial start time (t0) is formulated by the following equation.(12) where Max_iter denotes the maximum iterations, a1 stands for parameter that controls the exploration feature while a2 signifies a parameter that control the exploitation skill,
is a vector contains random values ranged from 0 to 1. Additionally,
denotes the generation rate parameter that aids to further improve the exploitation feature and it is formulated as follows.
(13)
Wherewhere r1 and r2 stand for arbitrary numbers generated using the uniform distribution ranged from 0 to 1; GP denotes the generation probability and it is set to 0.5 to acquire a better balance among the explorative and exploitative skills;
signifies the generation rate; V is set to a unit.
The concentration update formula is expressed as follows:(14)
Algorithm 1: Pseudo-code of the proposed EO algorithm
1: Define the algorithm’ parameters: a1 = 2; a2 = 1; Iter = 0 (counter); GP = 0.5
2: Create an initial population at random comosed of N particles,
3: For minimization problem, set large value for the fitness of the equilibrium candidates ()
4: While Iter<MaxIter do
5: For i = 1: N
6: Evaluate the fitness of the particle (f(Δi))
7: If , then
and
8: Elseif and
, then
and
9: Elseif &
and
, then
and
10: Elseif &
&
and
, then
and
11: End If.
12: End for.
13: .
14: Construct the equilibrium pool,
15: Renew the time
16: For i = 1: N
17: Select one equilibrium candidate at random from equilibrium pool
18: Generate the random vectors of
19: Update
20: Perfrom
21: Constitute
22: Constitute
23: Renew the postion of particle as:
24: End for
25: Iter = Iter+1
26: End While
27: Output: disply
2.3 Explanation AI based Locally Interpretable Model Agnostic Explanations (LIME)
The technique of "Agnostic Explanations" is a post-hoc, model-agnostic explanation approach that aims to provide interpretability to any black box machine learning model by creating a local, interpretable model for each prediction. LIME is independent of the classifier’s algorithm; the authors advise using it to explain any classifier [38]. LIME predicts locally and provides explanations for each observation. LIME fits a local model using similar data points to explain the observation. Linear models, decision trees, and others can be used as local models. The LIME explanation φ(x) at the point x produced by an interpretable model g can be expressed as:
(15)
Where G represents the class of interpretable models. The explanatory model x, for example, is a model g that minimizes losses like the sum of squared errors. This is the loss L, which shows how well the forecasts of the original model f can be explained.
πx: defines the neighborhood’s size in terms of, for example, x.
Ω(g): shows how complex this model is and suggests that the feature amount should be reduced.
The objective is to minimize the locality aware loss function L without making any assumptions about the function f, as LIME is designed to be model agnostic. The measure of how accurate g is in approximating f in the defined locality is captured by πx.
3. Dataset description and analysis
In this paper, the data was collected at two solar power plants in India over 34 days [39]. It has a pair of files contains a dataset on power generation and a dataset on sensor readings.
The dataset on Sensor readings includes time and date Observations recorded at 15-minute intervals, it has; "Plant ID", "SOURCE KEY", "AMBIENT TEMPERATURE", "MODULE TEMPERATURE", and "IRRADIATION". The samples of data are listed in Table 2.
While Power-generation data includes Date and time for each observation taken at 15-minute intervals, it has; Plant ID (common to the file), SOURCE KEY sort—The source key in this file represents the inverter id, DC_POWER, AC_POWER, DAILY_YIELD, and TOTAL_YIELD, the samples of data are listed in Table 3.
As noted, the database contains two files which are power generation data and sensor readings data. Table 4 illustrates the statistical analysis of the entire dataset. Scatter plots are used to monitor and describe the relationship between data aspects. Scatter plots show dataset trends and individual data values. These data can establish correlations [40]. Hence, a scatterplot diagram displays dataset data. Scatter plots depict each pair of attributes as dispersion plots. Scatter plots reveal the strongest and weakest associations. This lets us explain each feature’s relationship.
Fig 2 shows the solar energy dataset scatterplot graphs. Scatter graphs correlated scatter plots differently. With 23 days’ worth of data on solar power generation, the data visualization is used to spot faults and abnormalities in solar power plant output. Fig 3 illustrates that the DC POWER generation per day graph shows that the amount of power made by the sun changes from day to day. On some days, there is less change in how much DC POWER is made. On the other days, the amount of DC POWER produced goes up and down a lot.
The daily DC POWER generation statistic indicates the average daily power generation. Fig 4(A) shows 2020-05-25 has the highest average DC POWER generation and 2020-05-18 the lowest. A system fault or changing weather may explain this large DC POWER generation mismatch. DC POWER days are shown here. Irradiation histograms mirror daily DC power generation. Solar power stations’ DC power comes from the sun. Radiation impacts generation. Radiation Fig 4(B) displays the average daily irrigation compared to Fig 4(A). 2020-05-25 has the most radiation, 2020-05-18 the least. DC POWER and IRRADIATION graphs are near-perfect. The sky is cloudless because radiation, solar panel temperature, and ambient temperature are similar (Fig 4(C)). Rain, clouds, and bad weather likely caused this decline. It’s unlikely. Since the amount of energy generated from the solar panel is affected by environmental factors, our proposed model not only predicts this amount but also explains the reasons for these findings. The Correlation between variables is a well-known measure of similarity between two random variables. Pearson correlation coefficient (ρcc) is a measure of the degree to which two random variables are dependent on one another [41].
daily power tracing, (a) daily DC power, (b) daily irradiation, (c) daily module temperature and ambient temperature.
The Pearson correlation coefficient is provided for a pair of variables x with values xi and y with values yi by the equation:
(16)
Where co denotes covariance and σ denotes variance.
The coefficient ρcc can take on a value between 1 and +1. Strong positive correlation occurs for values close to +1, strong negative correlation occurs for values close to 1, and no association occurs for values close to 0 [42]. Since Pearson’s correlation establishes a straight line of dependence between two variables, linear analysis is assumed when comparing them.
The power_ generation dataset file provides the generated power, whereas the weather dataset file provides the independent attributes used in solar energy prediction. Here, the direction, shape, and magnitude of the dispersion of the data points between the two files’ characteristics are used to determine the existence of a relationship.
Four types of power-generated values can be predicted: DC_POWER, AC_POWER, DAILY_YIELD, and TOTAL_YIELD. The ρcc analysis was executed to identify the outputs that depended on the input data the most. Table 5 shows the results of the ρcc analysis; it demonstrates that there is a strong correlation between AC_POWER and DC_POWER; consequently, DC POWER is chosen as the predictability parameter for our work. Since solar panels produce DC power [43].
4. The proposed explainable solar power generation forecasting model
The proposed model has four main phases as illustrated in Fig 5 which are data preparation, hyper-parameter optimization, model evaluation, and model explanation. The detailed with each phase will be described in this section.
4.1 Data preparation phase
Data preparation involves cleaning and processing raw data for accurate ML predictions. Data preparation, the hardest part of ML, simplifies real-time initiatives. So, in this phase, the data will be aggregated from the two tables: data analysis and visualization and splitting the dataset.
Considering Plant ID, time, and SOURCE KEY; the power generation files; and the sensor reading files combined in a single file, this is done to make the data more conducive to the prediction model, as seen in Table 6.
Splitting a dataset into training and testing sets is a widely used method for model validation. It involves fitting and validating statistics and machine learning models on both the training and testing sets. By allocating a distinct validation dataset, it becomes possible to evaluate and compare the predictive efficacy of different models, mitigating the potential issue of overfitting on the training set. This approach helps to ensure that the model’s performance is not biased towards the training data and can generalize well to new data [44].
To effectively analyze a dataset, it is necessary to divide it into a training set comprising n rows and a testing set comprising m rows, where the total number of rows is denoted by Rs = n + m. The splitting ratio, represented by m/Rs, is denoted by the symbol γ. In datasets that contain predictor variables x and u, both of which are less than Rd, the training set is defined as DStrain = {(xi,yi)}, where I = 1,…, n, and the testing set is defined as DStest = {(ui,vi)}, where I = 1,…, m. If a predictor variable is categorical, it is presumed to be coded to a numerical variable.
The objective is to develop a model g(x; β) that approximates E(y |x), where β represents a collection of unknown parameters in the model. The training set is utilized to estimate the value of β, while the testing set is used to evaluate the approximation error. To determine the prediction error of the estimated model g(x; β) derived from the training set, a loss function L denoted by L(y, g(x; β)) is used [44].
4.2 LSTM hyper-parameter optimization based on EO
Like other neural network models, LSTM applications require manual setting of model parameters such as the number of cells, model optimization function, and training epochs. The network model’s performance depends on these factors. This research proposes using the EO method to optimize the internal parameters of LSTM and then using the optimized neural network to predict power generation.
In order to expedite the construction of LSTM models, it is essential to investigate hyper-parameter tuning and to generate recommendations that can serve as a solid starting point for the type of network being developed. To get the ideal values for the Hyper-parameters (the numbers of LSTM cells P1c, the numbers of epoch’s P2e, and optimization function ‘s type P3o), the interval values for each parameter are set as shown in Table 7.
The flow of the proposed model phases for power forecasting is depicted in Fig 6, and the subsequent sections will provide a description of each stage.
4.2.1 Initialization.
Particles are used in this stage of EO, with each particle standing in for the concentration vector that holds the optimal solution. In order to generate a random vector of initial concentrations within the search space, Eq 7 is employed. Our particles are denoted by the symbols P1c, P2e, and P3o; their values are denoted by the ranges [10 – 25], [25 – 100], and [0 – 1] (where 0 indicates to Adam optimizer while 1 indicates to SGD optimizer) respectively. Moreover, their initial value is determined by Eq (7).
Every meta-heuristic algorithm has a specific objective that aligns with its inherent characteristics. The objective of EO is to attain a state of equilibrium, and by achieving this state, EO can potentially solve the optimization problem with near-optimal results. However, during the optimization phase, the concentration levels required to reach the equilibrium state remain unknown to EO. Once the population reaches equilibrium, the top four particles are analyzed and chosen as potential candidates, thus being added to a candidate list. Additionally, another candidate is chosen based on the average of the top four particles. This list of five equilibrium candidates serves as a valuable tool for EO, allowing it to explore and exploit opportunities effectively. The first four candidates help EO enhance its ability to diversify and exploit on average, while all five candidates are stored in an equilibrium pool vector, as shown in Eq (8).
4.2.2 Constructing and training LSTM.
Training a model is the process of feeding artificial data to a parametrized machine learning algorithm in order to generate a model with optimal learned trainable parameters that minimize an objective function. At this phase, Data is supplied in batches to LSTM. In the previous phase, the parameters of the LSTM are defined; in the subsequent phase, the performance is evaluated. Fig 7 shows the LSTM Architecture.
4.2.3 Evaluating fitness function.
The fitness function is a crucial component of all meta-heuristic methods. For solving hyper-parameter tuning problems using a meta-heuristic, EO must be provided till it permits the search core of the optimization process to be identified. The fitness function utilized by EO to solve the hyper-parameter tuning problem is computed by minimizing the MSE of the training process.
is the output of training process of LSTM model. The LSTM model has a number of LSTM cells P1c and the optimization algorithm P3o.
4.2.4 Updating concentration.
The concentration is modified according to Eq 9 based on Eq (18). Where Eq (19) calculates the fitness value based on the newly obtained concentration and then updates the individual ideal concentration and the global optimal concentration of the particles. Fig 8 shows the updated values of the three parameters across the EO optimizer progress and its effects on RMSE values.
4.3 Model evaluation
Understanding the strengths and shortcomings of a machine learning model requires evaluating the model using a variety of evaluation measures. In addition to its role in model monitoring, model evaluation is crucial for determining a model’s efficacy at the early stages of research.
Numerous statistical metrics are used to evaluate the precision of models created to forecast the performance of solar energy generation. These criteria largely concentrate on quantifying the disparities between the projected values and the real measurements. The statistical methods used for this purpose include the coefficient of determination R2, RMSE, MAE, COV, and EC. The Eqs (20)–(24) mathematically establish the link between these statistical factors, as demonstrated in the reference [45].
R2: the R-square’s primary goal is to calculate the correlation between predicted and actual data. A dataset contains n values designated as y1, y2,…,n (also known as yi or as a vector y = [y1, y2,…,n]T), each of which is associated to a forecasted value f1,…, fn. It is changed between its lowest value (zero) and its maximum value (one). The better the correlation and the better the artificial model, the closer "R2" is to one.
MAE: the mean absolute error is determined as the average absolute error of datasets, and it is calculated using Eq 20:
(21)
RMSE: the root mean square error metric is commonly utilized to measure the error between the predicted and measured values, especially for assessing the difference between the predicted and target datasets. As the RMSE value becomes smaller, the accuracy of the model improves.
COV: the coefficient of variation is a statistical measure that indicates how far individual datasets in a set of data deviate from the average value. For the model to be more accurate, it needs to have a low coefficient of variation (COV).
EC: the efficiency coefficient value is a statistical indicator that determines the accuracy of the model. To ensure that the model is fitted correctly, it is necessary that the EC be equal to 1.
Where d, and y stand for the measured and predicted values, respectively, and N stands for the number of iterations. yi is the predicted datasets, and is the mean.
4.4 Model explanation based on LIME
The goal is to get a better understanding of how to apply XAI techniques to solar power generation forecasts and how to interpret "black box" machine learning models for usage in solar power station applications. In this paper, the Long-Short Memory (LSTM) is assumed to be the primary black-box model. The preceding section outlined the process of training the LSTM model. To achieve optimal performance, the LSTM model’s hyper-parameters are optimized with the EO algorithm. The LIME tool provides aid in determining an interpretable model as part of an interpretable representation at the level of the neighborhood.
5. Experiments, results and discussion
This section demonstrates the results of all the experiments carried out to evaluate the efficiency of the proposed model. TensorFlow and Keras, along with GPU processing from Google Collab, were used to analyze the results of all of the experiments.
The LSTM network receives data in batches, and the batch size, which determines the number of rows processed by the model before updating its weights, is determined by the user (in our case, it was set to 10). While processing a batch, LSTM retains its state, and between batches, the state can either be maintained or cleared. By default, the state is cleared. The experiment was conducted using the parameters listed in Table 8.
The experiment involved setting the number of LSTM cells to 10 and using "mean_squared_error" as the loss function for LSTM, along with ADAM as the optimization algorithm and a learning rate of 0.1. The training was conducted with a batch size of 1, and testing was conducted with a batch size of 1 as well. The number of epochs was kept constant at 25.
Table 9 illustrates the experiment test results for solar power generation forecasting using LSTM, where the R2, RMSE, COV, MAE and EC are 0.67, 2.2, 1.31, 1.15 and 0.71 respectively.
The LR, the DT and Gradient Boosting were executed to verify the efficiency of the LSTM; Table 9 shows the experiment results for solar power generation forecasting with different ML algorithms and LSTM.
In Table 9, the LR train results are 1, 1.16, 1.72, 0.00 and 0.92 for R2, RMSE, COV, MAE and EC respectively. While the LR tests results are 1, 0.96, 1.58, 0.18 and 0.86 for R2, RMSE, COV, MAE and EC respectively. The DT train results are 1, 0.99, 1.8, 0.00 and 1.1 for R2, RMSE, COV, MAE and EC respectively. While the DT tests results are 1, 2.12, 1.1, 1.01 and 0.71 for R2, RMSE, COV, MAE and EC respectively. The Gradient Boosting train results are 1, 0.99, 1.8, 0.00 and 1.1 for R2, RMSE, COV, MAE and EC respectively. While the Gradient Boosting tests results are 1, 2.07, 1.52, 0.86 and 0.64 for R2, RMSE, COV, MAE and EC respectively.
LR gets better results for R2, RMSE, and MAE measures, but it does not achieve good results for COV and EC; LSTM, on the other hand, does. Therefore, it is proposed that LSTM be used, and attempts are made to improve its efficiency through the use of an optimizer technique.
Table 10 displays the calculation of the training time for the three examined models throughout the training process which indicates the complexity of each one. The training running time for DT, LR, and LSTM are 4.84E-05 min, 1.47E-05 min, and 1.41E-01 min respectively. As demonstrated, LSTM has a longer processing time which means that it is a more complicated one, but it is worthwhile to utilize and strive to enhance its performance due to its superior efficacy.
Fig 9 depicts the actual value and the predicted value of the training model for the LSTM algorithm, as well as the actual value and the predicted value of the testing phase.
The experiment involving the application of the EO algorithm included several important hyper-parameters in the construction of LSTM, such as the training optimizer, training epoch, and number of LSTM cells. The primary training optimizers used were Gradient Descent (SGD) and Adam optimizers. Gradient Descent is an optimization algorithm that follows the negative gradient of an objective function to find its minimum. However, a fixed step size (learning rate) across all input variables is a limitation of Gradient Descent. On the other hand, Adam, short for "Adaptive Moment Estimation," is an extension of Gradient Descent that automatically adjusts the learning rate for each input variable based on the objective function. It further smooths the search process by utilizing an exponentially decreasing moving average of the gradient to update variables [46], the lower and upper order of each parameter is illustrated earlier in Table 7. For the P3o, “0” indicates ADAM optimizer while “1” indicates gradient optimizer.
The EO algorithm was assigned specific parameter values. The maximum number of iterations was set to 50, while a1 was assigned a value of 3 and a2 was assigned a value of 1, which are both recommended values. The values of these parameters are summarized in Table 11.
Fig 10 illustrates the EO optimizer execution that occurs while the optimizer is being performed. A sampling of the results of the EO can be shown in Fig 11, which covers a few different epochs.
After applying the EO algorithm, the best solution obtained includes 24 LSTM cells, "MSE" as the LSTM loss function, ADAM as the optimization algorithm with a learning rate of 0.1, a training batch size of 10, and a testing batch size of 10. The number of epochs used for training was fixed at 45.
Table 12 illustrates the Experiment of test results for LSTM based EO, where the R2, RMSE, COV, MAE and EC are 0.99, 0.46, 0.35, 0.229 and 0.95 respectively. Fig 12 depicts the actual value and the predicted value of the training model, as well as the actual value and the predicted value of the testing phase.
The PSO algorithm is utilized to optimize the hyper-parameters of LSTM. The primary objective is to identify the most significant parameters that influence the LSTM’s performance and therefore, the PSO algorithm has been implemented to discover the optimal parameters. The PSO is among the most popular metaheuristics. This method was inspired by natural swarm behavior, such as bird flocking and schooling. PSO has been widely used, and it has inspired a new study field known as swarm intelligence [47].
Table 13 illustrates the Experiment of solar power generation forecasting using LSTM based PSO test results, where the R2, RMSE, COV, MAE and EC are 0.9, 0.46, 0.35, 0.229 and 0.95 respectively. The training and testing model’s actual values and predicted values are illustrated in Fig 13.
After implementing the PSO optimization, it was discovered that the loss of the trained model is less when using the EO optimizer than when using the PSO optimization. In Fig 14, the red rectangle above the model’s performance covers the 20 to 40 epoch intervals after employing PSO, which is offset at the bottom by a second red rectangle covering the model’s performance at the same interval after employing EO. This indicates the efficacy of the model based on EO, as the loss is less. It also demonstrates that the performance of the EO-based model is stable and does not fluctuate, as is the case with the PSO-based model.
Fig 15 depicts the results of the executed experiments, where the best results are the blue-colored LSTM-based EO results; the accuracy presented by EC and R2 is the best, whereas the loss presented by MAE, RME, and COV is the least. While using LSTM gives the worst results (colored with red); the accuracy presented by EC and R2 is the worst, whereas the loss presented by MAE, RME, and COV is the highest. LSTM-based PSO results (colored with green) give moderated results.
5.1 Comparative analysis with other studies
To test and confirm the correctness of the proposed model, it was compared with the work of others who have employed LSTM. Table 14 demonstrates that the proposed model achieves superior outcomes to those achieved by other models that rely just on LSTM or on LSTM that has been optimized.
5.2 XAI for explain the forecasting model-based on LIME approach
LIME generates an explanation for a prediction based on the components of an interpretable model that resemble the black-box model near the point of interest.
Fig 16 illustrates the results of the LIME approach to explain specific predictions; it displays the solar DC power production predicting results with each attribute.
LIME provides explainability on a local scale in the form of an explanation for events that take place in close proximity to a prediction. For case “A” in Fig 16, in terms of their numerical contribution, the IRRADIATION is the important feature, whereas the AMBIENT TEMPERATURE and MODULE TEMPERATURE are the least important. It is possible to see the impact that each characteristic makes, whether it be good or negative, by looking at the explanations. For instance, AMBIENT TEMPERATURE and MODULE TEMPERATURE have a positive influence, whereas IRRADIATION has a negative effect on predictions.
For case “D” in Fig 16, in terms of their numerical contribution, the IRRADIATION and MODULE TEMPERATURE are the important feature, whereas the AMBIENT TEMPERATURE is the least important. It is possible to see the impact that each characteristic makes, whether it be good or negative, by looking at the explanations. For instance, AMBIENT TEMPERATURE and has a positive influence, whereas IRRADIATION ad MODULE TEMPERATURE has a negative effect on predictions and so on in all cases.
As shown before, the XAI permits an explanation of the factors influencing the predicted power output of the solar capacity across various environmental circumstances. However, some researchers employ Maximum Power Point Tracking (MPPT) techniques [52] and electricity distribution burdens [53] that may influence the amount of power generated.
6. Conclusion and future works
In this paper the LSTM model is proposed to forecast the power generated by the solar system under different environmental conditions. The performance of LSTM is evaluated in comparison to that of Decision DT and LR. It was demonstrated that the LSTM model performed more effectively than both the DT and LR models when the outcomes were in comparison. For enhancing the results he EO optimizer is proposed for tuning the LSTM hyper parameter.
The proposed model produced; R2, RMSE, COV, MAE, and EC values of 0.99, 0.46, 0.35, 0.229, and 0.95, respectively. These results improve the performance before turning hyper-parameter using the EO optimizer, which had R2, RMSE, COV, MAE, and EC values of 0.67, 2.21, 1.31, 1.15, and 0.71 respectively.
Additionally, the XAI-based LIME algorithm was used to explain the results, which helps improve the transparency and interpretability of the model’s predictions. This algorithm was successful in identifying the most important features that affected solar power generation, including weather conditions, time of day, and solar panel tilt angle.
In conclusion, the proposed X-LSTM-EO model, along with the use of the XAI-based LIME algorithm, offers a more accurate and transparent method for predicting solar power generation in solar plant systems. These findings have important implications for developing and deploying renewable energy sources, such as solar power.
The proposed model in this paper exhibits high prediction accuracy under various environmental conditions, demonstrating its universality and accuracy. It reduces uncertainty in PV power generation and safely integrates large-scale PV power generation into micro grids, reducing operating costs and improving efficiency and safety. It also introduces a new PV power generation forecast research direction: clustering data and noise reduction can reduce uncertainty. The research in this paper does not account for harsh weather conditions such as thunderstorms, sand, and dust.
Our future work will incorporate an efficient maximum power point tracking (MPPT) method into our proposed model. This method is crucial for enhancing the efficiency of PV power generation systems. In addition, we will consider the distribution of electrical loads in the proposed model. Finally, we will treat the limitations of our work by considering the harsh weather conditions.
References
- 1. Antonanzas J., Osorio N., Escobar R., Urraca R., Martinez-de-Pison F., & Antonanzas-Torres F. (2016, October). Review of photovoltaic power forecasting. Solar Energy, 136, 78–111.
- 2. Khan Z. A., Hussain T., & Baik S. W. (2023, May). Dual stream network with attention mechanism for photovoltaic power forecasting. Applied Energy, 338, 120916.
- 3. Almasad A., Pavlak G., Alquthami T., & Kumara S. (2023). Site suitability analysis for implementing solar PV power plants using GIS and fuzzy MCDM based approach. Solar Energy, 249, 642–650.
- 4. Qing X.; Niu Y. Hourly Day-Ahead Solar Irradiance Prediction Using Weather Forecasts by LSTM. Energy 2018, 148, 461–468.
- 5. Sobri S.; Koohi-Kamali S.; Rahim N.A. Solar Photovoltaic Generation Forecasting Methods: A review. Energy Convers. Manag. 2018, 156, 459–497.
- 6. Alshammari A. (2023). Generation forecasting employing Deep Recurrent Neural Network with metaheruistic feature selection methodology for Renewable energy power plants. Sustainable Energy Technologies and Assessments, 55, 102968.
- 7. Yin R., & He J. (2023). Design of a photovoltaic electric bike battery-sharing system in public transit stations. Applied Energy, 332, 120505.
- 8. Zulfiqar M., Kamran M., Rasheed M. B., Alquthami T., & Milyani A. H. (2023). A hybrid framework for short term load forecasting with a navel feature engineering and adaptive grasshopper optimization in smart grid. Applied Energy, 338, 120829.
- 9.
Osali, N. (2023, February). Optimal Scheduling of Active Distribution Networks Considering Dynamic Transformer Rating Under High Penetration of Renewable Energies. In 2023 8th International Conference on Technology and Energy Management (ICTEM) (pp. 1–7). IEEE.
- 10. Ahmed R., Sreeram V., Mishra Y., and Arif M. D., “A review and evaluation of the state-of-the-art in PV solar power forecasting: Techniques and optimization,” Renew. Sustain. Energy Rev., vol. 124, 2020, Art. no. 109792.
- 11. Khan M. H. R., & Righetti R. (2022). Ultrasound estimation of strain time constant and vascular permeability in tumors using a CEEMDAN and linear regression-based method. Computers in Biology and Medicine, 148, 105707. pmid:35725503
- 12. Kim E., Akhtar M. S., & Yang O. B. (2023). Designing solar power generation output forecasting methods using time series algorithms. Electric Power Systems Research, 216, 109073.
- 13. Cervantes J., Garcia-Lamont F., Rodríguez-Mazahua L., & Lopez A. (2020). A comprehensive survey on support vector machine classification: Applications, challenges and trends. Neurocomputing, 408, 189–215.
- 14. Ali I. M. S., & Hariprasad D. (2023). Hyper-heuristic salp swarm optimization of multi-kernel support vector machines for big data classification. International Journal of Information Technology, 15(2), 651–663.
- 15. Zhu X., Li M., Liu X., & Zhang Y. (2024). A backpropagation neural network-based hybrid energy recognition and management system. Energy, 131264.
- 16. Song J., Krishnamurthy V., Kwasinski A., & Sharma R. (2013, April). Development of a Markov-Chain-Based Energy Storage Model for Power Supply Availability Assessment of Photovoltaic Generation Plants. IEEE Transactions on Sustainable Energy, 4(2), 491–500. https://doi.org/10.1109/tste.2012.2207135
- 17. Long H., Zhang Z., & Su Y. (2014, August). Analysis of daily solar power prediction with data-driven approaches. Applied Energy, 126, 29–37. https://doi.org/10.1016/j.apenergy.2014.03.084
- 18. Phan Q. T., Wu Y. K., Phan Q. D., & Lo H. Y. (2022, May 2). A Novel Forecasting Model for Solar Power Generation by a Deep Learning Framework with Data Preprocessing and Postprocessing. 2022 IEEE/IAS 58th Industrial and Commercial Power Systems Technical Conference (I&CPS). https://doi.org/10.1109/icps54075.2022.9773862
- 19. Mat Daut M. A., Hassan M. Y., Abdullah H., Rahman H. A., Abdullah M. P., & Hussin F. (2017, April). Building electrical energy consumption forecasting analysis using conventional and artificial intelligence methods: A review. Renewable and Sustainable Energy Reviews, 70, 1108–1118. https://doi.org/10.1016/j.rser.2016.12.015
- 20. Ugurlu U., Oksuz I., & Tas O. (2018, May 14). Electricity Price Forecasting Using Recurrent Neural Networks. Energies, 11(5), 1255. https://doi.org/10.3390/en11051255
- 21. Qing X., & Niu Y. (2018, April). Hourly day-ahead solar irradiance prediction using weather forecasts by LSTM. Energy, 148, 461–468. https://doi.org/10.1016/j.energy.2018.01.177.
- 22. Abdel-Nasser M., & Mahmoud K. (2017, October 14). Accurate photovoltaic power forecasting models using deep LSTM-RNN. Neural Computing and Applications, 31(7), 2727–2740. https://doi.org/10.1007/s00521-017-3225-z
- 23. He Y., Gao Q., Jin Y., & Liu F. (2022). Short-term photovoltaic power forecasting method based on convolutional neural network. Energy Reports, 8, 54–62.
- 24. Singla P., Duhan M., & Saroha S. (2022). A hybrid solar irradiance forecasting using full wavelet packet decomposition and bi-directional long short-term memory (BiLSTM). Arabian Journal for Science and Engineering, 47(11), 14185–14211.
- 25. Singla P., Duhan M., & Saroha S. (2023). A point and interval forecasting of solar irradiance using different decomposition based hybrid models. Earth Science Informatics, 16(3), 2223–2240.
- 26. Singla P., Duhan M., & Saroha S. (2023). An integrated framework of robust local mean decomposition and bidirectional long short-term memory to forecast solar irradiance. International Journal of Green Energy, 20(10), 1073–1085.
- 27. Phan Q. T., Wu Y. K., Phan Q. D., & Lo H. Y. (2022). A novel forecasting model for solar power generation by a deep learning framework with data preprocessing and postprocessing. IEEE Transactions on Industry Applications, 59(1), 220–231.
- 28. Sabir D., Hafeez K., Batool S., Akbar G., Khan L., Hafeez G., et al. (2024). Prediction of Solar PV power using Deep Learning with Correlation-based Signal Synthesis. IEEE Access.
- 29. Wang Z., Zhang Y., Li G., Zhang J., Zhou H., & Wu J. (2024). A novel solar irradiance forecasting method based on multi-physical process of atmosphere optics and LSTM-BP model. Renewable Energy, 120367.
- 30. Ibrahim M. S., Gharghory S. M., & Kamal H. A. (2024). A hybrid model of CNN and LSTM autoencoder-based short-term PV power generation forecasting. Electrical Engineering, 1–17.
- 31. Ehteram M., Nia M. A., Panahi F., & Farrokhi A. (2024). Read-First LSTM model: A new variant of long short term memory neural network for predicting solar radiation data. Energy Conversion and Management, 305, 118267.
- 32. Hong Y. Y., & Martinez J. J. F. (2024). Forecasting solar irradiation using convolutional long short-term memory and feature selection of data from neighboring locations. Sustainable Energy, Grids and Networks, 38, 101271.
- 33. Liu W., & Mao Z. (2024). Short-term photovoltaic power forecasting with feature extraction and attention mechanisms. Renewable Energy, 120437.
- 34. Bukhari SM, Moosavi SK, Zafar MH, Mansoor M, Mohyuddin H, Ullah SS, et al. Federated transfer learning with orchard-optimized Conv-SGRU: A novel approach to secure and accurate photovoltaic power forecasting. Renewable Energy Focus. 2024 Mar 1;48:100520.
- 35. Abou Houran M, Bukhari SM, Zafar MH, Mansoor M, Chen W. COA-CNN-LSTM: Coati optimization algorithm-based hybrid deep learning model for PV/wind power forecasting in smart grid applications. Applied Energy. 2023 Nov 1;349:121638.
- 36. Khan UA, Khan NM, Zafar MH. Resource efficient PV power forecasting: Transductive transfer learning based hybrid deep learning model for smart grid in Industry 5.0. Energy Conversion and Management: X. 2023 Oct 1;20:100486.
- 37. Faramarzi A., Heidarinejad M., Stephens B., &Mirjalili S. (2020). Equilibrium optimizer: A novel optimization algorithm. Knowledge-Based Systems, 191, 105190.
- 38. Kawakura S., Hirafuji M., Ninomiya S., & Shibasaki R. (2022). Analyses of Diverse Agricultural Worker Data with Explainable Artificial Intelligence: XAI based on SHAP, LIME, and LightGBM. European Journal of Agriculture and Food Sciences, 4(6), 11–19. https://doi.org/10.24018/ejfood.2022.4.6.348.
- 39.
https://www.kaggle.com/datasets/anikannal/solar-power-generation-data?select=Plant_1_Generation_Data.csv.
- 40. Urbanowicz R. J., Meeker M., La Cava W., Olson R. S., & Moore J. H. (2018, September). Relief-based feature selection: Introduction and review. Journal of Biomedical Informatics, 85, 189–203. pmid:30031057
- 41. Ziegel E., Press W., Flannery B., Teukolsky S., & Vetterling W. (1987, November). Numerical Recipes: The Art of Scientific Computing. Technometrics, 29(4), 501. https://doi.org/10.2307/1269484.
- 42. Kumar S., & Chong I. (2018, December 19). Correlation Analysis to Identify the Effective Data in Machine Learning: Prediction of Depressive Disorder and Emotion States. International Journal of Environmental Research and Public Health, 15(12), 2907. pmid:30572595
- 43.
https://center4ee.org/how-solar-energy-works/
- 44. Joseph V. R. (2022, April 4). Optimal ratio for data splitting. Statistical Analysis and Data Mining: The ASA Data Science Journal, 15(4), 531–538. https://doi.org/10.1002/sam.11583
- 45. Zayed M. E., Zhao J., Li W., Elsheikh A. H., & Elaziz M. A. (2021, November). A hybrid adaptive neuro-fuzzy inference system integrated with equilibrium optimizer algorithm for predicting the energetic performance of solar dish collector. Energy, 235, 121289. https://doi.org/10.1016/j.energy.2021.121289
- 46.
Code Adam Optimization Algorithm From Scratch by Jason Brownlee on January 13, 2021 in Optimization. https://machinelearningmastery.com/adam-optimization-from-scratch/
- 47. Freitas D., Lopes L. G., & Morgado-Dias F. (2020, March 21). Particle Swarm Optimisation: A Historical Review Up to the Current Developments. Entropy, 22(3), 362. pmid:33286136
- 48. Zheng J., Zhang H., Dai Y., Wang B., Zheng T., Liao Q., et al. (2020, January). Time series prediction for output of multi-region solar power plants. Applied Energy, 257, 114001. https://doi.org/10.1016/j.apenergy.2019.114001
- 49. Wang K., Qi X., & Liu H. (2019, December). Photovoltaic power forecasting based LSTM-Convolutional Network. Energy, 189, 116225. https://doi.org/10.1016/j.energy.2019.116225
- 50. Wang K., Qi X., & Liu H. (2019, October). A comparison of day-ahead photovoltaic power forecasting models based on deep learning neural network. Applied Energy, 251, 113315. https://doi.org/10.1016/j.apenergy.2019.113315
- 51. Saxena N., Kumar R., Rao Y. K. S. S., Mondloe D. S., Dhapekar N. K., Sharma A., et al. (2024, January). Hybrid KNN-SVM machine learning approach for solar power forecasting. Environmental Challenges, 14, 100838. https://doi.org/10.1016/j.envc.2024.100838.
- 52. Salau A. O., & Alitasb G. K. (2024, March). MPPT efficiency enhancement of a grid connected solar PV system using Finite Control set model predictive controller. Heliyon, 10(6), e27663. pmid:38500997
- 53. Krishna V. M., Duvvuri S. S., Sobhan P. V., Yadlapati K., Sandeep V., & Narendra B. (2024, March). Experimental study on excitation phenomena of renewable energy source driven induction generator for isolated rural community loads. Results in Engineering, 21, 101761. https://doi.org/10.1016/j.rineng.2024.101761