Xiang Wang ,Liangsa Wang ,Han Li and Yibin Guo
1School of Civil Engineering and Architecture,Zhengzhou University of Aeronautics,Zhengzhou,450015,China
2School of Civil and Architectural Engineering,Jiangsu University of Science and Technology,Zhenjiang,212003,China
ABSTRACT The original whale optimization algorithm (WOA) has a low initial population quality and tends to converge to local optimal solutions.To address these challenges,this paper introduces an improved whale optimization algorithm called OLCHWOA,incorporating a chaos mechanism and an opposition-based learning strategy.This algorithm introduces chaotic initialization and opposition-based initialization operators during the population initialization phase,thereby enhancing the quality of the initial whale population.Additionally,including an elite opposition-based learning operator significantly improves the algorithm’s global search capabilities during iterations.The work and contributions of this paper are primarily reflected in two aspects.Firstly,an improved whale algorithm with enhanced development capabilities and a wide range of application scenarios is proposed.Secondly,the proposed OLCHWOA is used to optimize the hyperparameters of the Long Short-Term Memory(LSTM) networks.Subsequently,a prediction model for Realized Volatility (RV) based on OLCHWOA-LSTM is proposed to optimize hyperparameters automatically.To evaluate the performance of OLCHWOA,a series of comparative experiments were conducted using a variety of advanced algorithms.These experiments included 38 standard test functions from CEC2013 and CEC2019 and three constrained engineering design problems.The experimental results show that OLCHWOA ranks first in accuracy and stability under the same maximum fitness function calls budget.Additionally,the China Securities Index 300(CSI 300)dataset is used to evaluate the effectiveness of the proposed OLCHWOA-LSTM model in predicting RV.The comparison results with the other eight models show that the proposed model has the highest accuracy and goodness of fit in predicting RV.This further confirms that OLCHWOA effectively addresses real-world optimization problems.
KEYWORDS Whale optimization algorithm;chaos mechanism;opposition-based learning;long short-term memory;realized volatility
The volatility of financial markets refers to the degree of volatility of the prices of financial assets,which serves as a crucial risk indicator.Realized Volatility(RV)has become one of the most commonly used volatility measures,which is widely employed to assess the Value at Risk (VaR) of investment portfolios and to determine the pricing of derivatives based on options.Consequently,the prediction of RV in financial markets has become a topic of great interest to both theoretical and practical communities.
Since RV has the characteristics of long memory and aggregation [1],the AutoRegressive Fractionally Integrated Moving Average (ARFIMA) [2] and the Heterogeneous Autoregressive Model(HAR)[3]that can capture these two characteristics have become the most widely used models for RV prediction.Autoregressive has become the most widely used model for RV prediction.The research of scholars using the ARFIMA model to capture the long memory of economic data started in the 1990s.Subsequently,scholars have conducted numerous improved studies on the limitations of ARFIMA in predicting RV.Andersen et al.[4],Giot et al.[5],and Degiannakis [6] all proposed improved models for the defects of the ARFIMA model.Furthermore,Zhou et al.[7] and Izzeldin et al.[8]both employed the ARFIMA model to study RV and demonstrated its excellent performance in RV prediction.Muller et al.proposed the HAR model in 1997.Later,Barndorff-Nielsen et al.[9]proposed improved versions of the HAR model.Meanwhile,numerous empirical studies have demonstrated the excellent performance of the HAR model in RV prediction[10,11].However,the HAR and ARFIMA are both econometric models,and they both have the disadvantages of being limited to linear time series patterns and not providing accurate forecasts.
In recent years,artificial intelligence technology has led to the application of Long Short-Term Memory (LSTM) [12],which can capture nonlinear time series features,to the RV prediction problem.Maknickiene et al.[13] used LSTM to predict the rate of return of the exchange rate in 2012,and the results demonstrated that the prediction accuracy was improved compared to BP neural networks.Chen et al.[14] used the LSTM model to predict the returns of the Chinese stock markets in 2015.According to the results,the LSTM model’s prediction accuracy was better than the Stochastic Volatility(SV)model.In 2018,Kim et al.[15]used a mixture of Generalized AutoRegressive Conditional Heteroskedasticity(GARCH)and LSTM for predicting Korean stock market volatility and the results were impressive.Recently,scholars have also attempted to solve the RV prediction problem using LSTM.Hu et al.[16] proposed a combined model that integrates GARCH,LSTM,and Artificial Neural Network(ANN)in 2020 for predicting copper price volatility,which was more accurate than the other five models.The Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (CEEMDAN) and LSTM model was applied by Lin et al.[17] in 2022 to predict the RV of the CSI300,S&P500,and STOCXX50 with good results,solving the problem of noise associated with RV time series.Chen et al.[18] used the Principal Component Analysis (PCA) and LSTM model to predict the volatility of Chinese and American stock index futures within the same year,and the experimental results were more accurate than the other seven models.The results suggest that PCA can improve the LSTM prediction performance by reducing attributes.While LSTM-based volatility prediction is highly accurate,the determination of the hyperparameters has always presented a challenge.In this paper,a meta-heuristic algorithm will be introduced to automate the process of hyperparameter optimization for LSTM.
The Genetic Algorithm (GA) [19] was first proposed and achieved great success in solving optimization problems in the 1960s,inspired by biological research.Metaheuristic algorithms have made significant progress since the 1980s.Kirkpatrick et al.[20] proposed the Simulated Annealing(SA) algorithm in 1982.Glover [21] proposed the Tabu Search Algorithm (TSA) in 1986.The Ant Colony Optimization(ACO)was proposed in 1992 by Dorigo[22].In 1995,Venter et al.[23]proposed the concept of Particle Swarm Optimization (PSO).Storn et al.[24] introduced the Differential Evolution Algorithm(DE)in 1997.Many new meta-heuristic algorithms have been proposed in recent years,such as Monarch Butterfly Optimization (MBO) [25],Naked Mole-Rat Algorithm (NMR)[26],Moth Swarm Algorithm (MSA) [27],Harris Hawks Optimization (HHO) [28],Slime Mould Algorithm (SMA) [29],African Vultures Optimization Algorithm (AVOA) [30],Carnivorous Plant Algorithm (CPA) [31],and the Hunger Games Search (HGS) [32].These algorithms have emerged as prominent representatives and have achieved significant success in solving optimization problems across various domains[33–39].
Mirjaliliab et al.[40]developed the Whale Optimization Algorithm(WOA)by mimicking humpback whales’natural search,surround,and attack behaviors in 2016.The WOA algorithm has demonstrated its effectiveness in optimizing diverse problem domains,such as reservoir scheduling [41],workshop scheduling [42],rolling bearing fault diagnosis [43],turbine heat consumption prediction[44],and residue hydrogenation reaction kinetic model parameter optimization[45].While the basic WOA algorithm is characterized by its simplicity and a limited number of parameters,it also has some shortcomings.First of all,the accuracy and speed of heuristic algorithms are directly affected by the quality of the initial population[46].In general,the more diverse the initial population,the stronger the algorithm’s ability to perform global searches[47].Random initialization of the original WOA does not produce an even distribution of the initial population throughout the whole search space,which affects the algorithm’s efficiency in solving problems.Additionally,the algorithm has a tendency to converge towards local extrema during the late stage of population evolution[48,49],which results in poor convergence accuracy.
There are two primary motivations for this study.First,this paper introduces a chaos mechanism and an opposition-based learning strategy to compensate for the limitations of the WOA.As of now,there are several works [50–55] that use chaos and opposition-based learning strategies to improve the WOA,among which literature[53]is the closest to the idea presented in this paper.However,the algorithm employed differs in two ways.In the first instance,the opposition-based learning operator is different.This paper adopts the classical opposition-based learning operator,similar to that described in the literature[52],while the partial-opposition-based learning operator,as discussed in the literature[53].In the second instance,this paper introduces the Jumping rate (Jr) and discusses its parameter sensitivity,while in literature[53],Jris fixed at 0.5.Second,to improving the accuracy of RV prediction,this paper proposes a prediction model based on the improved WOA-LSTM model.As of now,some applied research has been conducted on the improved LSTM model based on WOA [56–60],but this research takes a step further.In the first instance,the improved WOA algorithm has stronger optimization capabilities compared to the original WOA algorithm.In the second instance,the WOALSTM model is used to predict RV,thereby enhancing its usefulness.
In summary,this paper proposes an improved whale optimization algorithm called OLCHWOA,which utilizes chaos mechanisms and opposition-based learning techniques.In addition,an RV prediction model based on OLCHWOA-LSTM is developed.Through simulation experiments,this study has drawn several conclusions: (1) The experimental results of 38 test functions and three engineering design problems with constraints establish the superior performance of OLCHWOA with statistical significance when compared to the other five algorithms;(2)The RV prediction for the CSI 300 index in mainland China indicates that OLCHWOA-LSTM Models have a competitive advantage.As a result of our research,this paper has made two important contributions.First,the utilization of the chaos mechanism and opposition-based learning strategy improves the global search ability of the WOA algorithm and further advances related research.Additionally,this paper has developed an RV prediction model based on OLCHWOA-LSTM,which enhances the RV prediction method by introducing a novel approach for neural network-based RV prediction models.
In the remainder of the paper,the following sections will be discussed: Section 2 introduces the concept of OLCHWOA.The simulation experiment is presented in Section 3.Section 4 introduces the OLCHWOA-LSTM model and discusses its prediction performance on the RV of the CSI 300 dataset.Our conclusions are presented in Section 5.
WOA is a novel heuristic optimization algorithm based on the mathematical modeling of the unique hunting method of humpback whales.In nature,humpback whales swim upward in a spiral position from the depths of the ocean and expel numerous bubbles of different sizes.Prey will be forced toward the center of the bubble net by the surrounding bubble net,at which point the whale opens its mouth and swallows the prey.
Since the initialized population of the WOA lacks a priori experience,it is assumed that the prey position is set as the global optimization,and other individual whales converge towards the prey position as a means of updating their own positions.The WOA algorithm for locating the optimal position relies on three mechanisms,which are described as follows.
2.1.1 Encirclement Hunting
By swimming towards the nearest whale in the group,the whales narrow the circle around their prey.The formula for updating whale positions is used in this process.
whereis the next position of the whale,is the current optimal whale,Dis the encircling step.Cis the effect of the distance betweenand individualWhenC>1,the degree of influence is enhanced,and vice versa.The convergence factor,denoted asa,linearly decreases from 2 to 0 as the number of iterations increases.r1andr2represent random numbers within[0,1].
2.1.2 Spiral Hunting
The whales hunt by spiraling towards the prey’s position,and this can be expressed by the following formula:
whereD′is the distance betweenandand the constantbdetermines the shape of the spiral.lis a random number within[-1,1],wherelis-1 when whales are closest to the prey position,andlis 1 when whales are farthest from the prey position.
In addition,the whales swim around their prey while simultaneously conducting spiral hunting simultaneously.The probabilityPrfor an individual whale choosing encircling or spiral hunting is 50%.The behavior is as follows:
2.1.3 Random Hunting
Additionally,whales can go on random hunts,which increases the search area for the whales.Following is the formula for this process:
When the parameter |A|<1,the individual whale moves away from the random individualto surround the prey,and then tends to the optimal individualWhen the parameter |A|≥1,the individual whale deviates from the prey position to search for a better prey.Therefore,the value of parameterAdetermines whether the individual whale adopts Eq.(1)or Eq.(9)to update its position.In light of the fact that parameterAis largely affected by the convergence factora,ais an important consideration for the exploration and development of balanced algorithms.
The quality of the initial population has a significant impact on the optimization efficiency of metaheuristic algorithms [61].Currently,the majority of metaheuristic algorithms employ random initialization to generate the initial population,which results in uneven distribution of the population across the solution space,reducing diversity and making the algorithm prone to premature convergence[62].Chaos is a common nonlinear phenomenon distinguished by its non-repetition,ergodicity,and dynamism [63].Using chaos mechanisms to generate the initial whale population,rather than random generation,can enhance the algorithm’s search efficiency during the search process [64],facilitating a faster exploration of the search space.Chaos mapping has been employed to address various optimization problems[65].The logistic map is a simple and effective chaotic system whose expression is as follows:
whereμis the control parameter,the value range is 0<μ≤4.0.Whenμ=4,the logistic map is completely chaotic.Krepresents the number of iterations.ch1is a random number within the interval[0,1],whilechi0.25,0.5 and 0.75.
Assuming that the initial whale population is composed ofNindividuals,the specific process of using the Logistic chaotic initialization operator to generate the initial whale population is as follows.The first individual is randomly generated within the upper and lower bounds.Then,Eq.(11)is used to iterate the first individual in order to obtain the remainingN-1 individuals.Finally,Eq.(12)is used to map variables to individual whales.
whereXi,min_j,Xi,max_jdefine the upper and lower bounds of thej-thdimension of thei-thindividual,respectively,Xijis the mapped whale individual.
Fig.1 illustrates the distribution of a random population and a logistic chaotic population in a two-dimensional plane.Under different population sizes,it is found that the population generated by the logistic chaotic initialization operator exhibits better uniformity compared to the population generated by random initialization.
Figure 1:Comparison of two initialization methods in two-dimensional plane(a)Population size=20,(b)Population size=200
Opposition-based learning(OBL)is an enhancement strategy introduced by Tizhoosh within the domain of swarm intelligence in 2005[66].At its core,OBL involves the simultaneous consideration of feasible solutions and their corresponding opposite solutions,selecting the best solutions for advancement to the next generation of the population.A significant limitation associated with random initialization is that when the current solution is considerably distant from the optimal solution,it may lead to prolonged search times or the algorithm getting stuck in local optima[67].Interestingly,according to probability theory,solutions that are opposite to each other have a 50% probability of being closer to the optimal solution[68].The integration of opposite solutions can result in significant enhancements in search efficiency while simultaneously reducing computational expenses[69].
The Opposition-based Learning strategy of this paper introduces two types of operators called the Opposition-based Initialization operator(OI)and the Elite Opposition-based Learning operator(EOL)to enhance WOA’s global search ability.Their detailed descriptions are as follows.
2.3.1 Opposition-Based Initialization Operator(OI)
As shown in Section 2.2,this paper introduces the chaotic initialization operator to generate the initial population,which ensures an even distribution of the initial population.To increase the search efficiency,this paper attempts to introduce an Opposition-based Initialization operator(OI)to further optimize the initial population.
The feasible solution is given byXi=(xi1,xi2,...,xij),wherexij∈[lij,uij],i=1,2,...,N,j=1,2,...,D.Each dimension’s maximum and minimum values are represented bylijanduij.The opposite solution can be defined asXi=whichis obtained by Eq.(13)as follows:
After obtaining the opposing population,the fitness values are compared between the original population and the opposing population.The individuals with the highest fitness values are then selected to be included in the initial population.Let the fitness function bef(.).The selection operator can be expressed as follows: The original population is compared with the opposite population in order to select the most suitable individuals for entry into the population.
2.3.2 Elite Opposition-Based Learning Operator(EOL)
Based on the ideas of Wang et al.[70]and Zhou et al.[71],this paper proposes an Elite Oppositionbased Learning operator(EOL).By introducing the Jumping rate(Jr),whales have a greater chance of jumping out of the local solutions,thereby improving the algorithm’s global search capability and thus improving the algorithm’s ability to search globally.According to this principle,if rand(0,1) whereηis the generalized coefficient,η∈(0,1),ajandbjare the upper and lower bounds. Additionally,if the generated elite opposition-based solution crosses the boundary [aj,bj],it is reset according to Eq.(18). Step 2:Compared with the current solution and the elite opposition-based solution,evaluate each solution’s fitness value,and the best individual will be selected to stay in the population according to Eq.(14).With the EOL operator,the population can be updated with the information contained in the current population,which will increase the convergence speed and the capabilities of WOA global exploration. In the same way that other meta-heuristic algorithms encounter issues,WOA may also converge to a local optimum prematurely.An improved Whale Optimization Algorithm based on chaos mechanism and opposition-based learning(OLCHWOA)is designed to improve WOA’s global search capability and prevent a decrease in population diversity during later iterations. During the initialization stage,the logistic chaotic initialization operator is used first to create a diverse population of high quality.Afterward,the opposite population is constructed using the OI operator.Then,sort the chaotic initial population and its opposite population and select the topNindividuals with higher fitness values to enter the initial whales.This will allow the algorithm to fully extract useful information from the solution space,as well as improve the search efficiency of the algorithm while ensuring the quality of the initial solution.During the loop iteration stage,the EOL operator is applied to the current optimal group,taking into account the three hunting behaviors of whales.By utilizing this operator,the algorithm can be enhanced in terms of its global optimization capabilities while maintaining the diversity of the population.The flowchart of OLCHWOA is depicted in Fig.2. Time complexity stands as a fundamental metric for assessing algorithmic efficiency.Assuming a population size ofP,dimensionality ofD,and the number of iterations denoted asT,the time complexity analysis for both the WOA and OLCHWOA algorithms is as follows. The standard WOA algorithm consists of two phases: random population initialization phase and subsequent whale position update.In the initialization phase,the time complexity of WOA can be expressed asT1=O(P?D).In the position update phase,whales employ encircling hunting,spiral hunting,or random hunting mechanisms.For each iteration,the computational complexity isO(P?D),and acrossTiterations,it accumulates toT2=O(T?P?D).Consequently,the aggregate time complexity of WOA can be represented asTWOA=T1+T2,resulting in a time complexity ofO(P). The proposed OLCHWOA algorithm consists of three stages: chaotic and opposition-based learning population initialization,whale position updates,and an opposition-based search phase.In the chaotic and opposition-based learning initialization stage,the time complexity for OLCHWOA’s initialization is denoted asand is expressed as=O(P?D?2).The whale position update stage closely parallels that of the WOA and maintains a time complexity equivalent toT2,consistent with the WOA.In the opposition-based search stage,which incorporates a probability parameterJrgoverning the execution of the opposition-based search strategy,the time complexity for this stage is represented as=O(P?D?Jr).Consequently,the overall time complexity of the OLCHWOA algorithm can be summarized asTOLCHWOA=+T2+leading to a final time complexity ofO(P).It is worth noting that OLCHWOA shares an equivalent time complexity with WOA,signifying that the proposed enhancement strategy in this study does not compromise the algorithm’s solving efficiency. Generally,standard test functions and engineering design problems are used as optimization objects to assess algorithm performance.This section presents the test functions used,the results of the experiment,and the analysis.All experiments are implemented in the software of Python version 3.7.The computer system is 11th Gen Intel(R)Core(TM)i7-11700 processor and 32 G RAM. A total of 28 standard test functions were selected from the CEC 2013 benchmark[72],while 10 standard test functions were chosen from the CEC 2019[73]to evaluate the algorithm’s performance.As most functions have multiple local optima,it is difficult to determine their global optimum accurately,which allows one to fully examine the algorithm’s optimization ability. There are three categories of standard tests from CEC 2013:f1–f5,f6–f20,andf21–f28,which represent unimodal,basic multimodal,and composite test functions,respectively.As the unimodal function has only one global optimum,it is used to determine convergence speed and accuracy.Multimodal functions are suitable for evaluating the global search capability because it has multiple local optimal solutions.Composite test functions are created by combining,shifting,rotating,and biasing other test functions.They have a variety of shapes,and they have several local optimization points,which are used to evaluate whether the algorithm can achieve a balance between local search and global exploration.The test functions in CEC 2019 are known for their high complexity and challenging nature.Tables 1 and 2 provide detailed information on the test functions used,whereDis the dimension,Rangeis the search range of the function,andfminindicates the theoretical optimal value. As a precaution and to ensure fairness,all experiments were run separately 30 times.Then,the mean value(Mean)and standard deviation(Std)for each test were calculated.Generally,theMeanreflects the average precision that an algorithm can achieve after a certain number of evaluations.TheStdreflects the algorithm’s stability.All algorithms used the maximum fitness function call(Max_Fitness)as the termination condition,with specific settings provided in each section. Compared to WOA,OLCHWOA and OLWOA increase the parameterJr.The parameterJrrefers to the calling probability of the EOL operator,which is used to balance exploration and exploitation.WhenJr=0,the probability of the EOL operator is 0.Therefore,the OLWOA algorithm only calls the OI operator and not the EOL operator.IfJr=1,the probability of the EOL operator is 1.This means that the OLWOA algorithm will call the EOL operator every time an iteration occurs. Furthermore,the termination criterion for the experiments in this section is to reach the maximum number of fitness function calls.The population sizeP,can affect the actual number of iterations the algorithm undergoes.To investigate the influence of parametersJrandPon the performance of OLCHWOA and determine their optimal values,four representative functions (unimodal function F2,multi-modal function F6,F13,and composite function F25)from the CEC2013 were selected for testing.These experiments were designed with five different levels ofJr={0,0.2,0.5,0.8,1.0}and four distinctP={10,20,30,50}.A maximum fitness function calls of 2000 was set,and each experiment was rigorously conducted 30 times to ensure statistical robustness.Table 3 presents the experimental results for different combinations ofJrandP. In the table,Meansignifies the average fitness values obtained from 30 independent runs of the algorithm,whileStddenotes the standard deviation.Notably,the experiment data highlighted in bold font corresponds to the optimal results achieved.Additionally,theRankcolumn assigns rankings to the seven comparative algorithms.These rankings are determined by sorting the algorithm’s performance across various test functions.The numbers in parentheses indicate how many times the algorithm achieved the best result on such test functions.Based on these counts,the numbers outside the parentheses determine the final rankings of the seven algorithms.Ranking first also implies that the algorithm achieved the best results on more functions.From Table 3,it can be observed that different settings ofJraffect the algorithm’s performance on the test functions,leading to the following conclusions. Conclusion 1: With a fixed population sizeP,asJrincreases from 0.0 to 1.0,the algorithm’s performance exhibits an initial improvement followed by a decline.This is evident whenP=10,Jr=0.8 yields the best results for F2,whileJr=0.5 is optimal for F6,F13,and F25.ForP=20/30/50,Jr=0.5 ranks first in both the mean and standard deviation of final convergence accuracy across all four functions.This highlights the effectiveness of augmenting the probability of using the elite opposition-based search operator to improve the optimization accuracy of WOA.However,the algorithm is limited by the maximum number of fitness function calls during the search.AsJrincreases,the number of fitness function calls per iteration also increases,which reduces the overall number of iterations.This,in turn,leads to a decrease in optimization accuracy.Therefore,Jr=0.5 is the most appropriate value. Conclusion 2: WithJrfixed,asPincreases from 10 to 50,the algorithm’s performance shows a consistent improvement trend.TakingJr=0.5 as an example,for F2 and F6,the best results are obtained whenP=50.For F13,P=30 performs the best,and for F25,P=30 and 50 exhibits the best performance.This conclusion differs somewhat from Conclusion 1,mainly due to the simplicity of unimodal functions,where increasingPcan fully exploit the benefits of chaotic initialization and opposition-based initialization operators,effectively improving the algorithm’s optimization accuracy.However,multimodal and composite functions are relatively complex,making them more challenging to optimize.AsPincreases,the algorithm is also constrained by the maximum fitness function calls during the search,leading to a decrease in optimization efficiency.Therefore,P=30 is the most suitable value. 3.3.1 Performance Comparison for CEC 2013 To illustrate the merits of the proposed OLCHWOA algorithm,this section conducts comparative analyses involving several optimization algorithms,including PSO,HHO,AVOA,WOA,OLWOA,CHWOA,and OLCHWOA.The selection of these algorithms is grounded in three primary considerations.Firstly,PSO,introduced in 1995,is a well-established heuristic algorithm known for its enduring competitiveness.Secondly,both HHO and AVOA are distinguished by their simplicity in principles,minimal parameter requirements,and robust global search capabilities,rendering them highly competitive and advanced optimization algorithms that have emerged in recent years.Lastly,WOA serves as the foundational basis for OLCHWOA.Subsequently,OLWOA and CHWOA,arising from the incorporation of opposition-based learning and chaos mechanisms into WOA,respectively,naturally establish a basis for comparison with OLCHWOA,which integrates both enhancement operators.The effectiveness of each operator within OLCHWOA will be validated through ablation experiments.Following the discussion presented in Section 3.3.1,the algorithm has established the optimal parameter values forJrandP.The parameter configurations for OLCHWOA and the other six comparative algorithms are provided in Table 4. In this experiment,theMax_Fitnesswas set to 10,000.The experimental results for the seven compared algorithms on unimodal,multimodal,and composite functions are presented in Tables 5 to 7.Additionally,Table 8 summarizes the rankings of the seven algorithms.The following conclusions can be clearly drawn from the results presented in Tables 5–8. Conclusion 1: On unimodal functions,the OLCHWOA algorithm distinctly excels.This is supported by OLCHWOA achieving the highest rank on four test functions (F1,F2,F4,and F5)in terms of mean values,as well as one test function (F1) in terms of standard deviation.These results underscore OLCHWOA’s capacity to achieve optimal performance across a significant range of unimodal functions. Conclusion 2: On multimodal functions,both the OLCHWOA and AVOA algorithms exhibit the best performance in terms of algorithm convergence accuracy,with a relatively greater advantage over other algorithms.In terms of algorithm stability,OLCHWOA surpasses AVOA,as evidenced by OLCHWOA obtaining the first rank in mean and standard deviation values for five test functions,while AVOA only secures the first rank in these categories for five and three functions,respectively. Conclusion 3:Regarding composite functions,the OLCHWOA algorithm ranks second in mean values and first in standard deviation.This is prominently observed in OLCHWOA,ranking first in mean values for one test function (F24),second for three test functions (F21,F25,F26),and first in standard deviation for four test functions(F21,F25,F26,F28).These outcomes confirm the algorithm’s ability to produce commendable results when faced with the most complex benchmark problems. Conclusion 4:Among the three types of functions,the OLCHWOA algorithm demonstrates the highest performance,followed by the AVOA and OLWOA algorithms.This is primarily evident in OLCHWOA ranking first in mean values for all 28 test functions and first in standard deviation for 10 of them,placing it in a leading position,as detailed in Table 8.Additionally,Table 8 underscores that OLWOA obtains an overall superior ranking when compared to CHWOA,signifying that the opposition-based search operator exerts a more significant influence than the chaotic initialization operator in enhancing the standard WOA algorithm.This observation implies that OLCHWOA achieves a well-balanced outcome and exhibits a heightened potential for approaching the theoretical optimal solutions of these test functions.This is largely attributed to the contributions of the opposition-based search operator. 3.3.2 Comparison of Olchwoa and Other Algorithms under Different P Values The sensitivity analysis in Section 3.2 discusses the impact of the population sizePonly on the OLCHWOA algorithm.To verify whether the algorithm proposed can still maintain its relative advantage as thePincreases compared to other algorithms,this section selects eight representative test functions from CEC2013 for a comprehensive examination.Specifically,considering the scenarios ofP=50 andP=100,maintaining all other parameters identical to those outlined in Section 3.3.1.The results of these experiments are outlined in Tables 9 and 10.After conducting a meticulous analysis,the following conclusions have been drawn: Conclusion 1:WhenP=50,OLCHWOA outperforms the other six algorithms.This superiority is underscored by OLCHWOA achieving values that closely approach the global optimum on seven functions(F4,F7,F9,F13,F19,F24,and F26).The comprehensive ranking,based on both the mean and standard deviation,secures the top position. Conclusion 2: WhenP=100,OLCHWOA maintains a remarkable performance by attaining optimal values in five functions(F4,F7,F9,F24,and F26),with its comprehensive ranking,based on mean and standard deviation being first and second,respectively. Conclusion 3:Combining the results from Section 3.3.1 whenP=30,it can be observed that as the population size increases,OLCHWOA’s solution accuracy hardly decreases.Notably,its performance remains consistently competitive relative to the other comparative algorithms.OLCHWOA demonstrates efficient global search capabilities,and its effectiveness remains unhindered with the augmentation of the population size. 3.3.3 Convergence Analysis To illustrate how the improved strategy impacts algorithm convergence speed more intuitively,Fig.3 compares the average convergence curves of seven algorithms in different test functions.In view of space constraints within this paper,this section focuses on eight representative test functions.Specifically,F1 and F5 represent the convergence curves for unimodal functions.F6,F13,F15,and F20 illustrate the convergence curves for multimodal functions,while F24 and F26 depict the convergence curves for composite functions.The algorithm parameter settings align with those established in Section 3.3.1.The insights gleaned from Fig.3 are as follows: Conclusion 1:From the perspective of solution accuracy,as the number of fitness calls increases,OLCHWOA consistently gravitates towards the global optimum and ultimately attains higher convergence accuracy in comparison to the other six comparative algorithms across F1,F5,F6,F13,F20,and F24.This observation underscores the robust ability of the OLCHWOA to effectively escape local optima across the three categories of test functions.Concurrently,the OLWOA algorithm,similar to OLCHWOA,also demonstrates commendable convergence accuracy.This parallel suggests that the incorporation of opposition-based learning greatly enhances the global search capability of WOA. Conclusion 2:From the perspective of the initial population’s quality,OLCHWOA exhibits lower initial fitness values in the average convergence curves for the eight test functions,surpassing the original WOA and CHWOA.This suggests that the introduced chaotic initialization operator and opposition-based initialization operator in this paper effectively improve the quality of the initial population. Figure 3:Convergence graphs of eight functions Conclusion 3: From the perspective of convergence speed,OLCHWOA exhibits slower convergence speed,which is particularly evident in complex functions such as F13,F15,and F20.Additionally,CHWOA demonstrates the fastest convergence on functions F1,F6,F13,and F15,while OLWOA exhibits the fastest convergence on function F5.AVOA showcases the most rapid convergence on functions F20 and F26,while OLCHWOA converges most swiftly on function F24.This observation suggests that the introduction of the chaotic initialization operator and oppositionbased initialization operator,as detailed in this paper,has the potential to enhance convergence speed,especially on simpler test functions.However,their effectiveness is limited when faced with more complex problem sets.Per the“no free lunch theorem”[74],improving an algorithm’s performance in terms of both convergence speed and solution accuracy simultaneously presents a significant challenge.The proposed OLCHWOA algorithm significantly improves solution accuracy while sacrificing some convergence speed,which is one of the limitations of this algorithm. 3.3.4 Wilcoxon Rank Sum Test This paper compares the seven algorithms using theMeanandStdin Section 3.3.However,the results of only 30 independent runs cannot convincingly support the superiority of the OLCHWOA algorithm since there is still a certain probability that the algorithm performs better by chance.For this reason,this section applies the Wilcoxon rank sum test to measure the significance of the differences between different algorithms at the statistical level [75].The study takes into account the results obtained by seven algorithms independently solving 28 test functions for 30 independent runs as samples and tests them under the condition of a confidence level of 0.05 to determine if there were any significant differences between the results obtained by OLCHWOA and the other six algorithms.A significance threshold ofp<0.05 was applied.If thep-value exceeds this threshold,the optimization outcomes of the two algorithms are considered indistinguishable.Table 11 presents the results of the Wilcoxon rank-sum test.Results withp-values greater than 0.1 are shown in bold. Table 1:Description of the 28 CEC 2013 test functions Table 2:Description of the 10 CEC 2019 test functions Table 3:Sensitivity analysis of parameters Jr and P under different functions Table 4:Parameters setting of comparison algorithms Table 5:Performance comparison of seven algorithms on unimodal functions Table 7:Performance comparison of seven algorithms on composition functions Table 8:Total number of rankings of seven algorithms on the standard benchmark functions Table 9:Performance comparison of seven algorithms for CEC2013 with P=50 Table 10:Performance comparison of seven algorithms for CEC2013 with P=100 Table 11:Wilcoxon rank and p-value As indicated in Table 11,both OLCHWOA and OLWOA exhibitp-values exceeding 5%for F5,F13,and F15,while OLCHWOA and WOA attainp-values exceeding 5%for F19.Nevertheless,for the remaining test functions,all obtainedp-values are below 0.05.This statistical analysis underscores that OLCHWOA differs significantly from and outperforms the compared algorithms in these instances.Using the experiments presented in Sections 3.3.1,3.3.3,and 3.3.4,it is concluded that the OLCHWOA combined with chaos mechanism and opposition-based learning strategy has higher convergence accuracy,improved ability to jump out of local optimal solutions and better stability than six other optimization algorithms. To provide a comprehensive evaluation of the merits and drawbacks of OLCHWOA,it was compared with four recent WOA variants,including ACWOA[76],RDWOA[77],TBWOA[78],and MEWOA[79].These algorithms have been published in reputable journals and are widely recognized as benchmarks.Table 12 presents the average convergence accuracy and stability for 10 CEC2019 functions.The experimental results for these four WOA variants are sourced from reference[80],while the experiments for OLCHWOA are based on our simulation results.To ensure a fair comparison,this paper replicated the experimental conditions mentioned in the reference(Max_fitness=15000). Table 12:Performance comparison for CEC2019 As shown in the table,OLCHWOA,MEWOA,and ACWOA consistently achieved the highest average rankings in terms of mean convergence accuracy.OLCHWOA obtained five first-place rankings,MEWOA obtained four first-place rankings,and ACWOA obtained one first-place ranking.Similarly,when considering standard deviation,these algorithms also emerged as frontrunners,with 5,2,and 1 instances ranking first,respectively.The OLCHWOA algorithm consistently obtained the highest average ranking,indicating its superior global search capability compared to the other four recent WOA variants,especially when it comes to solving complex optimization problems. Given the intricacies presented by constraints in real-world optimization challenges,traditional algorithms often struggle to find solutions.This section evaluates OLCHWOA’s performance in three specific scenarios: the pressure vessel design problem,the three-bar truss design problem,and the welded beam design problem.Table 13 succinctly outlines the dimensions and the number of constraints.For more detailed information,please refer to references[79,80]. Table 13:Description of the 3 engineering design problems 3.5.1 Pressure Vessel Design Problem The objective of the pressure vessel design problem is to find the design solution with the minimum cost.Here,x1,x2,x3,x4,andf(x) represent the shell thickness,head thickness,vessel internal radius,cylindrical section length,and minimum cost,respectively.To evaluate OLCHWOA’s optimization results in the pressure vessel design problem,a comparison was made with ten algorithms selected from reference [79].OLCHWOA’s simulation results were obtained from our experiments,and the algorithm parameters were set consistent with Section 3.3.1.The maximum fitness function calls for the experiments was set to 15,000,following the conditions outlined in the reference.The experimental results in Table 14 show that OLCHWOA outperformed other algorithms,achieving the best optimization results. Table 14:Comparison of results on the pressure vessel design problem 3.5.2 Three-Bar Truss Design Problem The objective of this problem is to minimize the volume of its members while satisfying stress constraints for the bars.This endeavor entailed a meticulous comparison that involved the careful selection of 7 algorithms from reference [79].Here,x1,x2,andf(x) represent the cross-sectional areas of the two bars and the minimum volume,respectively.Similar to Section 3.5.1,the simulation results of OLCHWOA were obtained from our experiments.The experimental results in Table 15 show that algorithms such as DEPSO,WOAmM,MEWOA,and OLCHWOA obtain approximate optimal solutions.This suggests that these algorithms faced challenges despite having only three constraint conditions. Table 15:Comparison of results on the three-bar truss design problem 3.5.3 Welded Beam Design Problem The objective of optimizing this problem is to minimize the cost.Here,x1,x2,x3,x4,andf(x)respectively represent the length of the steel connecting rod,the thickness of the weld,the height of the steel rod,the thickness of the steel rod,and the minimum cost.The optimization results obtained by OLCHWOA in the welded beam design problem proposed in this paper were compared with those of eight algorithms from reference [80].The results are presented in Table 16.It was observed that e-mPSOBSA and CLPSO achieved the best results,while the OLCHWOA ranked third,providing solutions that are highly competitive among the eight algorithms. Table 16:Comparison of results on the welded beam design problem In this paper,the LSTM model serves as the benchmark model for RV prediction.LSTM utilizes memory units to store information and gate structures to discard unnecessary information,resulting in an extended memory.Unlike traditional neural networks,LSTM is composed of memory blocks.Fig.4 shows the detailed structure of a memory block,which consists of a memory unitct,input gateit,forgetting gategt,and an output gateot.According to Eqs.(19)–(24),the three gatesit,gt,otand the memory unitctcan be calculated,wherextrepresents the input at timet,htrepresents the hidden state,UandWdenote the weight matrix,bdenotes the bias term,σ(·)is a sigmoid function. Figure 4:Structure of LSTM The OLCHWO-LTSM model is designed to evaluate the effectiveness of the OLCHWOA in optimizing the LSTM-based RV prediction model.Fig.5 depicts the execution flow diagram for the RV prediction model based on OLCHWO-LTSM.The diagram includes the following modules:Data Processing Module,Model Training Module,Parameter Optimization Module,and Evaluation Module. Data Processing Module:Firstly,the dataset is split into a training and a testing set.The first 90%of the data is used to train the LSTM model,while the remaining 10%is used to test and validate the algorithm’s predictive performance.Prior to inputting variables,data normalization is performed. Model Training Module:To enhance the model’s ability to make nonlinear predictions,the LSTM architecture utilizes a three-layer structure.This configuration includes a single LSTM hidden layer along with two fully connected layers,as visually represented in Fig.5. Parameter Optimization Module:Within this module,the OLCHWOA algorithm is strategically utilized to coordinate the optimization of four crucial hyperparameters inherent to the LSTM model.These parameters include the number of nodes within the LSTM layer (ls1),the number of nodes in the fully connected layer (ls2),the dropout parameter of the LSTM (dp),and the proportion of the validation set(vs).In essence,this optimization endeavor aims to identify the optimal parameter configuration,thereby improving the predictive capacity of the LSTM model.The parameter search intervals specified for these hyperparameters of interest are as follows:ls1∈[1,200],ls2∈[1,200],dp∈[0.01,0.9],andvs.∈[0.01,1].It is worth noting that the remaining parameters governing the operation of the LSTM model adhere to a uniform standardization.This involves setting a fixed batch size of 512,running 50 iterations,using the Adam optimizer,and employing the Mean Squared Error(MSE)as the loss function. Evaluation Module diligently deploys the fine-tuned LSTM model to facilitate the prediction of RV values. Figure 5:RV prediction framework based on OLCHWOA-LSTM 4.3.1 Data The China Securities Index 300 (CSI 300) futures were introduced on April 16,2010,and have since become the most actively traded stock index futures product in the Chinese market.This article focuses on high-frequency data of the CSI 300 Index,covering the period from January 04,2010,to March 20,2020,which comprises a total of 2,480 trading days.The specific calculation of RV data is based on Anderson et al.[81].The process is shown below. Using a sampling frequency of 5 min per trading day and obtaining the opening and closing prices every 5 min,48 high-frequency returns can be obtained for a single trading day.In total,there would be 119,040 high-frequency returns.Price series are expressed asPt,d,t=1,2,3,...,2480,d=0,1,2,3,...,48,wherePt,0represents the opening price at 9:30 on trading dayt,andPt,drepresents the closing price at thed-th5-minute mark.Rtindicates the daily return,which is calculated using the closing prices of the two adjacent trading days.Here is the formula. wheret=1,2,3,...,2480.Likewise,the high frequency returnRt,dfor the 5-min periodd-th on daytis calculated as indicated by the following formula: wheret=1,2,3,...,2480,andd=0,1,2,3,...,48. The RV on daytcan be estimated by the following formula: Jumpiness is a concept that pertains to the substantial shifts in volatility observed when significant information emerges within financial markets,often resulting in pronounced jumps.Drawing upon this notion of jumpiness,this study introduces continuous volatility denoted asCand jump volatility denoted asJas inputs into the LSTM model.Additionally,volatility exhibits asymmetric characteristics,indicating that positive and negative news in the market have varying effects on volatility.Negative news tends to increase volatility.Based on this perspective,two predictive variables are introduced:the absolute value of volatility on down days,denoted asrdM,and the absolute value of daily volatility,denoted asrdabs. To evaluate the performance of the OLCHWOA-LSTM model in predicting RV for the CSI 300 Index,the previous period’s realized volatility(RV),continuous volatility(C),jump volatility(J),the absolute value of the down day’s volatility(rdM,),and the absolute value of the daily volatility(rdabs)are used as inputs.The model is designed to forecast RV for the next trading day based on a time series of five variables spanning the past 22 trading days.The data is sourced from the RESET HF Database,and Table 17 provides descriptive statistics for each indicator. Table 17:Descriptive statistics for predicting RV variables 4.3.2 Evaluation Metrics For the purpose of better reflecting the predicted value error,this paper uses four commonly used evaluation metrics,namely Mean Absolute Error(MAE),Mean Square Error(MSE),Heteroscedasticity adjusted MAE (HMAE),and Heteroscedasticity adjusted MSE (HMSE).The formulas for calculating the four evaluation indicators are as follows: In this section,the TensorFlow2.2 version is used to build the deep learning framework of LSTM,and the rest of the computer configuration is the same as in Section 3.ARFIMA and HAR are used as comparison models in this paper.The following is a brief introduction to their principles. HAR is based on the heterogeneous market hypothesis.In order to present volatility generated by different traders,the HAR model uses daily RV,weekly RV,and monthly RV to show volatility generated by short-term,medium-term,and long-term traders,respectively.The model uses 1-day,5-days,and 22-days as the trading time scales for each of the three investors.This model is described by the following formula: whereRV1,RV5,andRV22denote daily RV,weekly RV,and monthly RV,respectively. The ARFIMA model takes into account both long-term and short-term memory of time series.The specific formula is as follows: whereμis the series mean,|d|<0.5,(1-L)dis the fractional difference factor,orderpandqare short memory factors,and orderdindicates long-term memory. The HAR and ARFIMA models are the most classic RV prediction models and are considered reasonable as comparative models.They are implemented using the HARModel package and forecast package in the R language.LSTM is the basic model for RV prediction.To evaluate the performance of OLCHWOA in RV prediction,the PSO-LSTM model,HHO-LSTM model,WOA-LSTM model,OLWOA-LSTM model,CHWOA-LSTM model,and OLCHWOA-LSTM model were compared. Table 18 lists the parameter settings of the above 9 models,which are set according to the results of the training set experiments.The HAR model contains four parameters:beta0,beta1,beta5,andbeta22,which represent the average level of volatility and the impact of three different frequencies of market participants on RV,daily,weekly,and monthly,respectively.The results of parameter estimation show that all four coefficients are significant at the 1%confidence level.The ARFIMA model contains eight parameters.The difference termdis a score between-0.5 and 0.5,which is used to capture long-term memory and can be obtained by calculating the Hurst exponent,and its formula isd=Hurst-0.5.AR1toAR5are the autocorrelation coefficients of the model,whileMA1andMA2are the moving average coefficients,and these eight parameters are determined by the AIC criterion.LSTM model parametersls1,ls2,dp,andvshave the same meanings as those in Section 4.2.Additionally,PandMax_Fitnesshave the same meanings as those described in Section 3.2.There are several types of LSTM models that are based on three-layer network structures,but their optimal parameters vary. Table 18:Parameters setting of nine models Table 19 shows the experimental results of the nine comparison models mentioned in Section 4.4 on the RV dataset.According to Table 19,the following conclusions can be drawn: Table 19:Comparison of forecast results on four metrics Conclusion 1:The OLCHWOA-LSTM model has the best performance.This is demonstrated by the fact that the OLCHWOA-LSTM model outperforms the HAR model,ARFIMA model,LSTM model,PSO-LSTM model,HHO-LSTM model,WOA-LSTM model,OLWOA-LSTM model,and CHWOA-LSTM model in indicator MSE by 26.19%,25.58%,19.05%,15.31%,15.71%,6.76%,6.73%,and 5.11%,respectively.MAE,HMSE,and HMAE also exhibit varying degrees of improvement.Consequently,the improved algorithm has advantages not only on standard test functions,but also in practical application problems. Conclusion 2:Models optimized for LSTM parameters based on heuristic algorithms generally outperform classical models such as HAR and ARFIMA.It appears that the four metrics of OLWOALSTM,CHWOA-LSTM,and OLCHWOA-LSTM all outperform HAR and ARFIMA. Conclusion 3: ARFIMA and HAR perform better than standard LSTMs.It is evident from this example that LSTM models require parameter optimization.It is impossible to exploit LSTM’s nonlinear prediction capabilities without optimization.This further demonstrates the necessity of conducting this study. The OLCHWOA-LSTM model presented is an interesting study that deepens understanding of how LSTM technology contributes to research on RV.Research in the past has focused on optimizing LSTM hyperparameters using an exhaustive method [82–84],which had the shortcomings of being time-consuming and yielding poor results.This limitation has just been addressed in this study.Moreover,the OLCHWOA-LSTM model is highly competitive in predicting RV problems.There are two main reasons for this. First,the OLCHWOA-LSTM model can capture the nonlinear characteristics of RV and,therefore,statistically significantly outperforms HAR and ARFIMA,which are currently the most commonly used models in the industry.For RV prediction,ARFIMA can capture long-term memory and aggregation characteristics,while HAR can capture both aggregation and market heterogeneity characteristics.Compared to HAR and ARFIMA,the OLCHWOA-LSTM model can effectively capture the long memory and aggregation characteristics of RV by leveraging the inherent capabilities of LSTM.Furthermore,it can also account for the heterogeneity and asymmetry of RV through its input variables.Additionally,the LSTM model is capable of capturing nonlinearity in the time series of RV.The OLCHWOA-LSTM model not only exhibits the characteristics of ARFIMA and HAR models,but it also offers the advantages of nonlinear feature acquisition and asymmetric capture,making it statistically significant. Second,the OLCHWOA-LSTM model is statistically significantly superior to other machine learning models as a result of its hyperparameter optimization.RV prediction is essentially a time series prediction problem,and the LSTM model is specifically designed to handle it.By automating the hyperparameter optimization process,OLCHWOA further enhances the benefits of LSTM when dealing with time series data. This paper proposes a streamlined and efficient enhancement to the Whale Optimization Algorithm (WOA),called OLCHWOA in this manuscript.This improvement incorporates several new operators,including a chaotic initialization operator based on logistic sequences,an oppositionbased initialization operator,and an elite opposition-based learning operator based on the parameterJr.These operators are introduced into the original WOA algorithm.The proposed algorithm offers several advantages:(1)The chaotic initialization and opposition-based initialization operators enhance the diversity of the OLCHWOA population in the early stages,endowing the algorithm with strong exploration capabilities and contributing to improved search efficiency.(2) The oppositionbased search operator enables the algorithm to maintain robust local exploitation capabilities even in the later iterations,allowing it to attain superior solution accuracy in optimization tasks and a greater probability of finding the global optima of complex functions.(3)OLCHWOA is straightforward to implement and delivers satisfactory performance.The introduced enhancement strategies improve the algorithm’s search efficiency without compromising its time complexity.(4) Extensive experiments and diverse application scenarios validate the reliability of the OLCHWOA algorithm.Four sets of experiments were conducted to investigate the performance of OLCHWOA in solving complex optimization problems.The results indicated thatJr=0.5/P=30 represents the most competitive parameter setting.Firstly,the impact of different combinations of parameters,including the jump rateJrand the population sizeP,on the performance of OLCHWOA are discussed.Secondly,the proposed OLCHWOA algorithm is evaluated on 28 CEC2013 standard test functions and 10 CEC2019 standard test functions.It was compared with various types of heuristic algorithms,including traditional PSO,WOA,recently proposed HHO and AVOA,partially improved algorithms of WOA like OLWOA,CHWOA,and variants of WOA algorithms such as ACWOA.Simulation results demonstrated that OLCHWOA outperformed the other compared algorithms significantly in terms of convergence accuracy and stability.Furthermore,to assess the algorithm’s capability to solve real-world complex optimization problems,OLCHWOA was experimented with three constrained engineering design problems.The results indicated that the proposed algorithm also exhibited superiority and applicability.Lastly,OLCHWOA was used to perform hyperparameter auto-tuning for the LSTM model and applied to the RV prediction problem.Experimental results based on the CSI 300 dataset demonstrated that the OLCHWOA-LSTM model achieved the highest accuracy and robustness in RV prediction.This illustrates the high competitiveness and efficient search capability of our algorithm in solving practical problems under similar conditions. Nevertheless,OLCHWOA still exhibits certain limitations.Firstly,although the proposed algorithm exhibits a high global search capability,it does not consistently achieve top performance across all functions in CEC2013 and CEC2019.It struggles to converge to the optimum in certain highdimensional functions.Furthermore,the use of the opposition-based learning strategy in OLCHWOA enables it to maintain strong exploitation capabilities in subsequent iterations.Nevertheless,this strategic inclusion increases the frequency of fitness function invocation per iteration,which slows down convergence,especially when dealing with complex composite functions. In future investigations,our team will focus on three specific domains to further enhance the performance of OLCHWOA: (1) The jump rateJrin this paper was set as a fixed value,but it may not necessarily be the optimal choice.Therefore,the adaptive adjustment mechanism of parameterJrwill be considered.(2)OLCHWOA exhibits a slower convergence rate,and research on enhancing its convergence speed will be prioritized.(3)While OLCHWOA has demonstrated success in continuous function optimization,additional validation is required to evaluate its effectiveness in addressing discrete optimization challenges. Acknowledgement:The authors thank the anonymous reviewers and editors for their valuable comments and suggestions. Funding Statement:The National Natural Science Foundation of China(Grant No.81973791)funded this research. Author Contributions:The authors confirm contribution to the paper as follows:study conception and design:Xiang Wang,Yibin Guo;data collection:Liangsa Wang,Han Li;analysis and interpretation of results: Xiang Wang,Liangsa Wang,Han Li;draft manuscript preparation: Liangsa Wang,Han Li.All authors reviewed the results and approved the final version of the manuscript. Availability of Data and Materials:Researchers can obtain the data supporting the findings of this study by contacting the corresponding authors. Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.2.4 Olchwoa Algorithm
2.5 Computational Complexity
3 Simulation Experiment
3.1 Benchmark Functions
3.2 Experiment 1:Parameter Sensitivity Analysis
3.3 Experiment 2:Comparison of Olchwoa with Other Metaheuristic Algorithms
3.4 Experiment 3:Comparisons between Olchwoa and Other Whale Variants
3.5 Experiment 4:Application of Engineering Problems
4 Application of Olchwoa Algorithm in RV Forecasting
4.1 Lstm Model
4.2 Rv Prediction Model Based on OLCHWO-LTSM
4.3 Experimental Settings
4.4 Environment,Comparison Models and Their Parameter Settings
4.5 Experimental Results
4.6 Discussion
5 Conclusions and Future Works
Computers Materials&Continua2023年12期