1. Institute of Artificial Intelligence & Robotics (IAIR), Key Laboratory of Traffic Safety on Track of Ministry of Education, School of Traffic and Transportation Engineering, Central South University, Changsha 410075, China
2. School of Mechatronic Engineering, Hunan Agricultural University, Changsha 410128, China
csuliuhui@csu.edu.cn
Show less
History+
Received
Accepted
Published
2024-07-08
2024-09-23
2025-06-15
Issue Date
Revised Date
2025-02-10
PDF
(6614KB)
Abstract
The futures trading market is an important part of the financial markets and soybeans are one of the most strategically important crops in the world. How to predict soybean future price is a challenging topic being studied by many researchers. This paper proposes a novel hybrid soybean future price prediction model which includes two stages of data preprocessing and deep learning prediction. In the data preprocessing stage, futures price series are decomposed into subsequences using the ICEEMDAN (improved complete ensemble empirical mode decomposition with adaptive noise) method. The Lempel-Ziv complexity determination method was then used to identify and reconstruct high-frequency subsequences. Finally, the high frequency component is decomposed secondarily using variational mode decomposition optimized by beluga whale optimization algorithm. In the deep learning prediction stage, a deep extreme learning machine optimized by the sparrow search algorithm was used to obtain the prediction results of all subseries and reconstructs them to obtain the final soybean future price prediction results. Based on the experimental results of soybean future price markets in China, Italy, and the United States, it was found that the hybrid method proposed provides superior performance in terms of prediction accuracy and robustness.
As one of the most important crops in the world, soybean prices are of concern to governments, investors and farmers. However, soybean prices are affected by a combination of many factors and have great uncertainty[1]. Governments find it difficult to monitor and predict the future trend of soybeans, and are unable to formulate appropriate policies in response. Investors lack scientific decision-making methods to guide them and can only trade soybeans based on inadequately informed speculations, which exacerbates the instability of soybean prices. Farmers also need scientific and practical price forecasts to help maximize profits and provide confidence in their forward planning[2]. Global commodity market trends, investor perceptions and political events, and currency factors have influenced soybean future price, and the price series are highly nonlinear and volatile. Therefore, it is of great significance to study how to develop a general and effective soybean future price forecasting model[3].
Many scholars have conducted relevant studies on the forecasting of agricultural prices with good results. The earliest invented method was statistically-based. The most used statistical method is the autoregressive integrated moving average (ARIMA) model. Darekar & Reddy[4] and Panasa et al.[5] used the ARIMA model to forecast monthly soybean prices and maize prices in India to provide a reasonably predictive analysis. Bhardwaj et al.[6] first used the Box-Jenkins autoregressive integrated moving average model to forecast agricultural prices in New Delhi and found that the ability to capture the volatility of the data was not satisfactory, and then used the generalized autoregressive conditional heteroscedastic model for forecasting and the results obtained were better than ARIMA in all indicators, proving that the model has better forecasting performance. In addition, the accuracy of forecasting can also be further improved by combining models from different statistical methods. Şahinli[7] proposed a Holt-Winters model combined with ARIMA to improve the accuracy of forecasting, and the performance of the proposed model was more reliable compared with ARIMA alone.
In recent years, machine learning has become increasingly popular and intelligent models are widely used in endeavors such as image processing[8], price prediction[9], and fault diagnosis[10]. Mahto et al.[11].and Jaiswal et al.[12] used artificial neural networks and deep long-term short-term memory-based models on predicting agricultural price series and compared them with ARIMA, respectively. Zong & Zhu[13] used radial basis function (RBF) and back propagation neural networks to forecast Chinese agricultural prices. Xu[14] conducted univariate modeling of spot prices and bivariate modeling of spot and futures prices with neural networks to forecast the price series of the last 500 agricultural commodities in the USA. Xu & Zhang[15] evaluated the impact of different model settings of neural networks, such as the number of hidden neurons and the proportion of training and validation sets, on the prediction performance of the canola and soybean oil price forecasting models.
It is evident that the performance of single models is still not robust for high-complexity and noisy time series, so most of the research at this stage integrates preprocessing[16], optimization algorithms[17] and postprocessing[18] to improve the performance of hybrid models. Zhang & Na[19] introduced a hybrid model that includes fuzzy information granulation and uses the mind evolutionary algorithm to optimize the hybrid model with the support vector machine and concluded that the prediction of the UN Food and Agriculture Organization issued price. Li et al.[20] proposed a model combining wavelet transformation and exponential smoothing and experimentally demonstrated superiority over long short-term memory (LSTM) and support vector regression. Wang et al.[21] proposed a combined forecasting method based on a global optimization method, using three decomposition algorithms and five forecasting models to forecast soybean and maize futures prices, demonstrating the superiority of the combined model. Liang & Jia[22] introduced a hybrid Gray Wolf Optimizer-convolutional neural network-LSTM model for forecasting prices of maize, soybean and others, and introduced the Baidu index, Google trends and computed transfer entropy to improve real-time forecasting. Zhang et al.[23] proposed a quantile regression-RBF neural network model using a hybrid algorithm, combining gradient descent and genetic algorithms, to achieve global and local search. In particular, the introduction of this algorithm provided enhanced optimization capabilities for the quantile regression-RBF model, thus further improving the performance and stability of the model.
At this stage, price forecasting of agricultural products has achieved satisfactory forecasting accuracy, but in other price forecasting research fields, secondary decomposition and other optimization methods have been introduced to optimize the forecasting accuracy of time-series data. Liu & Long[24] used a novel framework for forecasting stock closing prices with higher prediction results compared with established models. The deep hybrid framework consists of a data processing component, a deep learning predictor component and a predictor optimization method. The empirical wavelet transform is used for data preprocessing. The LSTM is the main part of the hybrid framework and is optimized by a combination of dropout strategies and particle swarm optimization algorithms. Sun & Huang[25] proposed a secondary decomposition model based on empirical mode decomposition (EMD) and variational mode decomposition (VMD), and used LSTM to forecast the entire carbon market in China. Considering price dynamics that include cyclical growth, seasonal variations and irregular fluctuations, Zhu et al.[26] proposed a hybrid model combining a loess-based seasonal-trend decomposition procedure, support vector regression and autoregressive moving average to forecast hog prices for the next farming cycle. Liu et al.[27] proposed a hybrid model consisting of secondary decomposition, ensemble method and error correction. The decomposition is performed by wavelet decomposition to obtain the wind speed subseries. The SampEn algorithm is used to estimate the unpredictability of the subsequences. The most unpredictable subsequences are decomposed again by the VMD. The subsequences are obtained as predicted subsequences by the ensemble method neuron network, and the modified predicted sequence is reconstructed to obtain the final predicted sequence. Liu & Zhang[28] proposed a hybrid AQI time series forecasting model based on secondary decomposition, imperialist competitive algorithm, feature selection and echo state network. This hybrid model has broad application prospects and research value in the field of AQI forecasting.
In summary, although scholars have achieved good results in soybean future price forecasting, there is still scope for exploration of hybrid methods. The most mainstream hybrid learning methods are currently based on standard machine learning or deep learning with signal preprocessing methods or optimization algorithms. This study uses a secondary decomposition-based signal preprocessing method that combines a highly adaptive optimization algorithm and a neural network model with powerful information mining capabilities. Based on this, a new forecasting method is proposed and applied to the field of agricultural futures forecasting for the first time.
In this paper, an improved deep learning model for soybean future prices prediction is proposed, which includes ICEEMDAN (improved complete ensemble empirical mode decomposition with adaptive noise), Lempel-Ziv complexity (LZC) determination method, variational mode decomposition optimized by beluga whale optimization (BWO) algorithm (BVMD), sparrow search algorithm (SSA), and deep extreme learning machine (DELM), and is designated as ICEEMDAN-LZC-BVMD-SSA-DELM. The main contributions of this study can be summarized as follows.
(1) We propose a new hybrid deep learning model for high-accuracy forecasting of international soybean future closing price. The proposed deep learning forecasting model framework consists of ICEEMDAN decomposition of soybean futures closing price series, reconstruction of high-complexity subsequences based on LZC method, VMD secondary decomposition based on the BWO algorithm, the DELM forecasting algorithm and hyperparameter optimization based on the SSA. The experimental results on three soybean future data sets from China, Italy, and the United States reflect the ability of the proposed hybrid model to forecast with high accuracy. In addition, this forecasting method has good generalizability and can be extended to other work.
(2) A hybrid data preprocessing strategy of soybean future price was applied by integrating the LZC evaluation method. The ICEEMDAN algorithm was first used to decompose the original soybean future price. Then the LZC evaluation method was used to evaluate the resulting high-frequency subsequences, which was reconstructed and decomposed to further reduce the complexity of the series and help improve the accuracy of the forecast.
(3) The VMD algorithm optimized by the BWO algorithm is proposed as the secondary decomposition algorithm. Adopting the beluga whale optimization algorithm to optimize the parameters of the decomposition level K and penalty factor α of VMD can better preserve the signal characteristics and reduce the modal mixing problem so that the subsequent predicted soybean future price subsequence is smoother.
(4) The deep learning predictor for the secondary decomposition subsequences is optimized by the SSA algorithm, which consists of multiple extreme learning machine-auto encoder (ELM-AE) stacks with randomly generated input layer weights and thresholds. By optimizing the predictor using the SSA algorithm, we significantly improved the accuracy of soybean future price prediction.
2 Methodology
2.1 Framework of the proposed hybrid model
The framework of the hybrid model ICEEMDAN-LZC-BVMD-SSA-DELM proposed in this paper is shown in Fig.1. The process of the hybrid model is summarized as follows.
Fig.1(a): The soybean future price series from China, Italy, and the United States were decomposed using the ICEEMDAN algorithm to obtain multiple primary decomposition subsequences.
Fig.1(b): The complexity of each subsequence of soybean future price decomposed by ICEEMDAN is evaluated by the LZC determination method, and the selected subsequences with high complexity are reconstructed.
Fig.1(c): The reconstructed high-complexity sequences are decomposed again using VMD optimized by BWO algorithm to further decompose them into lower-complexity price subsequences. The price subsequences of low complexity derived from the LZC determination method is assembled with the price subsequences obtained from the BVMD secondary decomposition to obtain the combined soybean future price subsequences.
Fig.1(d): The DELM optimized by SSA is used to forecast all the subsequences obtained from the secondary decomposition, and then the forecasting results of all the subsequences are reconstructed to get the final soybean future price forecasts, and the weights and thresholds of DELM are optimized by the SSA algorithm.
2.2 Improved complete ensemble empirical mode decomposition with adaptive noise
To effectively suppress problems such as the modal blending phenomenon between the modes of the EMD algorithm and further improve the accuracy and stability of the decomposition, the ICEEMDAN algorithm includes adaptive noise suppression techniques[29]. The ICEEMDAN decomposition proceeds as follows.
Step 1: Introduce the operator in the EMD algorithm.
Step 2: Construct the noise signal and use EMD to calculate the first-order residual , which is added in the initial stage and is used to remove the noise. is the number of times the noise is added.
Step 3: When is 1, the first-order Intrinsic Mode Function (IMF) component is found, which is .
Step 4: Using the average of the local means, the second-order residuals and the second-order IMF components are found as:
Step 5: Derive j-th order mode as:
Step 6: Repeat Step 5 for all residuals and IMF components.
2.3 Complexity evaluation and data reconstruction
2.3.1 Lempel-Ziv complexity evaluation
The LZC algorithm was proposed by Lempel and Ziv to evaluate the randomness and chaos of sequences of a particular length[30]. The greater the complexity of a sequence, the more it converges to a random state and the richer the frequency component contained in the sequence[31]. It is widely used in applications such as electrocardiography signals, gene sequences and spoken texts.
2.3.2 Data reconstruction
Based on the complexity results of the subsequence and the Lempel-Ziv complexity principle, the process of dividing the high and low frequencies of soybean future price into the three markets and the reconstruction of the features are shown below.
Step 1: Compute the Lempel-Ziv complexity Ci for each subsequence Si, i = 1, 2, ..., m.
Step 2: Set the critical value as and find out the first k subsequences satisfying the following formula.
Step 3: Identify the sequence between subsequences S1 to Sk as high-frequency sequences and the sequence between subsequences Sk+1 to Sm as low-frequency sequences.
Step 4: The sequences between S1 and Sk are summed and reconstructed as the final high-frequency reconstructed sequence SHigh, and the data between Sk+1 and Sm are summed and reconstructed as the final low-frequency reconstructed sequence SLow.
Therefore, the determination of k in the subsequence reconstruction process based on the Lempel-Ziv complexity is crucial.
2.4 Improved variational model decomposition
2.4.1 Beluga whale optimization algorithm
The BWO algorithm is a meta-heuristic algorithm inspired by the observation and simulation of beluga whales during swimming, feeding and dying, which correspond to the three phases of exploration, exploitation and whale fall. In addition, a Lévy flight strategy is introduced in the development phase to keep it from being limited to local optimality in this phase further[32].
2.4.2 Variational mode decomposition
VMD is a non-recursive signal processing algorithm[33], which decomposes the raw signal into a series of patterns with a specific spectral domain bandwidth, which is excellent for processing non-smooth, nonlinear and noisy signals.
For VMD, to obtain the best decomposition results, it is necessary to find the most suitable number of modes K and penalty factor α. K is the parameter that controls the number of modes obtained from the decomposition. When the value of K is too large, although more modes can be decomposed and the local characteristics of the signal can be better preserved, it also tends to over-fit the noise and local fluctuations, and even the phenomenon of modal confusion can occur. However, when the value of K is too small, the signal is decomposed into fewer modes, which is helpful in preserving the global features of the signal, but some detailed information is also lost, and a complete decomposition cannot be achieved.
Secondly, α is the parameter that controls the bandwidth of each mode. When the value of α is too large, it allows the bandwidth of the modes to be wider and better preserve the global features, but there will be excessive smoothing of the signal, which will cause some loss of details and local features. When the value of α is too small, it narrows the bandwidth of each mode, allowing better preservation of details and local features, but may also spread the signal too much[27].
2.4.3 Parameter optimization for variational mode decomposition
To investigate how to choose the appropriate K and α to achieve an optimal balance between signal feature preservation and noise removal, BVMD is proposed, with the pseudocode shown in Algorithm 1.
In this study, the fitness function for VMD optimized by the BWO algorithm is the envelope entropy[34]. The smaller the envelope entropy, the more signal features are retained, the more noise signals are removed, and the VMD effect is more thorough.
The optimal K and α of the VMD are determined by comparing the corresponding fitness function values at different update positions. The formula of envelope entropy is:
where, is the envelope signal, is the normalized form of , and is envelope entropy.
Although several advantages of VMD and its improved models have been reported in the literature, the application of hybrid forecasting models based on VMD is not yet widespread in the field of agricultural price forecasting. Therefore, there is a need for researchers to conduct more relevant studies in the future to explore its potential for application in agricultural price forecasting and to further improve its performance.
2.5 SSA-optimized DELM
2.5.1 Sparrow search algorithm
The SSA algorithm was derived from the foraging and anti-predatory behavior of sparrow populations[35]. In the SSA algorithm, the magnitude of the fitness value indicates the strength of the finder to search for food, and finder position is updated during iteration as follows[35].
where, t is the number of current iterations; j is the number of dimensions, is the maximum number of iterations, α is a random number in (0,1], ST is the safety value and R2 is the warning value. means the population is in a safe area and the finder can forage randomly, whereas means there is a predator around the population and it needs to be moved immediately to bring safe area for foraging.
If joiners in the population perceive that a finder has foraged for better food, they will immediately compete for it. If the joiners are successful, they are given access to finder food, otherwise they will continue to watch the finder for food. Update the location of the joiners using this formula[35]:
where, A is a 1 × d matrix with each element randomly assigned 1 or −1. means that the i-th accession with a low fitness value is not getting food and needs to go elsewhere to feed. In addition, 10%–20% of sparrows in the entire population are randomly generated, named guards, and they randomly generate initial positions, as Li et al.[35]:
where, is the step control parameter; is the position of the safest sparrow; is the adaptation value of the current individual sparrow; and denote the current best and worst adaptation values, respectively, and when , it means that the sparrow is in the edge position at this time and is relatively threatened; means that the middle sparrow is aware of predator arrival at this time and thus tries to approach the nearby sparrow; is the minimum constant.
2.5.2 Deep extreme learning machine
The ELM has the advantages of strong generalization ability and fast learning speed[36]. For high-dimensional feature expressions, the output weight β of the hidden layer is expressed as:
where, is the connection weight of the hidden node to the output node, C is the regularization parameter and X is the input data.
For the equal dimensional feature expression, the output weight β of the hidden layer is expressed as:
The DELM is a stack of multiple ELM-AEs. The DELM enables the mapping of data features to improve both the forecasting accuracy and the generalization capability of the model[37]. The structure of DELM is shown in Fig.2.
In the training process of the DELM, the training time series is generally used as the output of the first ELM-AE layer to find the output weight β1. The output of the hidden layer of the DELM-AE model is then used as the input data of the second ELM-AE layer[37]. This method is repeated for each layer of the ELM-AE.
2.5.3 The deep extreme learning machine optimized by the sparrow search algorithm
The SSA is used to perform parameter optimization of the input layer weight and bias of the DELM. The optimization process flow is shown in Fig.3.
3 Experiments and results
3.1 Soybean future data
This paper uses experimental time-series data from three typical soybean markets, China, Italy, and the United States, to validate the model. The three market price studies were developed based on soybean future price, which are widely used in the study of soybean prices as indicators of uniform markets, complete data and timely updates.
The data resolution in the experiment was a single day, and the specific price movements and detailed characteristics are shown in Fig.4–Fig.6 and Tab.1, respectively. Specifically, soybean prices for each market included 1500 data values, divided into two parts: a training set and a test set. The training set was data values 1–1200 and the test set was data values 1201–1500. The Chinese soybean future price data set was chosen from the closing prices of the main contract of the Yellow Soybean 1 futures on the Dalian Exchange in China from 29 August 2016 to 2 November 2022. The United States soybean future price data set was selected from the closing prices of soybean future on the Chicago Board of Trade for the period 30 March 2017 to 27 January 2023. Additionally, to demonstrate model generalizability across financial instruments, the Italy soybean ETF (exchange traded fund) price data set was introduced. ETFs typically track specific indices or commodities, such as soybeans, which are directly linked to futures contracts. However, unlike futures contracts that are traded in the futures market, ETFs are listed and traded on exchanges like stocks, allowing for real-time buying and selling. On the WisdomTree Soybeans ETF, traded on the Italian Stock Exchange (Borsa Italiana), for the period from 19 June 2015 to 29 October 2021. The European Soybean ETF (exchange-traded fund) tracks the price of soybeans, and the unit of measurement is the ‘share,’ which represents a fraction of the total value of the ETF, linked to soybean future. All experiments were conducted on the MATLAB R2021b platform, which was run on a computer with Intel Core i5-12600K 3.70 GHz, 32G RAM, RTX 3070ti and Windows 10 Professional Edition operating system.
3.2 Experimental evaluation metrics
3.2.1 Performance evaluation indexes
Four main evaluation metrics were used to evaluate the predictions obtained in this study, including mean absolute error (MAE), mean absolute percentage error (MAPE), root mean square error (RMSE) and R2 calculated as:
where, N is the number of samples, is the actual closing price and is the predicted closing price. RMSE, MAE and MAPE are inversely proportional to prediction accuracy, and R2 is positively proportional to prediction accuracy. The optimal values of MAE, MAPE, RMSE and R2 are 0, 0%, 0 and 1, respectively.
3.2.2 Performance improvement indexes
To compare the performance of the models, the percentage improvement of the evaluation indicators is used in this study to further evaluate the experimental results.
The percentage improvement of each evaluation indicator was calculated according to the formulae:
A metric with a subscript of 1 is the performance of the baseline model, and a metric with a subscript of 2 is the performance of the proposed model.
3.3 Model analysis
To demonstrate the superiority of the models proposed in this paper, 14 models were selected for comparison, including six single machine learning prediction models and eight hybrid prediction models. The single prediction models include ELM, RBF, deep belief network (DBN), LSTM, gated recurrent unit (GRU) and DELM. The hybrid prediction models include SSA-DELM, ICEEMDAN-DELM, ICEEMDAN-SSA-DELM, VMD-DELM, BVMD-DELM, BVMD-SSA-DELM, ICEEMDAN-LZC-BVMD-DELM, and the proposed model. For the simplicity of illustration and comparison, in some figures, S1 to S6 is used to denote the six single forecasting models above, and D1 to D8 to denote the eight hybrid forecasting models.
The results of the soybean future price error assessment for the single forecasting models and the hybrid forecasting models are shown in Tab.2.
3.3.1 Comparative analysis of single models
To verify that DELM can accurately predict soybean future prices, experiments were conducted comparing DELM with other single models.
The ELM is a fast feed-forward neural network that reduces training time by randomly generating weights for the hidden layers. The RBF is a generating function that maps the input space to a high-dimensional space and is used to handle nonlinear classification and regression problems. The DBN is a multilayer neural network model consisting of multiple stacked restricted Boltzmann machines. The DBN has a self-encoder feature that The LSTM and GRU are variants of recurrent neural networks for processing time-series data. The LSTM has a memory capability to store historical information, and the GRU is a simplified version of the LSTM.
To visualize the forecasting performance of all single models, prediction error plots are given in Fig.7. The prediction results of a single model for the Chinese soybean future price series are given in Fig.8. Tab.2 shows the accurate results of all the single models.
After the experiments, this paper found that for soybean futures price forecasting in the three markets, compared with the other five neural network models, the DELM had better evaluation performance, with MAPE values of 0.941, 1.23 and 1.32 for soybean future price forecasting in the three markets China, Italy, and the United States, respectively, which were the lowest values. In the error evaluation plot, DELM has the smallest variation, indicating that it outperforms the other models, which proves that DELM has the best stability and forecasting performance among the six single models proposed.
3.3.2 Comparative analysis of decomposition methods
In this study, we first calculated the Lempel-Ziv complexity of soybean future data for three markets based on the principles of the LZC algorithm (Tab.3).
According to the calculation results of the complexity of data subsequences in each market in Tab.3, where “Res” is presented as the residuals—indicating the differences between the observed and predicted values. The calculation results of η with the change of k value in the three markets are shown in Tab.4.
Subsequences were divided into high and low-frequency sequences based on complexity similarity[38]. When k is 4, the value of η will be greater than the set threshold (0.8). This shows that the first four subsequences obtained by ICEEMDAN decomposition are all high-frequency sequences, which can be reconstructed into a new subsequence. At the same time, the remaining ones represent low-frequency sequences and residual sequences and are retained. The process of the secondary decomposition for Chinese market is shown in Fig.9. In the next step, the decomposed subsequences are used in forecasting models to assess their impact on prediction accuracy, as analyzed below.
Fig.9: Decomposition algorithms can help deep learning models to better handle large-scale data and can also enhance the generalization ability of neural networks, thus reducing prediction accuracy. The soybean future price is full of uncertainty and randomness, so decomposition is imperative. In this study, the experimental results of soybean future price prediction using ICEEMDAN and using the VMD algorithm in combination with DELM are compared with the prediction results of DELM alone, with the specific percentage improvements shown in Tab.5. It is concluded that the use of the decomposition algorithm reduces the degree of nonlinearity in the raw data and improves the forecasting accuracy of soybean future price. Using MAPE as the evaluation metric, the ICEEMDAN-DELM for the three markets improved the forecasting accuracy by 36.3% compared with DELM, and the VMD-DELM improved the forecasting accuracy by 24.9% compared with DELM, which represents the effectiveness of the data decomposition algorithm.
Fig.9: The secondary decomposition method can decompose high-complexity time-series data into multiple stable regular subsequences efficiently. In contrast, ICEEMDAN is an adaptive decomposition algorithm, VMD requires empirical manual adjustment of the decomposition level K and penalty factor α, which increases the operational difficulty of this hybrid model and the risk of experimental failure due to blind adjustment. Therefore, BVMD was introduced to decompose the decomposed ICEEMDAN high-frequency subsequences. To fully evaluate the optimized decomposition method proposed in the soybean future price forecasting model, the BVMD-DELM was compared with the VMD-DELM in experiments (Tab.5). It is evident that compared with VMD, BVMD optimized by BWO algorithm can better decompose the original series to reduce the difficulty faced by the forecasting network. For example, in comparison with VMD-DELM, the MAE, MAPE and RMSE of BVMD-DELM in China are reduced by 54.1%, 54.3% and 56.0%, respectively. While the R2 value improved by 5.7%, demonstrating that the forecasting accuracy of the model will be significantly improved after optimization by the BWO algorithm.
Fig.9: Based on the high-frequency reconstructed series from the three markets, the randomness of the time series and the components included remain very high, which requires a secondary decomposition to continue the research. The data for the secondary decomposition of soybean future price in the three markets compared with the primary decomposition are shown in Tab.5. Comparing the prediction results of ICEEMDAN-LZC-BVMD-DELM for the three markets with ICEEMDAN-DELM and BVMD-DELM, respectively, using MAE as the error evaluation index, there was an improvement of 82.4% and 67.4% for China. This demonstrates that the secondary decomposition preprocessing based on ICEEMDAN-LZC-BVMD can provide better data identification and information extraction for the original time series compared with the conventional primary decomposition preprocessing and is a completely necessary data preprocessing method. The ICEEMDAN-LZC-BVMD-DELM shows an increase in MAE values of close to 90% in the three markets compared with the DELM. In the case of the original soybean futures price series with high complexity, containing considerable information and redundancy, which makes identification and prediction difficult, the ICEEMDAN-LZC-BVMD method can reduce the nonlinearity of the original data and can effectively improve the prediction accuracy. The ICEEMDAN and BVMD are used for secondary decomposition to obtain multiple subsequences, which can greatly reduce the difficulty of sequence identification, and the prediction effect of DELM is greatly improved.
3.3.3 Forecasting model optimization method enhancement analysis
To be able to thoroughly evaluate the performance of the SSA optimization algorithm, four comparison experiments were conducted for the three markets, with the specific parameters shown in the table below.
The results of the error assessment of the parametric optimization model are given in Tab.2. The percentage improvement for the parametric optimization method is given in Tab.6. The prediction results of soybean price series of the hybrid models for the three markets are shown in Fig.10–Fig.12. The prediction errors of the hybrid models for the three markets are shown in Fig.13–Fig.15. The scatter plot of hybrid models for three markets are shown in Fig.16–Fig.18.
Fig.10–Fig.18 and Tab.6 reveal that the forecasting results of the model with the parametric optimization method outperformed the model without the parametric optimization method in all cases. This demonstrates that the use of the parameter optimization algorithm results in a more robust data analysis capability. The SSA parameter optimization method proposed in this paper can analyze the characteristics of each subsequence, select the optimal parameters and obtain better prediction results, effectively improving the accuracy and generalization of the prediction model. Therefore, the SSA method has good potential for improving the accuracy of soybean future price forecasting systems.
3.3.4 Comparison of the evaluated models
To evaluate the effectiveness of the ICEEMDAN-LZC-BVMD-SSA-DELM, single network models (ELM, RBF, DBN, LSTM, GRU and DELM) and hybrid models (SSA-DELM, ICEEMDAN-DELM, VMD-DELM, BVMD-DELM, ICEEMDAN- SSA-DELM, BVMD-SSA-DELM and ICEEMDAN-LZC-BVMD-DELM) were compared with the proposed models. The evaluation of these models is shown in Fig.19–Fig.21.
All hybrid models gave better predictions than the single models. The hybrid models integrated multiple signal processing methods, which dramatically improved the performance of the models. As concluded above, the secondary decomposition method can effectively decompose highly complex time-series data into multiple subsequences with stable patterns. ICEEMDAN is an adaptive decomposition algorithm, but VMD requires empirical manual adjustment of the decomposition level K and penalty factor α, which increases the operational difficulty and poses the risk of experimental failure. Therefore, BVMD is introduced to decompose the reconstructed sequences of ICEEMDAN high-frequency subsequences. The secondary decomposition can boost the forecasting accuracy of the neural network by noise reduction and nonlinearity removal and also improve the generalization ability of the neural network by reducing the non-smoothness of the time-series data.
DELM was used to predict the soybean future price, and it performed the best of the six single models. However, the random generation of input layer weights and thresholds in DELM led to unstable predictions, negatively impacting prediction accuracy. To address this, the SSA algorithm was used for parameter optimization.
In a comparative analysis of the Chinese and the United States soybean future price data sets against the Italy soybean ETF price data set, it is found that the MAPE results of the single network models and hybrid models for the Italy data set were improved more substantively compared with the other two data sets. This finding indicates that the proposed model enhances overall forecasting performance and excels in specific markets (e.g., extremes or outliers), as illustrated in Fig.22, which shows the MAPE comparison of representative models across the markets. While the improvement in R2 was not statistically significant, as shown in Fig.23, this may be attributed to the already high R2 value of the Italy data set. Nonetheless, values close to 1 indicate that the proposed model effectively fits the data and the proposed model performs quite well on the Italy soybean ETF price data set.
Of the models compared, the proposed model achieved the highest prediction accuracy in the three markets and for all evaluation metrics. It fully illustrates that the ICEEMDAN-LZC-BVMD-SSA-DELM model can effectively predict soybean futures price series and has strong prospects for application and further research in the field of agricultural futures price forecasting.
4 Conclusions
To achieve high-precision forecasting of soybean future price series, an improved deep learning model for soybean future prices prediction with hybrid data preprocessing strategy is proposed.
The main conclusions from the analysis of the comparative experimental results are as follows. (1) Of the six single models compared, the DELM model performs the best. (2) The secondary decomposition method has greatly improved the forecasting effect of the proposed model. In contrast to the non-decomposition and primary decomposition methods, the subsequences obtained by the secondary decomposition proposed are more stable and orderly, which greatly reduces the difficulty of time-series identification and is an excellent data preprocessing method. (3) The BWO algorithm used in this study can significantly enhance the decomposition effect of VMD, and the SSA algorithm can effectively increase the prediction accuracy of DELM network. (4) The proposed hybrid model achieves sound experimental results for soybean futures closing price time series in the three markets evaluated. Notably, the proposed model also performs well in predicting the typical more flexible Italy soybean ETF price data set, demonstrating its strong potential for generalization.
The hybrid forecasting framework proposed in this paper provides new ideas for agricultural price forecasting research and also has clear potential for generalization and application to the forecasting of other financial time series.
Isengildina O, Irwin S H, Good D L. Evaluation of USDA interval forecasts of corn and soybean prices. American Journal of Agricultural Economics, 2004, 86(4): 990–1004
[2]
Yang Q H, Du X Q, Wang Z H, Meng Z C, Ma Z H, Zhang Q. A review of core agricultural robot technologies for crop productions. Computers and Electronics in Agriculture, 2023, 206(13): 107701
[3]
Kenyon D E. Producer ability to forecast harvest corn and soybean prices. Applied Economic Perspectives and Policy, 2001, 23(1): 151–162
[4]
Darekar A, Reddy A A. Predicting market price of soybean in major india studies through ARIMA model. Indian Journal of Pulses Research, 2017, 30(2): 73–76
[5]
Panasa V, Kumari R V, Ramakrishna G, Kaviraju S K. Maize price forecasting using auto regressive integrated moving average (ARIMA) model. International Journal of Current Microbiology and Applied Sciences, 2017, 6(8): 2887–2895
[6]
Bhardwaj S P, Paul R K, Singh D R, Singh K N. An empirical investigation of Arima and Garch models in agricultural price forecasting. Economic Affairs, 2014, 59(3): 415–428
[7]
Şahinli M A. Potato price forecasting with Holt-Winters and ARIMA methods: a case study. American Journal of Potato Research, 2020, 97(4): 336–346
[8]
Gao Z M, Khot L R, Naidu R A, Zhang Q. Early detection of grapevine leafroll disease in a red-berried wine grape cultivar using hyperspectral imaging. Computers and Electronics in Agriculture, 2020, 179: 105807
[9]
Yin H L, Jin D, Gu Y H, Park C J, Han S K, Yoo S J. STL-ATTLSTM: vegetable price forecasting using STL and attention mechanism-based LSTM. Agriculture, 2020, 10(12): 612
[10]
Craessaerts G, De Baerdemaeker J, Saeys W. Fault diagnostic systems for agricultural machinery. Biosystems Engineering, 2010, 106(1): 26–36
[11]
Mahto A K, Alam M A, Biswas R, Ahmed J, Alam S I. Short-term forecasting of agriculture commodities in context of indian market for sustainable agriculture by using the artificial neural network. Journal of Food Quality, 2021, 2021: e9939906
[12]
Jaiswal R, Jha G K, Kumar R R, Choudhary K. Deep long short-term memory based model for agricultural price forecasting. Neural Computing & Applications, 2022, 34(6): 4661–4676
[13]
Zong J J, Zhu Q Y. Price forecasting for agricultural products based on BP and RBF Neural Network. In: 2012 IEEE International Conference on Computer Science and Automation Engineering. IEEE, 2012, 607–610
[14]
Xu X J. Corn cash price forecasting. American Journal of Agricultural Economics, 2020, 102(4): 1297–1320
[15]
Xu X J, Zhang Y. Canola and soybean oil price forecasts via neural networks. Advances in Computational Intelligence, 2022, 2(5): 32
[16]
Liu H, Yin S, Chen C, Duan Z. Data multi-scale decomposition strategies for air pollution forecasting: a comprehensive review. Journal of Cleaner Production, 2020, 277: 124023
[17]
Liu H, Wu H P, Li Y F. Smart wind speed forecasting using EWT decomposition, GWO evolutionary optimization, RELM learning and IEWT reconstruction. Energy Conversion and Management, 2018, 161: 266–283
[18]
Yin S, Liu H, Duan Z. Hourly PM2.5 concentration multi-step forecasting method based on extreme learning machine, boosting algorithm and error correction model. Digital Signal Processing, 2021, 118: 103221
[19]
Zhang Y L, Na S G. A novel agricultural commodity price forecasting model based on fuzzy information granulation and MEA-SVM model. Mathematical Problems in Engineering, 2018, 2018: e2540681
[20]
Li G Q, Chen W, Li D H, Wang D J, Xu S W. Comparative study of short-term forecasting methods for soybean oil futures based on LSTM, SVR, ES and wavelet transformation. Journal of Physics: Conference Series, 2020, 1682(1): 012007
[21]
Wang J, Wang Z, Li X, Zhou H. Artificial bee colony-based combination approach to forecasting agricultural commodity prices. International Journal of Forecasting, 2022, 38(1): 21–34
[22]
Liang J Y, Jia G Z. China futures price forecasting based on online search and information transfer. Data Science and Management, 2022, 5(4): 187–198
[23]
Zhang D Q, Zang G M, Li J, Ma K P, Liu H. Prediction of soybean price in China using QR-RBF neural network model. Computers and Electronics in Agriculture, 2018, 154: 10–17
[24]
Liu H, Long Z H. An improved deep learning model for predicting stock market price time series. Digital Signal Processing, 2020, 102: 102741
[25]
Sun W, Huang C C. A novel carbon price prediction model combines the secondary decomposition algorithm and the long short-term memory network. Energy, 2020, 207: 118294
[26]
Zhu H M, Xu R, Deng H Y. A novel STL-based hybrid model for forecasting hog price in China. Computers and Electronics in Agriculture, 2022, 198: 107068
[27]
Liu H, Duan Z, Han F Z, Li Y F. Big multi-step wind speed forecasting model based on secondary decomposition, ensemble method and error correction algorithm. Energy Conversion and Management, 2018, 156: 525–541
[28]
Liu H, Zhang X Y. AQI time series prediction based on a hybrid data decomposition and echo state networks. Environmental Science and Pollution Research International, 2021, 28(37): 51160–51182
[29]
Colominas M A, Schlotthauer G, Torres M E. Improved complete ensemble EMD: a suitable tool for biomedical signal processing. Biomedical Signal Processing and Control, 2014, 14: 19–29
[30]
Aboy M, Hornero R, Abasolo D, Alvarez D. Interpretation of the Lempel-Ziv complexity measure in the context of biomedical signal analysis. IEEE Transactions on Biomedical Engineering, 2006, 53(11): 2282–2288
[31]
Kurtin D L, Scott G, Hebron H, Skeldon A C, Violante I R. Task-based differences in brain state dynamics and their relation to cognitive ability. NeuroImage, 2023, 271: 119945
[32]
Zhong C T, Li G, Meng Z. Beluga whale optimization: a novel nature-inspired metaheuristic algorithm. Knowledge-Based Systems, 2022, 251: 109215
[33]
Dragomiretskiy K, Zosso D. Variational mode decomposition. IEEE Transactions on Signal Processing, 2014, 62(3): 531–544
[34]
Liu Q, Zhao R Z, Yang B Z. Research of fault recognition method of rolling bearings based on K-VMD envelope entropy and SVM. Noise and Vibration Control, 2022, 42(3): 92
[35]
Li X H, Guo M M, Zhang R R, Chen G M. A data-driven prediction model for maximum pitting corrosion depth of subsea oil pipelines using SSA-LSTM approach. Ocean Engineering, 2022, 261: 112062
[36]
Mi X W, Liu H, Li Y F. Wind speed forecasting method using wavelet, extreme learning machine and outlier correction algorithm. Energy Conversion and Management, 2017, 151: 709–722
[37]
Tissera M D, McDonnell M D. Deep extreme learning machines: supervised autoencoding architecture for classification. Neurocomputing, 2016, 174(Part A): 42–49
[38]
Zhu B Z, Ye S X, Wang P, He K J, Zhang T, Wei Y M. A novel multiscale nonlinear ensemble leaning paradigm for carbon price forecasting. Energy Economics, 2018, 70: 143–157
RIGHTS & PERMISSIONS
The Author(s) 2024. Published by Higher Education Press. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0)
AI Summary 中Eng×
Note: Please be aware that the following content is generated by artificial intelligence. This website is not responsible for any consequences arising from the use of this content.