AN ENHANCED SHRINKAGE FUNCTION FOR DENOISING ECONOMIC TIME SERIES DATA USING WAVELET ANALYSIS
Sameera Abdulsalam Othman A, , Kurdistan M.Taher Omarb*
a Department of Mathematics, College of Basic Education, University of Dohuk, Kurdistan Region-Iraq
b Department of Mathematics, Faculty of Science, University of Zakho, Kurdistan Region, Iraq
Received: 2 Nov., 2023 / Accepted: 28 Jan., 2024 / Published: 30 Mar., 2024. https://doi.org/10.25271/sjuoz.2024.12.1.1223
ABSTRACT:
In the realm of economic (financial) time series analysis, accurate prediction holds paramount importance. However, these data often suffer from the presence of noise, particularly in highly random and non-stationary datasets like stock market data. Dealing with noisy data makes predicting noise-free economic models exceedingly challenging. This research paper introduces an innovative shrinkage (thresholding) function designed to improve the efficiency of wavelet shrinkage denoising in the context of financial time series data. The proposed function is constructed based on an arctangent model with adjustable parameters meticulously chosen to ensure the function maintains continuous differentiability. The application of this novel shrinkage function effectively reduces noise in stock data. Employing R program for data analysis and figure plotting, the performance of this approach is rigorously validated using closing price data from the Shanghai Composite Index, spanning the period from January 4, 2000 to August 28, 2023. The experimental results demonstrate that the proposed thresholding function outperforms classical shrinkage functions (hard, soft, and nonnegative garrote) in both continuous derivative property and denoising efficacy.
KEYWORDS: Wavelet Shrinkage Denoising, Arctangent Model, Financial Time Series Data, Noise Reduction, Stock Data, Shanghai Composite Index.
1. INTRODUCTION
Economic-mathematical modelling has established itself as an exceptionally potent approach for the characterization and examination of intricate socio-economic phenomena and operations. This methodology integrates mathematical models with innovative engineering solutions, thus ingraining it as an indispensable component of economics (Kolosinska and Kolosinskyi, 2013). Knowledge economics, as an abstract concept, necessitates translation into a tangible and concrete manifestation. This transformation is made achievable through the mathematical modelling of its processes as managerial entities (Svarc and Dabic, 2017).
Theoretical exploration and the formulation of mathematical models have yielded favourable results in addressing tangible economic and financial challenges. This includes tasks such as the analysis and prognosis of the flow of budgetary financial resources, the supervision of development, and operational support across varying budgetary tiers, as well as the evaluation of the quality and risk factors associated with the administration of economic systems (Neittaanmäki et al., 2016).
In the wake of the advent of market-oriented economies, global finance, and rapid technological progress, the demand for more sophisticated theoretical underpinning and precise analytical methodologies has surged. Time series models for finance and economics are now leveraging modern signal processing and mathematical tools. These tools encompass time-frequency analysis and the time-scale technique, notably employing the wavelet transform (Tiwari, 2016; Jiang, 2017).
The wavelet transform emerges as a particularly apt choice for scrutinizing economic data due to the nonstationary nature of most economic and financial time series, which exhibit evolving spectral compositions over time.
Prediction stands as a pivotal facet within economics and finance. Long-term prediction is instrumental in governing dynamic economic systems, while short-term prediction seeks to accurately track economic trends. In the examination of economic and financial data for forecasting, the wavelet technique has drawn a lot of interest. Its applications extend to forecasts in areas like crude oil prices, stock market trends, commodity market projections, and more (de Souza, 2010; Tiwari, 2014; Vacha, 2012). However, economic data records frequently contain noise originating from measurement inaccuracies and other sources. Noise within economic and financial systems contributes to heightened uncertainty within financial markets, intensifying the complexity of prediction tasks. To enhance prediction precision, denoising techniques are employed to alleviate data noise. This paper centers its focus on denoising procedures, and harnessing the wavelet transform to purify stock price data for predictive purposes.
Stock price volatility is subject to influence by diverse stochastic elements, thereby generating noisy observations. Through the utilization of wavelet shrinkage denoising and thresholding methods, this paper seeks to formulate a new shrinkage function designed to augment denoising efficacy. The overarching objective is to mitigate the repercussions of noise and elevate the accuracy of stock price predictions.
2. WAVELET TRANSFORM AND SHRINKAGE DENOISING
In the domain of processing and analyzing time series data in economics and finance, the ability to detect meaningful signals amidst noise and interference is paramount. To tackle this challenge, the wavelet transform (WT) has emerged as a valuable tool, leveraging its capacity to pinpoint transient features within signals. Essential properties like compression and sparsity are present in the WT, demonstrating that real-world signals often display sparsity—many smaller coefficients that can be disregarded, while a small number of significant coefficients carry the majority of the signal energy. With an output that rises when the input signal approaches the analysis template and falls when noise levels are higher, the WT essentially performs correlation analysis. This underlying idea supports the idea of wavelet denoising. Importantly, it's imperative to differentiate between wavelet denoising and smoothing, despite some authors interchangeably using these terms. Smoothing is concerned with eliminating high frequencies while preserving low ones, whereas denoising is centered on eliminating noise while retaining the signal, irrespective of its frequency content. Noise energy is distributed across all coefficients in the wavelet domain. Hence, denoising within the wavelet domain is accomplished by applying thresholding to WT coefficients, introducing the notion of wavelet shrinkage. This procedure entails selecting an appropriate threshold value and a thresholding rule to effectively attenuate noise power while retaining essential signal characteristics. Diverse wavelet thresholding strategies have been shown to exhibit near-optimal qualities from a minimax perspective in several research concentrating on signal denoising by wavelet shrinkage (Donoho, 1998).
The WT is widely used in non-stationary time series analysis to extract data in both the frequency and time domains. It can be viewed as a multi-scale variant of the specialized Fourier transform that dissects signals into smaller and altered forms of the original "mother" wavelet. The convolution of a time series, represented as x(t), and a wavelet function, represented as w(t), is the formal definition of the continuous wavelet transform (CWT) (Gomes,2015)
(1)
The translational parameter is denoted by d , the scale parameter is represented by a=
,
and the complex conjugate of ψ(t),
is denoted by ψ*, where "s" and "k" are integers. At
the coordinates (
,
) on the plane of time
scale, the CWT of x(t) yields a numerical result. This outcome displays the
relationship at that specific time-scale point between x(t) and ψ*(t). The
discrete counterpart of equation (1) is known as the discrete wavelet transform
(DWT):
(2)
The signal is broken down by DWT into components with different scales that correspond to successive frequencies. DWT plays a pivotal role in the multi-resolution approximation approach for analyzing signals across multiple scales or frequency bands.
In application, multi-resolution analysis starts with two-channel filter banks made up of low-pass and high-pass filters, and each filter bank is sampled half as fast as the preceding frequency (1/2 down-sampling). The length of the data determines how many rounds there will be in this decomposition. This down-sampling technique sustains a constant scaling parameter of 1/2 across successive wavelet transformations (Li and Kuo, 2008), thereby streamlining the computational implementation. DWT has seen extensive development and application in signal analysis across diverse domains.
By creating a suitable shrinkage (thresholding) function, DWT is used in this work to remove noise from stock price data. Risk Shrink, a useful spatially adaptable technique that trims empirical wavelet coefficients, is a part of groundbreaking research on wavelet shrinkage denoising (Donoho, 1994). Another well-known technique is the asymptotically minimax approach, which, when applied to curve estimation in the presence of noisy data, moves empirical wavelet coefficients toward the origin. Furthermore, Donoho and Johnstone (1995) describe Sure Shrink, as a technique for reducing noise by thresholding empirical wavelet coefficients. All these techniques are combined under the name "Wave Shrink," which is based on the idea of eliminating noise by setting wavelet coefficients to zero. A widely accepted technique for removing noise from distorted signals is called Wave Shrink.
Numerous scholars have employed wavelets in diverse applications in economics and finance, deploying DWT to dissect economic and financial data. This opened the door for wavelet analysis to be used in empirical economics and finance. The quasi-periodic components of signals are extracted using the Ensemble-Empirical-Model Decomposition (EEMD) and ensemble version of Empirical-Model-Decomposition (EMD) methods (Huang, 1998) in order to construct features. By using these nonparametrically produced components as inputs for classification models, human bias in feature generation can be minimized. Many fields have successfully used EEMD and DWT, particularly when dealing with non-stationary time series.
Non-stationary time series and highly oscillatory, connected to financial instruments like stock prices and market indices have been forecasted in a number of studies using EEMD and DWT.
Wavelet analysis is well known for its ability to examine the frequency content of analyzed processes with high joint time–frequency resolution and for estimating economic and financial time series using particular wavelet–based techniques. This method, which is based on an endogenously variable time window, allows for accurate identification of events that impact economic fluctuations, effective computational analysis, and the examination of economic relationships across various time horizons. In conclusion, wavelet analysis is crucial for economic forecasting even beyond its use in signal denoising. Conventional DWT has an advantageous decorrelation property, which suggests that it could be useful for financial and economic process forecasting. Surprisingly, the choice of analysis scale has a significant impact on the outcomes of forecasting methods used for financial and real economic variables. Wavelet transforms (CWT and DWT) simplify the analysis of simpler univariate and multivariate processes, which makes it easier to apply customized forecasting methods.
In the context of this analysis, we
introduce data (time series) denoted as , defined as follows:
(3)
represents normalized
time in this equation, f(·) is a deterministic function with potential
complexity and spatial inhomogeneity, The Gaussian noise standard deviation is
represented by the variable{
}, with
N(0,1). The objective is
to minimize the
risk when estimating
f(·) (Donoho, 1994; 1995):
(4)
The estimated sample value of f is denoted by
. The mean square error (MSE) is represented by the
risk. A well-liked tool
for estimating the function f(·) is WaveShrink. It achieves nearly optimal
minimax risk within a logarithmic factor of n across a broad range of
smoothness classes and loss functions, including
risk, and has broad
asymptotic near-optimality properties. The following steps are involved in the
WaveShrink method is done in three steps: a) Wavelet coefficient thresholding,
which involves shrinking empirical wavelet coefficients to zero, is done in three
steps: (b) forward wavelet transform of observed data, which turns data y into
the wavelet domain; (c) the reduced coefficients' inverse wavelet transform,
which returns the shrunken coefficients to the data domain.
The efficiency of shrinking empirical wavelet coefficients depends on the sparsity of the true coefficients of f(·).
This translates to practice as most coefficients (the non-important ones) have little effect on the functional form of f(), while a small number (the important ones) have a significant effect. A shrinkage function, sometimes referred to as a thresholding function, needs to have two characteristics to be accepted: a) It must retain large and significant coefficients and b) It must eliminate small and insignificant ones.
Hard shrinkage and soft shrinkage are the two basic shrinkage functions that Wave Shrink uses. Important coefficients are retained by the hard shrinkage function, but non-important coefficients are set to zero if their absolute values fall below a predetermined threshold ζ:
(5)
Here, ζ denotes the threshold value, with ζ ∈ [0, ∞).
Expanding upon the hard shrinkage function is the soft shrinkage function. Non-essential coefficients are set to zero and crucial coefficients that are not zero are reduced to zero if their absolute values fall below the threshold.:
(6)
There are advantages and disadvantages to both soft and hard shrinking. Large coefficients' shrinkage may cause a larger bias since the soft shrinkage is continuous but its derivative is discontinuous. Hard thresholding estimates, on the other hand, may have a higher variance and be unstable due to discontinuities in the shrinkage function (Gao 1998). The Wave Shrink denoising technique utilizes Breiman's (1995) non-negative garrote shrinkage function to solve the drawbacks of hard and soft shrinkages. The following is the definition of the non-negative garrote shrinkage function:
(7)
Several conclusions can be drawn
from the three shrinkage functions discussed: At the threshold , the hard functional
shrinkage shows a discontinuity. The soft shrinkage function maintains
consistency across its entire range. Since it is continuous, the non-negative
garrote shrinkage function is better than hard shrinkage. Moreover, the
positive garrote shrinkage function reaches the identity line as |x| increases
(just like hard shrinkage does), resulting in a lower bias for high
coefficients than soft shrinkage. These findings suggest that a good compromise
between hard and soft shrinkage functions is the non-negative garrote shrinkage
function.
It is worth noting, however, that
all three functions share the property of having a single threshold. (Gao,
1997) proposed a more general shrinkage function known as the firm (or
semisoft) shrinkage function, which includes two thresholds ( and
):Top of Form
(8)
![]() |
The waveforms of the hard (ζ =
3.41), soft (ζ = 3.41), positive garrote (ζ = 3.41), and semisoft ( = 2.341,
= 7.264) shrinkage
functions are shown in Figure 1. Although thresholding functions with improved
performance, such as the non-negative garrote and semisoft, which strike a
balance between hard and soft functions and have advantages over both, can
occasionally lack higher-order differentiability and have fixed structures that
limit their adaptability and flexibility. As a result, alternative thresholding
functions with continuous higher-order derivatives must be investigated.
3. DEVELOPMENT OF NOVEL NONLINEAR THRESHOLDING FUNCTIONS
To streamline the process of threshold selection and alleviate computational complexity, we aim to devise a novel shrinkage function that relies on a single threshold parameter, denoted as ζ. Simultaneously, this designed t thresholding function should endeavor to address the limitations observed in both the hard and soft shrinkage functions to the greatest extent possible.
Figure 1: Waveform Characteristics of the Shrinkage Functions.
The development of the novel thresholding function adheres to foundational principles. To address shortcomings in existing shrinkage functions, the proposed function prioritizes continuous derivatives, ensuring adaptability for gradient-based algorithms. Additionally, it maintains wavelet coefficients unchanged when their absolute values exceed the threshold, approximating a linear function (y=x). Crucially, the function zeroes coefficients below a given threshold (ζ), akin to y=0. These principles aim to minimize estimation variance and bias, overcoming limitations in conventional hard and soft shrinkage methods. The desired properties include yielding y=0 for coefficients within [-ζ, ζ], and y=x for coefficients beyond this range. These specifications guide the search for a function aligning with these criteria, ensuring the theoretical efficacy of the proposed shrinkage function.
The arctangent function is widely
utilized for improved approximation when approximating a function is close to
the origin. The arctangent function is the inverse of the tangent function,
denoted as . It has the domain of
real numbers and the range of
. The horizontal
asymptotes of the arctangent function are
and
. It is a one-to-one
relationship. Figure 2 depicts the arctangent function's waveform.
Figure 2: The Arctangent Function Waveform
When the x value is very small, the
arctangent function should be passes through the origin, i.e., , as shown in Figure 2.
When the x value moves away from
the origin, increases. These traits
match the intended criteria of the thresholding function. As a result,
arctan(x) can be used as the starting point for constructing the new shrinkage
function with continuous derivatives. To rase the adaptability and functionality
of the suggested shrinkage (thresholding) function, three shape-tuning
parameters are included:
(9)
In this context, determining real numbers α, β, and k is essential. A nonzero positive integer for k ensures the thresholding function's odd symmetry. To accommodate a diverse signal range, the suggested shrinkage function is not only continuous but possesses higher-order derivatives. Examining differentiability, the continuity and equality of left and right derivatives at x=ζ are both necessary and sufficient conditions for the function's differentiability concerning the threshold variable.
3.1 The Proposed Shrinkage Function's Partial Derivative
The is the partial
derivative of the presented shrinkage function.
(10)
This derivative ensures continuity and is necessary for the proposed shrinkage function to be smooth. The parameter β can be calculated as follows:
β = (11)
3.2 The Proposed Shrinkage Function's Second-Order Partial Derivative
![]() |
(12)
The parameter k can be determined as:
(13)
3.3 The Final Shrinkage Function:
Taking into consideration the above equations and parameters, the differentiable shrinkage function can be expressed as:
(14)
3.4 Parameter α
Optimization:
Prior to putting the shrinking function into use, the value of the parameter must be predetermined. The formula (14) can be simplified to determine the optimal value of
: (15)
Upon analyzing the formula (15), two critical results emerge:
(a) When , the shrinkage function should closely resemble hard thresholding,
making the value of α as small as possible. Notably, when α=1,
the value
reaches a minimum
(specifically,
=0.57). This implies that
the priested shrinkage function is similar to non-negative garrote shrinkage
for significant wavelet coefficients.
(b) When , The non-important
wavelet coefficients should be set to zero by the shrinkage function. in order
to reduce noise. α should be as large as possible in this case.
To reconcile these contradictory results,
the optimal α value should take into account both conditions
thoroughly. Furthermore, because it balances estimate variance and bias, the risk, or (MSE) will be
used as a criterion for determining the optimal α value. The
optimal value and the formulas for bias, variance, and
risk of the suggested
shrinkage estimate for a normal random variable are discussed in the section
that follows.
RISK ANALYSIS AND
DETERMINATION OF OPTIMAL THRESHOLDS
For convenience, we represent the
proposed shrinkage function as Let X follow a Gaussian
distribution, X~N(θ, 1). We determine the mean, variance, and
risk functions for the
shrinkage estimate of θ under the shrinkage function
(·) and threshold
, as follows (Gao 1998):
Mean:
,Variance:
Risk:
((16)
Here, represents the square
bias. We can further simplify Equation (16) to:
− 2θ ·
According to Equation (16), the bias
component contributes more to the than the variance
component. As a result, minimising bias by selecting an optimal α value
reduces the
risk function. The
risk function can
nevertheless accomplish an ideal trade-off between bias and variance even when
the variance is not at its lowest in this situation. To find the parameter
value that minimizes the
risk, the
risk functions for
various values of can be computed and compared. For example, we could take = 1,
2, 3, 5, 7 and set the threshold to = 3.41. Figure 3 depicts the resulting
waveforms of the
risk functions for the
presented shrinkage function.
Figure 3: shows the risk profiles of the
presented shrinkage function at variate α values.
Figure 4: depicts the risk functions for the
hard, soft, and proposed functions
In Figure 4. we have included for hard shrinkage (
= 3.41), soft shrinkage
(
= 3.41) have been
included, and the study presented shrinkage functions (
= 4.31) to facilitate
comparison.
This implies that the proposed
function with α = 1 offers an advantage in the VisuShrink method, which
employs the universal threshold. According to the VisuShrink method, the
threshold is given by:
(17)
![]() |
(18)
In this case, denotes the noisy
wavelet coefficients, and
denotes the number of
noisy data points. In addition to the VisuShrink method, we consider the
SureShrink procedure, which uses empirical wavelet coefficient thresholding to
suppress noise (Donoho,1995). The adaptive thresholding aims to minimise the Stein
unbiased estimate of risk (Sure) for threshold estimates. This execution's
computational effort is proportional to N·log(N), where N is the sample size.
5. APPLICATION METHODOLOGY
We a dataset comprising closing prices of the Shanghai Composite Index (SCI) was utilized and sourced from the Shanghai Stock Exchange (SSE) for our numerical analysis. This dataset encompasses a total of 5789 trading days, equivalent to approximately 283 trading months, spanning from Jan. 4, 2000, to Aug. 28, 2023. The raw data series illustrating the closing prices for each trading day are presented in Figure 5.
Figure 5: Plots Depicting the Original Daily Stock Price Data Series.
Figure 5 visually demonstrates that the observed data exhibit significant noise contamination (Wang et al., 2011). These noise sources could potentially include classical sources of noise as well as noise unique to the financial system. To mitigate the impact of these noises, - - various wavelet shrinkage (thresholding) functions were employed
To evaluate the performance of denoising techniques on actual stock price data, we can express the data as follows (Alrumaih,2022):
(19)
Here, s represents the underlying clean signal, and n denotes the additive noise. As previously discussed, various factors can contribute to the noise, including market fluctuations due to economic or political events, investor expectations, and psychological factors. To comprehensively account for these factors, we consider the noise n to behave like white noise. two metrics to assess denoising performance concerning actual stock prices are used: the Signal-to-Noise Ratio (SNR) and the Root Mean Square Error (RMSE), defined as follows (Jarrah,2023):
(20)
(21)
In these equations, denotes the denoised
signal, and N denotes the number of samples in the signal s. Better
denoising performance is indicated by a higher SNR value and a lower RMSE
value. While, in practice, it is complex to distinguish the clean signal s
and the noise n from the observed signal y. As a result, in this paper, we will
estimate the performance metrics SNR and RMSE using actual data y rather than
clean data s, albeit with some inherent estimation error.
a series of experiments to identify the most suitable wavelet type for analyzing the SCI data was demeanoured, which comprises closing prices. To facilitate a meaningful comparison, we both the hard thresholding function and the minimax threshold rule to denoise a subset of the SCI data were employed, consisting of approximately 5000 closing price records. The wavelet decomposition was carried out at level 6, utilizing a range of wavelet types denoted as "db1," "db2," and so forth up to "db12." The outcomes of these experiments are presented in Table 1 for reference.
Table 1: Comparative Analysis of Various Wavelet Types for the Subset of Daily Data
|
db1 |
db2 |
db3 |
db4 |
db5 |
db6 |
db7 |
db8 |
db9 |
db10 |
db11 |
db12 |
SNR |
26.607 |
23.54 |
23.64 |
22.85 |
22.77 |
22.73 |
22.71 |
22.69 |
22.68 |
22.66 |
22.66 |
22.65 |
RMSE |
13.429 |
12.973 |
12.895 |
13.382 |
11.479 |
13.526 |
13.548 |
13.568 |
13.584 |
13.59 |
13.59 |
13.60 |
Table 2: Evaluating Different Decomposition Levels for Daily Data Subset
Level |
1 |
2 |
3 |
4 |
5 |
6 |
7 |
8 |
Period |
2–4 |
4–8 |
8–16 |
16–32 |
32–64 |
64–128 |
128–256 |
256–512 |
SNR |
25.5 |
19.78 |
19.17 |
19.02 |
19 |
22.90 |
19.10 |
19.20 |
RMSE |
14.7 |
14.52 |
15.60 |
15.81 |
15.86 |
9.87 |
15.86 |
15.86 |
Table 1 reveals that
"db5" surpasses other wavelets with slightly higher SNR and lower
RMSE, highlighting its effectiveness for denoising SCI data. Wavelet analysis's
outcome depends heavily on the selected decomposition level. Through experimentation,
we identify six levels as optimal for denoising SCI data, providing the highest
SNR and lowest RMSE in Table 2. In subsequent analyses, A six-level analysis
and "db5" are employed, estimating a noise variance of approximately ≈ 17.32 (σ ≈ 4.16) using Eq. (18). To assess denoising efficacy, we employed various
shrinkage techniques and thresholding rules to eliminate noise from the
original data, considering noise levels of
≈ 17.32 (σ
≈ 4.16). Initially,
we focused on the universal threshold rule, conducting experiments to calculate and register the SNR and RMSE values for various thresholding functions. These outcomes are presented in Table 3.
Table 3 shows the results of the universal threshold rule experiment, and the hard shrinkage function performs the best in terms of denoising performance, as indicated by SNR and RMSE metrics. The soft shrinkage function, on the other hand, performs the worst. Furthermore, the proposed thresholding function outperforms the soft and nonnegative garrote shrinkage methods in denoising, though it lags slightly behind the hard shrinkage function. In conclusion, the presented shrinkage function's denoising performance is very close to the best-performing method.
Table 3: Evaluation of the Universal Threshold Rule for Daily Data.
|
Hard |
Soft |
Non-negative garrote |
Proposed |
SNR |
35.3289 |
31.2558 |
33.0593 |
34. 9226 |
RMSE |
7. 467 |
11. 3490 |
10. 3461 |
9.5070 |
CONCLUSION
In the realm of predicting economic indicators from noisy time series data, neglecting denoising processes hampers prediction accuracy. Noisy data are commonly denoised using wavelet shrinkage, yet traditional methods like soft shrinkage fall short with financial data. We aimed to enhance denoising by introducing a continuous derivative-based thresholding function for an adaptive gradient-based algorithm. Experimentation revealed "db5" as the superior wavelet for denoising SCI data, and a six-level decomposition yielded optimal results. Assessing denoising performance, hard shrinkage excelled, with our proposed function also outperforming soft and non-negative garrote methods. Denoising is vital for accurate economic predictions, and our proposed function shows promise for future applications.
REFERENCES
Alrumaih, R.M. and Al-Fawzan, M.A., 2002. Time series forecasting using wavelet denoising an application to saudi stock index. Journal of King Saud University-Engineering Sciences, 14(2), pp.221-233.
Breiman, L. (1995). Better subset regression using the nonnegative garrote. Technometrics, 37, 373–384.
de Souza e Silva, E. G., Legey, L. F. L., & de Souza e Silva, E. A. (2010). Forecasting oil price trends using wavelets and hidden Markov models. Energy Economics, 32, 1507–1519.
Donoho, D. L., & Johnstone, I. M. (1994). Ideal spatial adaptation via wavelet shrinkage. Biometrika, 81(3), 425–455.
Donoho, D. L., & Johnstone, I. M. (1995). Adapting to unknown smoothness via wavelet shrinkage. Journal of the American Statistical Association, 90(432), 1200–1224.
Donoho, D. L., & Johnstone, I. M. (1998). Minimax estimation via wavelet shrinkage. Annals of Statistics, 26(3), 879–921.
Gao, H. Y. (1998). Wavelet shrinkage denoising using the non-negative garrote. Journal of Computational and Graphical Statistics, 7(4), 469–488.
Gao, H. Y., & Bruce, A. G. (1997). WaveShrink with firm shrinkage. Statistica Sinica, 7, 855–874.
Gomes, J., & Velho, L. (2015). From Fourier analysis to wavelets. Basel, Switzerland: Springer.
Huang, N. E., Shen, Z., Long, S. R., Wu, M. C., Shih, H. H., Zheng, Q., et al. (1998). The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis. In Proceedings of the royal society of London a: Mathematical, physical and engineering sciences (Vol. 454, pp. 903–995). The Royal Society.
Jarrah, M. and Derbali, M., 2023. Predicting Saudi Stock Market Index by Using Multivariate Time Series Based on Deep Learning. Applied Sciences, 13(14), p.8356.
Jiang, M. H., An, H. Z., Jia, X. L., & Sun, X. Q. (2017). The influence of global benchmark oil prices on the regional oil spot market in multi-period evolution. Energy, 118, 742–752.
Kolosinska, M. I., & Kolosinskyi, Y. Y. (2013). Analysis and forecast of the basic principles of tourist market development in Ukraine using the methods of economic-mathematical modeling. Actual Problems of Economics, 10(148), 222–227.
Li, S. T., & Kuo, S. C. (2008). Knowledge discovery in financial investment for forecasting and trading strategy through wavelet-based SOM networks. Expert Systems with Applications, 34, 935–951.
Neittaanmäki, P., Repin, S., & Tuovinen, T. (2016). Mathematical Modeling and optimization of complex structures. Basel, Switzerland: Springer.
Svarc, J., & Dabic, M. (2017). Evolution of the knowledge economy: A historical perspective with an application to the case of Europe. Journal of the Knowledge Economy, 8(1), 159–176.
Tiwari, A. K. (2013). Oil prices and the macroeconomy reconsideration for Germany: Using continuous wavelet. Economic Modelling, 30, 636–642.
Tiwari, A. K., & Kyophilavong, P. (2014). New evidence from the random walk hypothesis for BRICS stock indices: A wavelet unit root test approach. Economic Modelling, 43, 38–41.
Tiwari, A. K., Mutascu, M. I., & Albulescu, C. T. (2016). Continuous wavelet transform and rolling correlation of European stock markets. International Review of Economics and Finance, 42, 237–256.
Vacha, L., & Barunik, J. (2012). Co-movement of energy commodities revisited: Evidence from wavelet coherence analysis. Energy Economics, 34, 241–247.
Wu, Z. H., & Huang, N. E. (2009). Ensemble empirical model decomposition: A noise-assisted data analysis method. Advances in Adaptive Data Analysis, 1(01), 1–14.