Next Article in Journal
Regularization, Bayesian Inference, and Machine Learning Methods for Inverse Problems
Next Article in Special Issue
Design and Test of an Integrated Random Number Generator with All-Digital Entropy Source
Previous Article in Journal
Ground State, Magnetization Process and Bipartite Quantum Entanglement of a Spin-1/2 Ising–Heisenberg Model on Planar Lattices of Interconnected Trigonal Bipyramids
Previous Article in Special Issue
A Multi-Method Survey on the Use of Sentiment Analysis in Multivariate Financial Time Series Forecasting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Combining Measures of Signal Complexity and Machine Learning for Time Series Analyis: A Review

by
Sebastian Raubitzek
* and
Thomas Neubauer
Information and Software Engineering Group, Institute of Information Systems Engineering, Faculty of Informatics, TU Wien, Favoritenstrasse 9-11/194, 1040 Vienna, Austria
*
Author to whom correspondence should be addressed.
Submission received: 8 November 2021 / Revised: 6 December 2021 / Accepted: 10 December 2021 / Published: 13 December 2021
(This article belongs to the Special Issue Review Papers for Entropy)

Abstract

:
Measures of signal complexity, such as the Hurst exponent, the fractal dimension, and the Spectrum of Lyapunov exponents, are used in time series analysis to give estimates on persistency, anti-persistency, fluctuations and predictability of the data under study. They have proven beneficial when doing time series prediction using machine and deep learning and tell what features may be relevant for predicting time-series and establishing complexity features. Further, the performance of machine learning approaches can be improved, taking into account the complexity of the data under study, e.g., adapting the employed algorithm to the inherent long-term memory of the data. In this article, we provide a review of complexity and entropy measures in combination with machine learning approaches. We give a comprehensive review of relevant publications, suggesting the use of fractal or complexity-measure concepts to improve existing machine or deep learning approaches. Additionally, we evaluate applications of these concepts and examine if they can be helpful in predicting and analyzing time series using machine and deep learning. Finally, we give a list of a total of six ways to combine machine learning and measures of signal complexity as found in the literature.

1. Introduction

The rise of artificial intelligence resulted in increased use of machine and deep learning for analysis and predictions based on historical data instead of employing mechanistic expert models. The outcomes of these predictions are encouraging, e.g., in solid-state physics, first-principle calculations can be sped up [1], or solar radiation can be predicted using machine learning methods [2]. In biology, machine learning can be applied, e.g., to genomics, proteomics, and evolution [3]. In medicine, researchers used machine learning to improve diagnosis using collected information from the past [4]. In finance, the applications range from risk management or the construction of portfolios to designing and pricing securities [5]. In agriculture, machine learning can be used to predict yields and give estimates on the nitrogen status [6].
However, while these approaches provide promising results in some research areas, the results are inferior in some others, depending on the available data and the system’s complexity. Though it can not be generalized what sort of data is necessary to perform machine learning successfully, the two main reasons, assuming one has optimized the algorithm for the regarded task, for machine learning to fail are:
(i)
Lack of data, meaning that the overall amount of data is not enough to train a model properly.
(ii)
Lack of good data, meaning that despite there is a sufficient amount of data available, the inherent information is not enough to achieve good results.
One way to improve machine learning approaches suffering from these problems is by studying complex systems, i.e., chaos theory. The reason for this is that real-life systems are highly complex and non-linear systems. Therefore, complexity analysis and ideas from chaos theory should be considered when analyzing or predicting real-life systems. Therefore, data-driven analysis and prediction methods can be improved, taking into account the complexity and non-linear behavior of the system under study. e.g., in the case of a general lack of data one can employ interpolation techniques such as fractal interpolation [7], or stochastic interpolation approaches [8], to increase the overall amount of data. The advantage of these interpolation techniques, in contrast to, e.g., linear or polynomial interpolation, is that the interpolated data can be tailored to match the complexity of the original data.
The study of complex systems had its peak in the 20th century, and many techniques have been developed to characterize and analyze chaotic systems, i.e., complexity measures. Furthermore, chaos theory has a widespread spectrum of applications ranging from medicine [9], finance [10] to physics [11], to name a few applications. Though chaos theory is not that popular among computer scientists, there is significant potential for future applications. For example, data-driven analysis/prediction approaches in agriculture could greatly benefit from considering the complexity of historical data. In agricultural systems, turbulence and complex non-linear behavior can be observed for many phenomenons affecting the overall system [12], such as weather [13], irrigation [14] or environmental pollution [15].
Though there are numerous other applications of measures of signal complexity, only a few publications exist which combine artificial intelligence and measures of signal complexity for time series analysis. When referring to complexity measures or measures of signal complexity, we mean all sorts of entropy, information, order-, disorder-measures one can perform on time series data. This article is a survey of those applications.
For this purpose, we searched google scholar for combinations of words synonymous with machine and deep learning applications. Those words, together with words synonymous for the study of complex systems, yielded the list of publications reviewed in Section 3. The result is a list of 18 publications that combine machine learning and complexity measures for real-life time series analysis in one way or another. E.g., in [16], a hybrid approach of fuzzy logic, fractal dimension, and a neural network are used to make predictions of time series. In [17], several indices of emerging economics are examined using the Hurst exponent and the fractal dimension, thus indicating a long term memory in the time series data, and afterward are forecast using machine learning methods.
The reason for this relatively low number of 18 combined applications of machine learning and chaos theory for time series analysis is most likely to be found in the education of computer scientists. Most modern computer scientists are trained in math and physics, but the study of complex systems (chaos theory) is, in most cases, not part of the curriculum. Therefore this survey aims to get computer scientists working with artificial intelligence in touch with ideas from the study of complex systems to boost their research. Consequently, we want to find out what combinations have already been successfully implemented in combination with machine learning, where one approach can benefit from the other, and how the combined approaches can be categorized. We focus on the discussed complexity measures and only briefly discuss the corresponding machine learning algorithms.
Finally, we highlight possible contributions of a combined approach:
  • One may save computational resources due to a clever feature selection based on a data’s inherent complexity.
  • Machine learning results can be checked not only on their accuracy but also on the algorithms capability to reconstruct the characteristics of the original data, thus providing a more profound check for generalization of a trained model.
  • Additional complexity-based features can improve the accuracy of the model.
  • The complexity of the data under study can be used to alter and improve the architecture of the employed algorithm, e.g., adapt the number of input nodes of a neural network to the long-term memory of the data.
  • Complexity measures provide additional arguments for analysis, e.g., why is an algorithm not capable of long-term predictions on a specific data set.
  • Complexity measures provide a tool to identify regions of predictability and/or non-predictability within a data set. E.g., given financial data, the inherent complexity may be used to identify volatile non-predictable periods.
  • Complexity measures can be used as a filter for predictions to improve machine learning ensemble approaches [18].
This article is structured as follows:
In Section 2, we explain the employed approach to search and find relevant publications. Further, we explain how we extracted and concluded the information from all listed publications. Section 3 gives the list of all publications that were used for this research and briefly summarizes the approach taken. We provide all complexity measures used in the featured publications in Section 4 and refer to where they were used. Section 5 lists all machine learning approaches and links them to the publications and the used complexity measures. We finally conclude our findings in Section 6 and sort all publications into six types of approaches for combining machine learning and complexity measures for time series data. Further, we give some ideas on future combinations and applications of machine learning and complexity measures. We also added a table summing up our findings, which can be found in Appendix B.

2. Methodology

For this review we performed an online search on google scholar. We searched for the following key words:
The following two lists give the range of search strings which we combined to yield the list of publications from Section 3.
Search strings to indicate machine learning/artificial intelligence:
  • Machine Learning
  • Artificial Intelligence
  • Neural Network
  • RNN or Recurrent Neural Network
  • LSTM or Long Short Term Memory
  • ANN or Artifical Neural Network
  • SVM or Support Vector Machines
Search strings to indicate methods from chaos theory:
  • Chaos
  • Hurst Exponent
  • Chaotic
  • Fractal
  • Fractal Dimension
  • Lyapunov Exponent
  • Entropy
Further, we added the words time series, forecast, and prediction to target time series data applications specifically.

Exclusion Criteria

The previously mentioned search strings (or the combined search strings) yielded many publications which had to be filtered. Thus, we used the following criteria to guarantee a focused review process. We excluded all publications that:
  • Did not specifically deal with time-series data.
    This is necessary as most measures of signal complexity, e.g., the Hurst exponent, the spectrum of Lyapunov exponents, etc., deal with time-series data only.
  • Mentioned, but did not employ whether machine learning techniques or ideas from chaos theory and/or complexity measures. i.e., no application was shown.
  • Deal only with theoretical implications of a combination of the mentioned techniques, i.e., without an actual application and results.
  • Included time series classification tasks, e.g., a time series data was classified as a certain behavior.
  • Aimed to calculate the complexity of a signal using a machine learning or neural network approach.
  • Analyzed only model data or artificial data.
  • Included phase space reconstruction approaches, such as methods to find a time series’s embedding dimension or time delay. (Note that for some complexity measures it is necessary to find a suitable phase space embedding, thus an embedding dimension and a suitable time delay. However, as we focus on the complexity measures, we did not include articles that only did a phase space reconstruction.)
Section 3 shows the remaining results that we chose for this review.
We next analyzed all used complexity measures and machine learning algorithms. We briefly explain the basic idea, and in terms of the complexity measures, describe how they are used in combination with machine learning approaches. Here our focus is on explaining the employed complexity measures and not the machine learning approaches. Still, we briefly discuss the employed machine learning approaches but refer to the mentioned sources for each algorithm for an in-depth treatment of the subject. Thus, we sort all featured publications by the employed complexity measures, their machine learning approaches, and lastly, according to their approach on combining complexity measures and machine learning. Finally, we discuss some future ideas.

3. Relevant Publications

There are several applications of complexity measures to machine learning. Here we give an overview of applications, results and discuss the applied techniques. For this reason 18 publications, combining machine learning and chaos theory are listed.
  • Ref [19]: A back-propagation neural network is used to capture the relationship between technical indicators and of the index in the market under study over time. Using the neural network, one can achieve better results than conventional ARIMA(Auto-Regressive Integrated Moving Average) models. R/S analysis is used to identify if Kuala Lumpur Composite Index (KLCI) is a random process or if it features long-term memory. A Hurst exponent indicates a long memory effect. Afterward, the time series is scrambled, and a new Hurst exponent calculated, which is significantly lower. This supports the thesis for KLCI to have a long-term memory.
  • Ref [16]: This research analyzes U.S./Mexico exchange rates for their predictability using different methodologies. The time series is clustered into several pieces. For each piece, the fractal dimension is calculated. Then fuzzy membership functions for fractal dimension, position, and the geometry of the sub-series are defined. Based on this, predictions are made using fractal-enhanced fuzzy logic and compared to neural network predictions. Here neural networks seem to work better on short-term predictions, but the fuzzy logic approach performs better in the long term.
  • Ref [20]: A layered back-propagation artificial neural network model with feedback connection was used for the study of the solar wind magnetosphere coupling and prediction of the geomagnetic D s t index. The local Hölder exponent, α , was used to analyze local scaling and singularity properties. The multifractality of the observed fluctuations led to α changing from point to point. The local Hölder exponent was added to the input layer of the neural network resulting in superior performance compared to an approach without the local Hölder exponent.
  • Ref [21]: The Nikkei stock prices over 1500 days are analyzed and predicted using fractal analysis, i.e., Hurst exponent, fractal dimension, and auto-correlation coefficient. The fractal dimension and the Hurst exponent show that the discussed time series follows a persistent behavior and can, therefore, be forecast theoretically. The fractal analysis also showed that data for three days provided the strongest correlation and, therefore, declared 3 and 5 input nodes. The results showed that the error is significantly lower for three input nodes than for five, and the training could be done using fewer training epochs.
  • Ref [22]: Here, it is stated that time series with a larger Hurst exponent can be predicted more accurately by back-propagation Neural Networks than those with H values close to 0.5 . The Hurst exponent for all 1024 trading day periods of the Dow-Jones index from 2 January 1930, to 14 May 2004, is calculated and analyzed. Periods with large Hurst exponent can be forecast with higher accuracy than those with low accuracy. The results suggest that the Hurst exponent can be used to guide data selection before forecasting. Furthermore, the embedding dimension of the time series under study and the corresponding time delay calculated via minimal mutual information were used to set the input nodes for the neural networks, i.e., The embedding dimension defined the number of input nodes, and the time delay defined the separation between the time steps fed into the neural network. This idea is based on Taken’s theorem for phase space reconstruction.
  • Ref [23]: Stock market closing prices were predicted using an enhanced evolutionary artificial neural network model (EANN). R/S analysis was used to calculate the Hurst exponent for different scales in the range of 3 , 50 for each time series. The reason for doing this for different sizes of subsets is that one identifies the length of maximal persistency, i.e., the size of subsets where the averaged Hurst exponent is maximal. The experiment shows that when finding the scale with the maximal Hurst exponent, deviations from that scale shift H towards 0.5 , meaning that information on persistency in the time series diminishes when observing other scales. Though those Hurst-exponent-based models did not outperform the regular EANN models, when considering trading strategies using the Hurst-exponent-based models, the average returns are higher than for other models. One interpretation is that the Hurst-based models capture essential information about upward/downward trends in the time series.
  • Ref [24]: According to the efficient market hypothesis, stock prices follow a random walk pattern and can not be predicted with high accuracy. In this research, the Dow Jones Industrial Average is analyzed for its predictability and complexity. Using the Hurst exponent, one can identify if a time series is a random walk process or features long-term memory. Here the Hurst exponent is used to identify regions in a time series that have non-random characteristics, i.e., H 0 , 5 . In these regions, predictions are made using artificial neural networks, decision trees, and k-nearest neighbor approaches. Using these techniques and ensembles, accuracy up to 65 % was achieved. Further, the embedding dimension and the time delay of the data under study were calculated to alter the architecture of the algorithm, i.e., adapt the input to the characteristics of the supposed underlying complex system.
  • Ref [25]: Many different chaotic time-series predictions are made using a NARX (Nonlinear Autoregressive model process with eXogenous input) dynamic recurrent neural network. The analyzed time series are Chaotic Mackey-Glass time series, Fractal Weierstrass time series, and BET time series (average of daily closed prices for nine representatives, most liquid listed companies at the Bucharest Stock Market). The Hurst exponent with H 0 , 5 suggests that all three time-series are predictable. For the first two, NARX neural networks perform very well, but for the third, despite having a high Hurst exponent, it worked not as well as for the others. The reason for this is that the third is the only real-life time series featured in the selection.
  • Ref [26]: The fractal dimension of data sets is used to select features to make predictions with. Afterward, Support Vector Machines (SVM) are used to make predictions. A total of 19 features is used to forecast The Shanghai Stock Exchange Composite Index (SSECI). Using fractal feature selection, this number is reduced to distinguish between "noise" and relevant data to predict future behavior. The results show that higher accuracy is achieved when fractal feature selection is applied. Further, fractal feature selection outperforms five other feature selection methods.
  • Ref [27]: Different AI methodologies, i.e., random forest, neural networks, and Fuzzy Inductive Reasoning (FIR), are proposed to perform short-term electric load forecasting (24h). A feature selection process based on Shannon entropy is used to reduce the number of features fed into the learning algorithms. FIR outperformed the random forest and neural network approaches.
  • Ref [28]: Several stock indices of developed and emerging markets are examined if they sport a long-term memory or follow a more random behavior. Rescaled range analysis is performed to calculate the Hurst exponent and the fractal dimension of those markets. The fractal analysis shows that those markets show persistence and can, therefore, theoretically be forecast. Afterward, machine/deep learning methods, i.e., adaptive neuro-fuzzy inference system, dynamic evolving neuro-fuzzy inference system, Jordan neural network, support vector regression, and random forest, are used to forecast further market development. The results show that these data sets can effectively be forecast, and random forest performed best in this study.
  • Ref [17]: A framework for forecasting exchange rates is presented. Different exchange rates are investigated using R/S analysis, the Hurst exponent, and the fractal dimension of time series, thus showing persistent behavior in the considered time series data. Afterward, the time series data are decomposed into wavelets and predicted using Random Forest and a bagging-based decision tree algorithm. The results show that both approaches performed well for the task at hand.
  • Ref [29]: The scope of this paper is to verify the existence of a relationship between long-term memory in fractal time series and the accuracy of neural networks to forecast them. Brazilian financial assets traded at BM&FBovespa, specifically public companies shares and real estate investment funds, were considered, and their long-term memory was estimated using R/S analysis and the Hurst exponent. Time lagged feedforward neural networks with back-propagation supervised learning process and gradient descent for error minimization were used to predict future prices. The study shows that one can achieve higher returns when considering time series with higher Hurst exponent and leaving out anti-persistent time series, i.e., data with a Hurst exponent  H < 0.5 .
  • Ref [30]: This research aims to predict stock market data, i.e., seven different recognized indexes, using three different neural network architectures, i.e., a feed-forward neural network, a cascade forward neural network, and learning vector quantization. Further, wavelet entropy and the Hurst exponent of the data under study are calculated for analysis and as an additional feature to be fed into the algorithms. The results show that the additional features improved the accuracy of the predictions.
  • Ref [31]: Eight different stock indices were analyzed and predicted using Hurst exponent, Shannon entropy, and Rényi entropy. Furthermore, the data was augmented using linear interpolation to generate more data points between actual data points. The final data set consists of eight stock indices, where for each of the indices, the features Open, High, Low, Close, Hurst exponent, Shannon entropy, Rényi entropy were considered. This generated an overall data set with 56 features. Then the whole dataset was normalized using the min-max normalization method, thus resulting in a total of 150 features. Predictions were made using Multi-Layer Regression (MLR), Support Vector Regression (SVR), and Feed Forward Back Propagation models. All models were tested using a different assemble of Hurst exponent, Rényi entropy, Shannon entropy components. The results show that the Hurst exponent component is crucial for making good predictions. The best results were obtained when using Feed Forward Back Propagation and including all three complexity features, i.e., Hurst exponent, Rényi entropy, Shannon entropy.
  • Ref [32]: Here, different approaches to predict the direction of US stock prices are compared with each other. Several algorithms were employed for this task, i.e., logistic regression, a multilayer perceptron (MLP), random forest, XGBoost, and a long short term memory (LSTM) recurrent neural network. Furthermore, effective transfer entropy (ETE) is calculated for analysis and as an additional feature to be fed into the algorithms. The results show that the predictions could be improved by adding ETE to the input and, second, that LSTM and the MLP perform best for this task.
  • Ref [7]: A fractal interpolation method is compared to a linear interpolation method to improve neural network time-series predictions. The fractal interpolation method uses the Hurst exponent to match the complexities of given sub-intervals of the time series data under study. Further, the neural network architecture is a Long Short Term Memory (LSTM) recurrent neural network. Furthermore, all results and data are analyzed using the Hurst exponent, the fractal dimension, and the spectrum of Lyapunov exponents. The results show that neural network predictions can significantly be improved using both linear and fractal interpolation techniques.
  • Ref [33]: The fractal interpolation method from [7] is used on a total of five datasets, with varying numbers of interpolation points. Further, these five data sets are then forecast using randomly parameterized ensembles of LSTM neural networks. These randomly parameterized ensemble predictions are then filtered using different complexity measures: The Hurst exponent, the spectrum of Lyapunov exponents, Fisher’s information, SVD entropy and Shannon’s entropy. The predictions could be improved by filtering the ensemble predictions. Further, the predictions outperformed baseline predictions using LSTM, GRU and RNN neural network approaches with one hidden layer. Further, all interpolated data sets are also analyzed using the previously mentioned complexity measures.

4. Complexity Measures

This section lists all complexity measures used in the discussed publications (Section 3), gives additional references and mentions how the complexity measures are used in combination with machine learning.
A table briefly summarizing and comparing the featured complexity measures can be found in Appendix A.

4.1. The Hurst Exponent

The rescaled range analysis (R/S analysis), as invented by Harold Edwin Hurst [34], is the procedure of determining the Hurst exponent “H” of time series data. The Hurst exponent is a measure of the long-term memory of a time series or a process.
As we only outline the main aspects of R/S analysis, we refer to [34,35] for a detailed discussion of the topic.
For a given a signal [ x 1 , x 2 , , x n ] , we find the the average over a period τ (a sub-interval of the signal 1 τ n ) as:
x τ , k = 1 τ j = k k + τ x j ,
where 1 k n for all elements i in this interval such that k i k + τ .
We further find an accumulated departure δ x i , τ , k over period a period i = 1 , 2 , , τ is:
δ x i , τ , k = j = k i x j x τ , k
The difference between maximal and minimal values of all x i in the interval k , k + τ , i.e., the range R of this interval τ is:
R τ , k = max δ x i , τ , k min δ x i , τ , k , satisfying k i k + τ .
The standard deviation for each sub interval is given as:
S τ , k = 1 τ i = k k + τ x i x τ , k 2 .
Averaging the range and the standard deviation over all possible k yields:
R τ = k R τ , k n k and S τ = k S τ , k n k ,
where 1 k n , k i k + τ and n k is the number of different ks. The Hurst exponent H is then defined using the scaling properties as:
R τ S τ τ H ,
The asymptotic behavior [34,36], for any independent random process with finite variance is then given as:
R τ S τ = π 2 v τ 1 2 ,
thus implies H = 1 2 . However, this is only true for completely random processes. For real-life data, we expect H 1 2 , as real-life processes usually feature long-term correlations.
The value of H varies as 0 < H < 1 . A value of H < 0.5 indicates anti-persistency, i.e., low values follow on high values and vice versa, thus heavily fluctuating, but not completely random. Values close to 0 indicate very strong anti-persistency. Contrarily, for values of H > 0.5 we expect persistent behavior and very strong persistency for values close to 1.
When it comes to combined applications with machine learning and R/S Analysis/the Hurst exponent, we find the following publications:
Ref [19]: In this research, the Hurst exponent was used to show long term memory in time series data as an argument for its predictability when using artificial neural networks for stock market data prediction.
Ref [21]: Here, the Hurst exponent was used to estimate the predictability of stock market data, and later to tailor the input nodes of a neural network architecture.
Ref [23]: For this study on stock market data predictability, the Hurst exponent was used to set the input window of an evolutionary neural network architecture.
Ref [22]: Here, it was shown that time-series data with a larger Hurst exponent can be predicted more accurately using an artificial neural network than data with a Hurst exponent closer to 0.5 .
Ref [24]: In this research, the Hurst exponent was used to identify regions in stock market data with higher predictability for several different machine learning approaches.
Ref [25]: Here, the Hurst exponent is used as a measure for the predictability of machine learning approaches.
Ref [17]: The researchers used the Hurst exponent to reject the random walk hypothesis of stock market data and as a measure for predictability.
Ref [28]: The researchers used the Hurst exponent to reject the random walk hypothesis of stock market data and as a measure for predictability.
Ref [29]: Here, the Hurst exponent was used to identify time series or regions with higher predictability to improve forecasts.
Ref [31]: Here, the Hurst exponent is used as an additional feature in the input of the employed algorithms to improve forecasts.
Ref [30]: Here, the Hurst exponent is used as an additional feature in the input of the employed algorithms to improve forecasts.
Ref [7]: Here, the Hurst exponent is used to, first, match a fractal interpolation to a data set and second, to analyze all data under study.
The Hurst exponent is used in [33] to filter LSTM ensemble predictions and to discuss all data under study.

4.2. Fractal Dimension of a Time-Series

The fractal dimension of a time series measures the complexity of a signal. The basic idea is can be understood as the area inhabited by the time series on a two-dimensional plane, i.e., the time series being placed upon a grid of equal spacing and checking the number of grid boxes necessary for covering it. Therefore, the fractal dimension is the ratio of the boxes covering the time series and the overall area of the plot. This process is referred to as box-counting. It has also been characterized as a measure of the space-filling capacity of a time series that tells how a fractal scales differently from the space it is embedded in; a fractal dimension does not have to be an integer. The fractal dimension of a self-affine time series can have values 1 < D < 2 .
Additionally, the fractal dimension is closely related to the Hurst exponent [21,37,38,39]. Therefore R/S analysis can be employed to calculate the fractal dimension of a time series using:
D = 2 H .
This originates from the fact that the long-range dependence (∼H) is closely related to self-similarity (∼D), and therefore the variability of the process.
Other ways to calculate the fractal dimension are the algorithm by Katz [40]; the algorithm by Higuchi [41]; and the algorithm by [42].
In [16], the fractal dimension is used as an additional feature for a fuzzy-logic-based forecasting technique, i.e., the fractal dimension is included when making the rules. Furthermore, a novel approach to calculate the fractal dimension of a geometric object, based on fuzzy logic is presented.
Ref [21]: This research employs the fractal dimension, alongside the Hurst exponent, of the time series under study to identify the best number of input nodes for a neural network. Further, the fractal dimension and the Hurst exponent are used to show that the time series under study is a non-random process that can in-theory be predicted.
In [26], the fractal dimension is used, together with an ant colony algorithm, as a tool for feature selection in a stock trend time series prediction task.
In [28], the fractal dimension, alongside the Hurst exponent, is used to indicate predictive behavior instead of a purely random behavior of the considered stock market data. Furthermore, Joseph’s effect and Noah’s effect are considered and discussed [43].
Ref [17]: Just as above, the fractal dimension, alongside the Hurst exponent, is used to show that the considered stock market data can in-theory be predicted. Furthermore, Joseph’s effect and Noah’s effect are considered and discussed [43].
Ref [7]: The fractal dimension alongside other complexity measures is used to analyze the data under study and the presented fractal interpolated data and how this relates to a neural network time series approach.

4.3. The Local Hölder Exponent

Following [20,44,45]:
Given a signal x i = x 1 , x 2 , , x n , we find the corresponding amplitudes as:
x ^ i = x i i = 1 n x i 2 such that : i = 1 n x ^ i 2 = 1 .
Next, very similar to R/S analysis, find windows or periods as i τ , i and the corresponding signal energies as:
E i , τ = i τ i x ^ i 2 .
Next we find the local Hölder or singularity exponent as:
α i τ = log E i , τ log τ .
Similar to the Hurst exponent or Fractal dimension, the exponent α quantifies degree of regularity or irregularity (singularity) in a distribution or a function at a point i for different possible window lengths τ , i.e., calculated via regression analysis.
Furthermore, for monofractal process, e.g., fractional Brownian motion, a l p h a i will be constant for all i, where for the case of multi-fractal processes, alpha changes from point to point, thus also provides a measure for the fractal behavior of time series data.
Ref [20]: In this research the local Hölder exponent was used, first o analyze the data under study, and second as an additional input feature to improve predictions.

4.4. Transfer Entropy (TE) and Effective Transfer Entropy (ETE)

Following [32,46,47]:
TE is used to measure the Granger-causal relationship between two processes, thus it can detect a non-linear Granger-causal relationship and is used in various disciplines, such as neuroscience and finance.
For two interacting time series X and Y, i.e., a variable Y can affect the future time point of the variable X. Then, if we consider the time series X to be Markov process of degree k, the state x i + 1 is thus affected by k previous states of the same variable:
p x n + 1 x n , x n 1 , , x 0 = p x n + 1 x n , x n 1 , , x n k + 1 , x i X .
Here p A B is a conditional probability of A given B, thus p A , B / p B . Further, the state x n + 1 depends on l previous states of Y, then we define the TE from a variable Y to a variable X as the average information included in Y excluding the information from the past state of X to the next state of X. Thus, employing Shannon’s entropy concept (see Section 4.6), we can define the TE from Y to X, where the state x n + 1 is affected by k previous states of X and l previous states of Y, as:
TE Y X ( k , l ) = i p x n + 1 , x n ( k ) , y n ( l ) log 2 p x n + 1 x n ( k ) , y n ( l ) i p x n + 1 , x n ( k ) , y n ( l ) log 2 p x n + 1 x n ( k ) = i p x n + 1 , x n ( k ) , y n ( l ) log 2 p x n + 1 x n ( k ) , y n ( l ) x n + 1 x n ( k ) .
Here the index i refers to i = { x n + 1 , x n ( k ) , y n ( l ) } , whereas x n ( k ) = x n , x n 1 , , x n k + 1 and y n ( l ) = y n , y n 1 , , y n l + 1 and p A , B is the joint probability of A and B.
Setting k = l = 1 to express the weak form of the efficient market hypothesis (when applied to financial data), i.e., the current price reflects all past information, we find TE as:
TE Y X = i p x n + 1 , x n , y n log 2 p x n + 1 x n , y n p x n + 1 x n = i p x n + 1 , x n , y n log 2 p x n + 1 , x n , y n p x n p x n + 1 , x n p x n , y n
and again i = { x n + 1 , x n , y n } .
Summed up, we can interpret TE Y X as the difference between the information for future values of X obtained from X and Y and the information for future values of X obtained only from X. Thus, a positive value of TE Y X indicates that Y affects future values of X; also, the larger the value, the more significant the information flow. Furthermore, because of its asymmetric properties, we can view TE Y X as the in-flow information from Y to X, and contrary to that, TE X Y as the out-flow information from X to Y.

Effective Transfer Entropy

TE measures a statistical dependency between two signals X and Y regardless of the data type. For TE to be calculated, a large amount of data is necessary. Another disadvantage is that TE includes noise from sample effects or the non-stationarity of the data set under study. To solve the noise issue, one can employ Effective Transfer entropy (ETE) [47]. To calculate ETE, one must first shuffle the elements of each time series understudy to keep the inherent probability distribution but break any causal relationships and dependencies. Then one obtains the TE from these randomized time series data, thus referred to as Randomized Transfer Entropy (RTE). The final step to determine ETE is then to subtract RTE from the original TE.
ETE Y X = TE Y X RTE Y X .
Here it is recommended to perform this process, starting with generating randomized time series data several times and average all obtained ETEs.
In [32], ETE is used to enhance the prediction of the direction of US stocks using a variety of different algorithms/approaches. Here it is used as an additional feature in the input of the algorithm.

4.5. Rényi Entropy

Following [48]: Rényi entropy is one of the families of functionals for quantifying diversity, uncertainty and randomness of a system. The Rényi entropy of order q is defined for q 1 (for q 1 as a limit) by:
S q X = 1 1 q log b i m p i q ,
where X is a discrete random variable, p i is the probability of the event { X = x i } , and b is the base of the logarithm. If the probabilities are all the same then all the Rényi entropies of the distribution are equal, with S q X = log b m . In any other case, the entropies are weakly decreasing as a function of q. Higher values of q, approaching infinity, give the Rényi entropy which is increasingly determined by consideration of only the highest probability events. Lower values of q, approaching zero, give the Rényi entropy which increasingly weights all possible events more equally, regardless of their probabilities.
Rényi entropy, and its alterations (different values of q), is one of the families of functionals to quantify the diversity, uncertainty, or randomness of a given system. Ref [31] uses Rényi entropy as an additional feature as input for the employed algorithms to improve predictions.

Related Complexity Measures

Taking into account the formal framework given above one can then find further generalizations and/or special cases [48]:
  • Hartley entropy or max-entropy:
    The Hartley or max-entropy is Rényi for q = 0
  • Shannon’s entropy:
    Shannon’s entropy is Rényi entropy with q = 1 . See Section 4.6.
  • Collision entropy:
    Rényi entropy with q = 2 is called collision entropy.
  • Min-entropy:
    In the limit q , the Rényi entropy converges to the min-entropy [49].
  • Kolomogorov-Sinai entropy:
    Another closely related entropy measure is the Kolomogorov-Sinai entropy. Here x t is the trajectory of the dynamic system in an n-dimensional phase space with time intervals Δ t , t > 0 . According to [48] the generalized Kolmogorov-Sinai entropy K q is then defined as:
    K q x i = lim r 0 lim Δ t 0 lim N 1 N Δ t 1 q 1 log b i 1 , i 2 , , i N m r p i 1 , i 2 , , i N q ,
    where the phase space is divided into n-dimensional hypercubes of side r and x i = x ( t = i Δ t ) . The p i 1 , i 2 , , i N is the joint probability that the trajectory x t = Δ t is in the box i 1 , x ( 2 Δ t ) is in the box i 2 , and x t = N Δ t is in the box i N .
    The Rényi entropy is a special case of the Kolmogorov-Sinai entropy, i.e., for a finite time interval ( N Δ t = 1 ), and for the probability that is not changing with the cell diameter ( m r = const . , p i 1 , i 2 , , i N = p i ).

4.6. Shannon’s Entropy

Given a signal [ x 1 , x 2 , , x n ] we then find the probability to occur for each value denoted as P x 1 , , P x n , thus we formulate Shannon’s entropy [50], as:
H Shannon = i = 1 n P x i log 2 P x i .
giving units as bits, as the base of the logarithm is set to 2. Shannon’s entropy is a measure for the uncertainty of a (random) process/signal.
For related complexity measures, see Section 4.5.
Ref [27] uses Shannon’s entropy first to analyze the data under study and second as a tool for feature selection to improve predictions.
Ref [31] uses Shannon’s entropy as an additional feature in the input for the employed Machine Learning algorithms to improve predictions.
Shannon’s entropy is used in [33] to filter LSTM ensemble predictions and to discuss all data under study.

4.7. Wavelet Entropy

Following [51,52,53]: Wavelet analysis introduces an appropriate basis and characterization of a signal by a distribution of amplitudes in the bass. Furthermore, one would construct this basis as an orthogonal basis, which has the advantage that arbitrary functions can be uniquely decomposed and recomposed, i.e., the decomposition can be inverted.
A family of wavelet packets can be understood as elemental signals obtained from appropriate linear combinations of wavelets, i.e., the wavelets constitute the signal. Wavelets are locally-oscillating forms of sines and cosines, i.e., they have a good localization on both frequency and time.
We can find a wavelet family ψ a , b t , where t is the time, as the set of elementary functions generated by a mother wavelet ψ t :
ψ a , b t = a 1 2 ψ t b a .
Here a , b R and a 0 are the scale/dilation and translation parameters, respectively.
We can then find two cases of wavelet transforms: The continuous wavelet transform (CWT) and the discrete wavelet transform (DWT). Both act on a signal x t L 2 R (the space of real square sumable functions).
The CWT of a continuous signal x t is defined as the correlation between the function x t and the wavelet family ψ a , b t for each a and b:
T a , b = 1 a d t x t ψ * t b a ,
where ψ * t is the complex conjugate of the analyzing wavelet function ψ t .
The DWT of a continuous signal x t L 2 R is found by first discretizing the wavelet family:
ψ m , n t = 1 a 0 m ψ t n b 0 a 0 m a 0 m .
Here, contrary to the CWT, we find two integer parameters m , n which control the scale/dilation and translation, respectively. For the fixed dilation and location step parameters a 0 and b 0 we further find find the conditions a 0 > 1 and b 0 > 0 . Here a common choice is a 0 = 2 and b 0 = 1 , which is referred to as dyadic grid scaling. From now on we will use ψ m , n with the dyadic grid scaling as:
ψ m , n t = 2 m 2 ψ 2 m t n .
Whereas this constitutes an orthonormal basis of L 2 R of finite energy signals.
Next we assume the signal be given by sampled values, i.e., x t = { x 1 , x 2 , , x M } . If we now carry out the decomposition over all resolution levels, i.e., with N = log 2 M , we find the wavelet expansion as:
x t = m = N 1 n C m n ψ m , n t = m = N 1 r m t .
Here we interpret wavelet coefficients C j k as the local residual errors between successive signal approximations at scales j and j + 1 , and we thus call r j t the residual signal at scale j.
As the { ψ m , n t } are an orthonormal basis for the space L 2 R we find the concept energy derived the same way as in Fourier theory as:
E m = r m 2 = n C m n 2 ,
with the corresponding energy at each sample time step n:
E n = m = N 1 C m n 2 .
The total energy thus is:
E tot = m < 0 n C m n 2 = m < 0 E m .
We can then find the total wavelet entropy S WT and the corresponding relative wavelet energy p m as:
S WT S WT p = m < 0 p m · ln p m , where p m = E m E tot .
Wavelet entropy is a measure of the order/disorder of a signal. Thus, it provides information about the corresponding dynamics and processes involved. This is because Shannon’s entropy idea, which for the wavelet entropy is used to analyze the different relative wavelet energies, gives a measure of the corresponding information of any distribution.
An ordered process can be interpreted as a periodic mono-frequency signal, thus having a narrow band spectrum. A wavelet representation of such a signal will almost entirely be resolved in one unique resolution level, i.e., all relative wavelet energies, but one will be almost zero. The non-zero one is thus the level that includes the representative signal frequency. For a signal like this, the corresponding wavelet entropy will be almost zero.
On the other hand, a random process would thus generate a very disordered signal. We expect a signal like this to have wavelet representations from all frequency bands. Further, one would expect these contributions to be of a similar order, much like the scaling seen in R/S analysis of a fractional Brownian motion. Thus, wavelet entropy will have its maximal values for signals like this.
In [30], wavelet entropy is used as an additional feature in the input for the employed algorithms to improve predictions.

4.8. Lyapunov Exponents of Time Series Data

The spectrum of Lyapunov exponents measures a system’s predictability depending on given initial conditions. For experimental time-series data, i.e., no different trajectories but only a single time series data, it still can be interpreted as a measure for predictability [54].
A Lyapunov exponent is a measure of exponential divergence, if, for a Lyapunov exponent λ , λ > 0 and convergence, if λ < 0 , respectively. A system embedded in an m-dimensional phase space has m Lyapunov exponents. A higher value of a Lyapunov exponent corresponds to lower predictability, i.e., a higher degree of chaos. In general, a positive Lyapunov exponent indicates chaos [9], thus in most cases, it is enough to calculate the first Lyapunov exponent, i.e., the first and largest exponent of the whole spectrum. A system that has more than one positive Lyapunov exponent is referred to as hyperchaotic [55].
There is a range of algorithms/approaches to calculate the spectrum of Lyapunov exponents or just the largest Lyapunov exponent from a time series data, the interest reader thus is referred to [54,56,57,58,59,60,61]. However, discussing each of these approaches and their similarities/differences is not within the scope of this article. To get an idea on how to calculate Lyapunov exponents from time series, we discuss the algorithm by [57], as it is a practical and very understandable approach to estimate the largest Lyapunov exponent of a signal.
Given a signal x i = x 1 , x 2 , , x n we find a reconstructed phase space (We note that we excluded articles dealing with phase space reconstruction in particular, but, because we focus on the spectrum of Lyapunov exponents as a measure for signal complexity, we neglect this discrepancy.):
X m τ i = x i , x i + τ , , x i + m 1 * τ ,
with time delay τ and an embedding dimension m. The time delay can be found via the method of average mutual information from [62] and the False Nearest Neighbours algorithm [63] can be used to determine the embedding dimension.
The algorithm from [57] looks for the nearest neighbor of each point on the trajectory, the corresponding Euclidean distance d 0 between two such neighboring points at instant 0 is defined as:
d i 0 = min X m τ j d X m τ j , X m τ i .
Thus, one finds the initial distance from the jth point to its nearest neighbor. Further, this algorithm is constrained, such that the nearest neighbors have a temporal separation which is greater than the mean period of the time series. The estimated largest Lyapunov exponent λ 1 then is the mean rate of separation of these nearest neighbors and is obtained from:
d j i C j e λ i Δ t ,
with C j being the initial separation. Taking the logarithm thus yields:
ln d j i = ln C j + λ 1 i Δ t ,
where Δ t is the sampling period of the time series. The corresponding distance d j i describes the distance between the jth pair of nearest neighbors after i discrete time steps. This can be interpreted as a set of approximately parallel lines, whereas the corresponding slope is roughly proportinal to the Largest Lyapunov Exponent λ 1 . The largest LYapunov exponent is then estimated using a least-squares fit to an averaged line defined as:
y i = 1 Δ t ln d j i .
Here · is the average over all values of j. This averaging allows an accurate evaluation of λ for short and noisy data.
In [7], the spectrum of Lyapunov exponents is used to analyze all the data under study, and the corresponding fractal interpolated data sets.
The spectrum of Lyapunov exponents is also used as filters for LSTM ensemble predictions in [33].

4.9. Fisher’s Information

As a measure of signal complexity, Fisher’s information is the amount of information extracted from a set of measurements, i.e., the quality of the measurements [64]. Another interpretation would be to use Fisher’s information to measure the order/disorder of a system or phenomenon. It is further well suited to investigate non-stationary and complex signals.
Given a signal [ x 1 , x 2 , , x n ] we can find a discrete version of Fisher’s information.
Here, the first step is to construct embedding vectors:
y i = x i , x i + τ , , x i + d E 1 * τ ,
with a fixed time delay τ and a corresponding embedding dimension d E . The so constructed embedding space, as a matrix, is:
Y = y 1 , y 2 , , y N d E 1 τ T .
The next step then is to perform a single value decomposition on this matrix [65]. This yields a total of Msingular values σ . We can then find normalized singular values as:
σ ¯ i = σ i j = 1 M σ j .
Finally, Fisher’s Information then is:
I Fisher = i = 1 M 1 σ ¯ i + 1 σ ¯ i 2 σ ¯ i .
As given above, Fisher’s information requires two parameters, first the time delay, which can be found using the calculation of the average mutual information from [62] for the time delay and for the embedding dimension the False Nearest Neighbours algorithm [63] can be used (There are more methods to determine a correct phase space embedding, but as a phase space reconstruction is not the focus of this article we refer to the given ones as they are the most famous ones.). Thus, constructing a phase space embedding of the signal.
Fisher’s information is used in [33] to filter LSTM ensemble predictions and to discuss all data under study.

4.10. SVD Entropy

SVD entropy is an entropy measure based on a single value decomposition [65], of a corresponding embedding matrix, thus similar to Fisher’s information (Section 4.9).
It can be used to assess the predictability of stock market data, as done in [66,67].
To calculate SVD entropy, one constructs an embedding space for a signal [ x 1 , x 2 , , x n ] with corresponding embedding vectors as [68]:
y i = x i , x i + τ , , x i + d E 1 * τ ,
with a corresponding embedding dimension d E and time delay τ . The time delay can be found using the calculation of the average mutual information from [62] and the embedding dimension can be found by employing the False Nearest Neighbours algorithm [63] (There are more methods to determine a correct phase space embedding, but as a phase space reconstruction is not the focus of this article we refer to the given ones as they are the most famous ones.). Thus, constructing a phase space embedding of the signal, represented by a matrix:
Y = y 1 , y 2 , , y N d E 1 τ T .
The next step is then to perform a single value decomposition [65], on this matrix to get M singular values σ 1 , , σ M , i.e., the singular spectrum. A corresponding normalized spectrum of singular values is then found via:
σ ¯ i = σ i j = 1 M σ j .
And finally, using the concept of Shannon’s entropy (see Section 4.6) yields SVD entropy as:
H SVD = i = 1 M σ ¯ i log 2 σ ¯ i .
SVD entropy is used in [33] to filter LSTM ensemble predictions and to discuss all data under study.

4.11. Some Honorably Mentions for Measures of Signal Complexity

Furthermore, we want to give a list of other complexity measures, that from the authors’ point of view, make sense to be combined with machine learning approaches in one of the discussed ways (see Section 6).
  • Generalized Hurst Exponent:
    The generalized Hurst exponents method is a tool to study the scaling properties of the data, and it is a generalization of the approach from Section 4.1. Note that the generalized Hurst exponent  H q , contrary to the regular Hurst exponent, depends on a variable q, which means it is associated with the qth order moment of the distribution of the increments [69].
  • Generalized Fractal Dimension:
    One can then find a connection between the Rényi entropy (see Section 4.5) and a generalized fractal dimension  D q . Following [48]:
    S q r = D q log b r ,
    where S q r is the Rényi entropy of order q, whereas q 1 (for q 1 as a limit). D q is the corresponding generalized fractal dimension of order q, b is the base of the logarithm and r is the size of the cells of the embedding. For the generalized fractal dimension, we thus compute the probability with which the data points fall into the i-th cell.
  • Approximate Entropy (ApEn):
    Approximate entropy was initially developed by Steve M. Pincus to analyze medical data [70] and afterwards adapted to general biologic network systems [71]. Later it spread it’s applications into various fields including finance [72] and psychology [73]. The motivation for this is that many approaches to calculate the entropy of a time series require huge data sets and cannot handle noisy data sets. ApEn assigns a non-negative number to a time series, where larger values indicate a greater randomness apparent in the process. From a statistical point of view, ApEn can be seen as an ensemble parameter of process auto-correlation, i.e., smaller values correspond to greater positive auto-correlation, larger values indicate greater independence.
  • Sample Entropy (SampEn):
    Like Approximate entropy, sample entropy is a measure of complexity [72]. The difference to ApEn is that it does not include self-similar patterns and that it does not depend on the length of the data set.
  • Range Entropy (RangeEn):
    The basic idea behind ApEn and SampEn is to calculate the Chebyshev distance between two state vectors in the reconstructed m-dimensional phase space [74]. However, this is limited because these formulations do not have upper limits.

5. Machine Learning Methods

In this section, we briefly discuss the used Machine Learning approaches. Note that, when we refer to machine learning approaches, we mean all sorts of data-based prediction/regression/classification algorithms that occurred in the listed research, such as k-nearest neighbors, Random Forest, Artificial Neural Networks, fuzzy Logic approaches, etc. Further, as there was a significant amount of papers featuring neural network approaches with varying architectures, we collected all of the neural network approaches together and again briefly discuss the employed architectures in the neural network context.

5.1. k Nearest Neighbours (kNN)

k nearest neighbours (kNN) [75] is a model-free algorithm for classification and regression. Like the name says, values or classes are predicted based on the k nearest neighbors, i.e., which class is dominant within these data points.
In [24], kNN, among other algorithms, is used to predict stock market data. Further, R/S analysis is used to identify non-random regions within the data.

5.2. Decision Trees

Decision trees are an algorithm for classification and regression that can be understood as that a tree-like structure to sort all occurring scenarios, thus containing only conditional control statements [76]. In [24], Decision Trees, among other algorithms, are used to predict stock market data. Further, R/S analysis is used to identify non-random regions within the data.

5.3. Random Forest

Random Forest [77], is an ensemble learning method for classification and regression analysis based on decision trees. The basic idea is to generate an ensemble of decision trees using bagging and later combine the results. The Idea of Bagging is to train many weak learners on the same data set and combine the result.
In [27], Random Forest, among other approaches, is used for electricity load forecasting. Furthermore, Shannon’s entropy is employed for feature selection.
Random Forest, among other algorithms, is used in [28] in conjunction with R/S analysis to predict stock market movement.
Two similar Random Forest algorithms are used in [17] to forecast different exchange rates. Furthermore, R/S analysis and the fractal dimension are used to analyze the data for predictability.
Random Forest, among other approaches, is used in [32] to predict the direction of US stocks. Here Effective Transfer entropy (ETE Section 4.4) is used as an additional feature to improve predictions.

5.4. Tree Based Extreme Gradient Boosting (XGBoost)

EXtreme Gradient Boosting (XGBoost) is a tree based algorithm. The term boosting refers to combining the results of many weak predictions to a strong one, i.e., an ensemble of weak classifiers. Here, the selection of weak classifiers has to be optimized. Further, boosting is generalizable by allowing optimization of an arbitrary differentiable loss function.
XGboost was first proposed in [78] as part of greedy function approximations. Nowadays, the standard decision tree-based ensemble method was developed by [79].
In [32], XGBoost among other approaches was used together with Effective Transfer Entropy (ETE, see Section 4.4) to predict the direction of US stocks.

5.5. Logistic Regression

Logistic regression is a tool for regression analysis very similar to linear regression, but instead of a linear function, one employs a sigmoid function [80]. Logistic regression is useful for binary or multinomial dependent variables.
In [32], a Logistic Regression among other approaches was used together with Effective Transfer Entropy (ETE, see Section 4.4) to predict the direction of US stocks.

5.6. Support Vector Machines

The basic idea of Support Vector Regression [81,82] is to perform linear regression in a high-dimensional feature space using the kernel-trick, i.e., employing kernel functions. Thus, a linear model in the new space can represent a non-linear model in the original space. Support Vector machines and/or regression was first proposed in [83].
As previously mentioned, we map the input onto a high-dimensional feature space. Here one uses non-linear mappings to represent non-linear separations in the original space, i.e., one constructs an optimally separating hyperplane. The data points on each side of the hyperplane thus have a distance to the hyperplane. The smallest of these distances is called the margin of separation, and we will refer to the margin of the optimal hyperplane as q and the corresponding data points are called the support vectors. All other training examples are irrelevant for defining the class boundaries. SVM thus finds that particular linear model, i.e., the maximum margin hyperplane.
In [26], support vector machines are used together with a feature selection method based on the fractal dimension to predict the direction of stock indices on a daily basis.
Several approaches, among them, support vector machines, are used in [28] to predict stock market movement. Furthermore, R/S analysis is employed to test for market efficiency.
In [31], different forecast methodologies such as SVM are used to predict stock market behavior. Furthermore, Shannon’s entropy, Rényi entropy, and the Hurst exponent were used as additional features to improve the algorithm.

5.7. Multi-Linear Regression

Multilinear regression is the generalization of the standard linear regression model for multiple of dependent and independent variables [84,85]. It can be used to predict, e.g., prices. The basic idea is to find the correlation between two factors, i.e., the dependent and the independent variables. The corresponding assumption here is that the conditional mean function is linear.
In [31] mulit-linear regression and other approaches are used to predict stock market behavior. Furthermore, Shannon’s entropy, Rényi entropy, and the Hurst exponent were used as additional features to improve the forecasts. Ref [16] uses Multi-linear regression in addition to neural networks to predict the U.S./Mexico exchange rate. Furthermore, the fractal dimension of the time series data under study is used to analyze and improve the fuzzy logic approach.

5.8. Fuzzy Logic Learning Approaches

There exists a variety of fuzzy logic-based prediction methods, such as adaptive neuro-fuzzy inference systems (ANFIS [86]), Dynamic Evolving Neural-Fuzzy Inference System (DENFIS [87]), or Fuzzy Inductive Reasoning, (FIR [88]). All of these approaches can be broken down to the fuzzification of the data under study and learning rules, or train an algorithm, such as a neural network, based on these fuzzy sets, and then make predictions with the so learned behavior [89].
The work in [16] builds fuzzy rules based on both the actual time series data and the corresponding fractal dimension of the observed object to predict the data under study.
In [27], FIR, among other approaches, is used for electricity load forecasting. Furthermore, Shannon’s entropy is employed for feature selection.
Both ANFIS and DENFIS are used in [28] in conjunction with R/S analysis to predict stock market movement.

5.9. Artificial Neural Networks

One of today’s most common approaches to predict time series are artificial neural networks (ANNs) [90]. Artificial neural networks (sometimes just neural networks) are learning algorithms that emerged from the idea of human or animal brains. As such, they are constituted of connected neurons. Here, the way how they are connected is crucial for tasks they are employed for. Each neural network consists of an input layer, hidden layers, and an output layer. For our research, we categorize the mentioned applications into two types of neural networks. First, feed-forward neural networks, i.e., neural networks that are connected in a non-cyclic way, and recurrent neural networks, which are connected cyclically, i.e., the input and output or sub-parts of the network are connected such that the neurons form loops [91].
Further, the neurons have specific (non-linear) activation functions and are controlled by adjusting weights for each neuron. Thus, the learning process constitutes of adjusting and readjusting the weights on each neuron, through, e.g., backpropagation [92].

5.9.1. Feed Forward Artificial Neural Networks

Feedforward neural networks are neural networks that are connected in a non-cyclic manner. The feed-forward neural network and its prototype, the multi-layer perceptron, were the first and most simple neural networks. Information propagates in only one direction, i.e., from input to output nodes.
Several different architectures for feed-forward artificial neural networks with backpropagation were tested in [19], such as with one hidden layer or two hidden layers, to predict stock market data. Here the Hurst exponent was used to indicate predictability of the time series under study, i.e., a long-term memory.
The study in [16] employs a three-layered feed-forward neural network with backpropagation to be tested against a combined fuzzy logic and fractal theory approach; further, the fractal analysis was applied to analyze all-time series that were to be predicted in this study.
In both [24,27] a standard three-layer feed-forward artificial neural network (one hidden layer) together with backpropagation was used to predict stock market data. In [24], the Hurst exponent was used to identify periods of time series that are easier to predict. Furthermore, in [27] Shannon’s entropy was used as a tool for feature selection.
Furthermore, in [31] a feed-forward artificial neural network with backpropagation was used. Again, the researchers employed a standard architecture of a three-layered ANN, thus one input layer, one output layer, and one hidden layer. Here, combinations of the Hurst exponent, Shannon entropy, and Rényi entropy were used as additional features to improve stock market predictions. In [30], again, a three-layered feed-forward artificial neural network was used to predict long-term trends of stock indices. Here wavelet entropy and the Hurst exponent were used as additional features for the algorithm to improve predictions.
Sometimes, or some types of feed-forward neural networks are referred to as Multi-Layer perceptrons. A multi-layer perceptron [90], refers to simple feed-forward artificial neural networks. The standard architecture consists of three layers, an input layer, a hidden layer, and an output layer. In contrast, the hidden and the output layer both consist of neurons with non-linear activation functions.
In [32], a MLP among other approaches was used together with Effective Transfer Entropy (ETE, see Section 4.4) to predict the direction of US stocks. In [23], an MLP is used together with the Hurst exponent to predict stock market data.
As it is very similar, we also discuss the cascade forward neural network in the context of feed-forward neural networks. The cascade forward neural network [93], has the same architecture as a feed-forward neural network but has a connection from the input and every previous layer to the following layers.
In [30], cascade forward neural networks are used together with R/S analysis and Wavelet entropy to improve forecasting for different time series data.
Another used architecture here is that of time-lagged neural networks. A time-lagged feed-forward neural network has a sliding time window as input. Thus, the input nodes (with the corresponding input) are ordered such that the input is a sequence. This architecture was used in [29] in combination with the Hurst exponent. Here, the Hurst exponent was used to identify time series data for better predictability and financial returns.
Though referred to as a time-delay neural network, which is usually used for speech recognition, the basic architecture of the neural network used in [21] is that of time-lagged neural networks. This research combines time-lagged neural networks with fractal analysis, e.g., Hurst exponent and fractal dimension, to predict stock price data. Here the fractal analysis improves the neural network architecture, i.e., tailor the input window.
Learning Vector Quantization (LVQ) is a supervised classification algorithm which can be used for classification and regression [94,95]. It is a precursor of Kohonen’s self-organizing maps [94], and is thus also inspired by the self-organizing capabilities of neurons in the visual cortex. It is a two-layered feed-forward neural network that uses a competitive (winner-take-all) learning strategy. In [30], (LVQ) was used, among other algorithms and by using the Hurst Exponent and wavelet entropy, to forecast financial time series data.

5.9.2. Recurrent Artificial Neural Networks

Contrary to feed-forward neural networks are recurrent neural networks, where information can propagate in loops in the network. This architecture was specifically invented to learn temporal dynamic behavior, whereas speech recognition and time series prediction [96], are exemplary applications.
In [20], an artificial neural network with feedback connection, i.e., a recurrent neural network, was used to predict geomagnetic properties. Further, the local Hölder exponent was used as an additional feature to improve predictions.
In [23], an elmann recurrent neural network was used, among other architectures, together with the Hurst exponent, to tailor the input windows of the algorithms.
In [25], a recurrent neural network with embedded memory, i.e., a non-linear autoregressive model process with exogenous input (NARX), is used to forecast several different time-series data. Furthermore, R/S analysis, i.e., the Hurst exponent, is used to indicate and verify the predictability of the neural network approach.
Ref [28]: A Jordan neural network [97], was used among other approaches to predict stock market behavior. Furthermore, R/S analysis was employed to test for market efficiency.
A long short term memory (LSTM [98]) recurrent neural network, among other other approaches is used in [32] to predict the direction of US stocks. Here Effective Transfer entropy (ETE Section 4.4) is used as an additional feature to improve predictions.
In [7], a Hurst-exponent-based fractal interpolation is used to increase the data points of different time series data. These data are then forecast using a long short term memory (LSTM [98]) recurrent neural network. The fractal interpolation, and the forecasts are analyzed using a three different complexity measures, i.e., the fractal dimension (Section 4.2), the Hurst exponent (Section 4.1) and the spectrum of Lyapunov exponents (Section 4.8).
In [33], a Hurst-exponent-based fractal interpolation is used to increase the data points of different time series data. Then ensemble predictions are produced using randomly parameterized LSTM neural networks. Furthermore, finally, these predictions are filtered using five different complexity measures, i.e., the spectrum of Lyapunov exponents, The Hurst exponent, Shannon’s entropy, Fisher’s information, and SVD entropy. Further, this article uses a single hidden layer LSTM, gated recurrent unit (GRU [99]) and simple recurrent neural network as baseline predictions, i.e., to compare the ensemble results to.

6. Summary and Conclusions

This review analyzed the applications of a combination of complexity measures and machine learning algorithms for time series prediction. We found a total of 18 publications fitting our criteria (defined in Section 2).
Given the analysis discussed above, we found six types of combinations which we can sort all listed publications (Section 3) into (A table summing up these findings can be found in Appendix B):
  • Complexity measures as an additional criterion for analysis: [7,16,17,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33]
  • Complexity measures to improve the architecture of the employed neural network/ algorithm: [21,23]
  • Complexity measures as additional features for machine learning algorithms: [16,20,30,31,32]
  • Complexity measures to find regions of increased predictability: [22,24,28,29]
  • Feature Selection using Complexity Measures: [26,27,29]
  • Filtering predictions/ensembles using Complexity Measures: [33]
The authors propose that combined approaches of machine learning and complexity measures have valuable contributions for both machine learning researchers and engineers as:
A complexity-based feature selection may save resources and improve the overall performance of the algorithm at hand.
An additional complexity feature can increase the accuracy of the algorithm at hand.
Analyzing the data and the predictions under study using complexity measures, one can argue for the quality of models based on their ability to recreate the complexity characteristics of the training data. Thus, the authors propose, complexity measures may become important for measuring the generalizability of models.
Models with higher accuracy and lower computational cost can be built taking into account the complexity of the data under study when designing the architecture of the algorithm, e.g., the neural network architecture.
Neural Network predictions can be sped up by circumventing the problem of finding the best set of hyper-parameters and instead choosingthe best predictions of different neural networks based on the complexity of the prediction.
Finding regions of increased predictability can yield higher returns for applications of machine learning in the finance sector.
Machine learning ensembles that are designed taking into account the complexity of the data under study may outperform today’s machine learning ensemble approaches.

Author Contributions

Conceptualization, S.R.; Data curation, S.R.; Funding acquisition, T.N.; Investigation, S.R.; Methodology, S.R.; Project administration, T.N.; Resources, T.N.; Supervision, T.N.; Validation, S.R.; Visualization, S.R.; Writing—original draft, S.R.; Writing—review and editing, S.R. and T.N. All authors have read and agreed to the published version of the manuscript.

Funding

Open Access Funding by TU Wien; The state of Lower Austria: Forum Morgen-01112018; FFG Project AI4Cropr, No. 877158.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors acknowledge the funding of the project “DiLaAg—Digitalization and Innovation Laboratory in Agricultural Sciences”, by the private foundation “Forum Morgen”, the Federal State of Lower Austria, by the FFG; Project AI4Cropr, No. 877158 and by TU Wien Bibliothek for financial support through its Open Access Funding Program. Open Access Funding by TU Wien.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Measures of Signal Complexity: Summary Table

The following table summarizes the assertions, the applicability and the advantages/disadvantages of all discussed complexity measures from Section 4.
Complexity MeasureApplicabilityAssertionProsCons
The Hurst Exponentunivariate
time series data
persistency,
anti-persistency,
randomness
- only one parameter, i.e., the Hurst exponent
- Hurst exponent lives within a fixed range,
i.e., H 0 ; 1
- average over all scales, i.e., blind to
multifractal behavior
Fractal Dimension
of a Signal
univariate
time series data
persistency,
anti-persistency,
randomness,
sginal-complexity
- only one parameter, i.e., the fractal dimension
- The fractal dimension of a signal lives within a fixed
range, i.e., d F 1 ; 2
- can be calculated from the Hurst exponent
via d F = 1 H
- average over all scales, i.e., blind to
multifractal behavior
- if calculated from R/S analysis, it states
the same as the Hurst exponent
- many algorithms to choose from
The Local Hölder
Exponent
univariate
time series data
regularity,
irregularity,
varying scaling
behavior
- can identify regions of different scaling behavior,
i.e., multi-fractality
- varies from time step to time step
- overall no comparable classification of time
series data
- yields one Hölder exponent for every time step
(Effective) Transfer
Entropy
multivariate
time series data
information
transfer between
time series data
- can detect non-linear Granger causal relationships
- effective transfer entropy compensates for noise and
non-stationarity
- only applicable between two time series data
- transfer entropy contains noise
- requires a large amount of data
Rényi Entropyunivariate
time series data
diversity,
uncertainty,
randomness
- one can choose between weitghing events with
higher or lower probability by adjusting a parameter q
- not recommended for real-valued data
- requires and additional parameter q
Shannon’s Entropyunivariate
time series data
diversity,
uncertainty,
randomness
- classical well understood complexity/information
measure
- not recommended for real-valued data
The Spectrum of
Lyapunov Exponents
of time series data
univariate
time series data
predictability,
chaoticity,
- for many applications, it is sufficient to consider the
largest Lyapunov exponent i.e., the first of the spectrum
- the spectrum of Lyapunov exponents can be used to
characterize the underlying dynamics of signals
- yields a spectrum of parameters
- computationally expensive
- only clearly defined for different trajectories
and not for univariate data
- for some algorithms a phase space embedding
of the time series under study is required, i.e.,
the embdding dimension and the time delay
Fisher’s informationunivariate
time series data
quality of
measurements,
order/disorder
- well suited for non-stationary and complex signals
- yields only parameter, i.e., the information content
of the data under study
- requires two additional parameters, the
embedding dimension and the time delay
SVD Entropyunivariate
time series data
predictability,
order/disorder
- well suited for non-stationary and complex signals
- yields only parameter, i.e., the entropy of the data
under study
- requires two additional parameters, the
embedding dimension and the time delay
Wavelet Entropyunivariate
time series data
order/disorder,
randomness
- yields only one parameter, i.e., the wavelet entropy
of a signal
- provides information on underlying dynamics
- wavelet transform needs be performed
beforehand
- one has to choose a wavelet representation of
a signal

Appendix B. Results: Summary Table

The following table summarizes our findings. The entries of the fourth column are sorted using the different approaches featured in Section 6.
PublicationEmployed Complexity MeasuresEmployed Machine Learning AlgorithmsApproach Type
[19]Hurst ExponentFeed Forward Neural Network1
[16]Fractal DimensionFeed Forward Neural Network
Fuzzy Logic Approaches
1, 3
[20]Local Hölder ExponentRecurrent Neural Network1, 3
[21]Hurst Exponent, Fractal DimensionTime Delayed Neural Network1, 2
[22]Hurst ExponentFeed Forward Neural Network1, 4
[23]Hurst ExponentFeed Forward Neural Network
Recurrent Neural Network
1, 2
[24]Hurst ExponentFeed Forward Neural Network
k-nearest Neighbours
Decision Tree
1, 4
[25]Hurst exponentNonlinear Autoregressive models with eXogenous input
Recurrent Neural Networks
1
[26]Fractal DimensionSupport Vector Machines1,5
[27]Shannon’s entropyFuzzy Inductive Reasoning
Random Forest, Feed Forward Neural Network
1, 5
[28]Hurst Exponent, Fractal DimensionAdaptive Neuro-Fuzzy Inference System
Dynamic Evolving Neuro-Fuzzy Inference System
Recurrent Neural Network, Support Vector Regression
Random Forest
1, 4
[17]Hurst Exponent, Fractal DimensionRandom Forest1
[29]Hurst exponent, Fractal DimensionTime Lagged Feed Forward Neural Network1, 4, 5
[30]Hurst Exponent, Wavelet EntropyFeed Forward Neural Network
Cascade Forward Neural Network
Learning Vector Quanitzation
1, 3
[31]Hurst Exponent
Rényi entropy
Shannon’s entropy
Multi-Linear Regression, Support Vector Regression
Feed Forward Neural Networks
1, 3
[32]Effective Transfer EntropyLogistic Regression
Feed Forward Neural Network (MLP)
Random Forest, XGBoost
Long Short Term Memory Neural Network
1, 3
[7]Hurst Exponent
The spectrum of Lyapunov exponents
Fractal Dimension
Long Short Term Memory Recurrent Neural Network1
[33]Hurst Exponent
The spectrum of Lyapunov exponents
Fisher’s information
SVD entropy
Shannon’s entropy
Long Short Term Memory Recurrent Neural Network1, 6

References

  1. Schmidt, J.; Marques, M.R.G.; Botti, S.; Marques, M.A.L. Recent advances and applications of machine learning in solid-state materials science. NPJ Comput. Mater. 2019, 5, 83. [Google Scholar] [CrossRef]
  2. Voyant, C.; Notton, G.; Kalogirou, S.; Nivet, M.L.; Paoli, C.; Motte, F.; Fouilloy, A. Machine learning methods for solar radiation forecasting: A review. Renew. Energy 2017, 105, 569–582. [Google Scholar] [CrossRef]
  3. Larrañaga, P.; Calvo, B.; Sanatana, R.; Bielza, C.; Galdiano, J.; Inza, I.; Lozano, J.A.; Armañanzas, R.; Santafé, G.; Pérez, A.; et al. Machine learning in bioinformatics. Brief. Bioinform. 2005, 7, 86–112. [Google Scholar] [CrossRef] [Green Version]
  4. Rajkomar, A.; Dean, J.; Kohane, I. Machine Learning in Medicine. N. Engl. J. Med. 2019, 380, 1347–1358. [Google Scholar] [CrossRef]
  5. Heaton, J.B.; Polson, N.G.; Witte, J.H. Deep Learning in Finance. arXiv 2016, arXiv:1602.06561v3. [Google Scholar]
  6. Chlingaryan, A.; Sukkarieh, S.; Whelan, B. Machine learning approaches for crop yield prediction and nitrogen status estimation in precision agriculture: A review. Comput. Electron. Agric. 2018, 151, 61–69. [Google Scholar] [CrossRef]
  7. Raubitzek, S.; Neubauer, T. A fractal interpolation approach to improve neural network predictions for difficult time series data. Expert Syst. Appl. 2021, 169, 114474. [Google Scholar] [CrossRef]
  8. Friedrich, J.; Gallon, S.; Pumir, A.; Grauer, R. Stochastic Interpolation of Sparsely Sampled Time Series via Multipoint Fractional Brownian Bridges. Phys. Rev. Lett. 2020, 125, 170602. [Google Scholar] [CrossRef]
  9. Henriques, T.; Ribeiro, M.; Teixeira, A.; Castro, L.; Antunes, L.; Costa Santos, C. Nonlinear Methods Most Applied to Heart-Rate Time Series: A Review. Entropy 2020, 22, 309. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Mandelbrot, B.B. Fractals and Scaling in Finance: Discontinuity, Concentration, Risk. Selecta Volume E; Springer: New York, NY, USA, 1997. [Google Scholar]
  11. Devynck, P.; Wang, G.; Antar, G.; Bonhomme, G. The Hurst exponent and long-time correlation. Phys. Plasmas 2000, 7, 1181–1183. [Google Scholar]
  12. Sakai, K. Nonlinear Dynamics and Chaos in Agricultural Systems, 1st ed.; Elsevier Science: Amsterdam, The Netherlands, 2001. [Google Scholar]
  13. Tsonis, A.A.; Elsner, J.B. Chaos, Strange Attractors, and Weather. Bull. Am. Meteorol. Soc. 1989, 70, 14–23. [Google Scholar] [CrossRef] [Green Version]
  14. Fu, Q.; Liu, Y.; Li, T.; Liu, D.; Cui, S. Analysis of Irrigation Water Use Efficiency Based on the Chaos Features of a Rainfall Time Series. Water Resour. Manag. 2017, 31, 1961–1973. [Google Scholar] [CrossRef]
  15. Chen, W.K.; Wang, P. Fuzzy Forecasting with Fractal Analysis for the Time Series of Environmental Pollution. In Time Series Analysis, Modeling and Applications; Springer: Berlin/Heidelberg, Germany, 2013; pp. 199–213. [Google Scholar]
  16. Castillo, O.; Melin, P. Hybrid Intelligent Systems for Time Series Prediction Using Neural Networks, Fuzzy Logic, and Fractal Theory. IEEE Trans. Neural Netw. 2002, 13, 1395–1408. [Google Scholar] [CrossRef]
  17. Ghosh, I.; Chaudhuri, T.D. Fractal Investigation and Maximal Overlap Discrete Wavelet Transformation (MODWT)-based Machine Learning Framework for Forecasting Exchange Rates. Stud. Microecon. 2017, 5, 105–131. [Google Scholar] [CrossRef]
  18. Raubitzek, S.; Neubauer, T. Machine Learning and Chaos Theory in Agriculture. ERCIM News 2020, 2020. [Google Scholar]
  19. Yao, J.; Lim Tan, C.; Poh, H.L. Neural Networks for Technical Analysis: A Study on KLCI. Int. J. Theor. Appl. Financ. 1999, 2, 221–241. [Google Scholar] [CrossRef] [Green Version]
  20. Vörös, Z.; Jankovičová, D. Neural network prediction of geomagnetic activity: A method using local Hölder exponents. Nonlinear Process. Geophys. 2002, 9, 425–433. [Google Scholar] [CrossRef]
  21. Yakuwa, F.; Dote, Y.; Yoneyama, M.; Uzurabashi, S. Novel Time Series Analysis & Prediction of Stock Trading using Fractal Theory and Time Delayed Neural Net-work. In Proceedings of the IEEE SMC’03 2003 IEEE International Conference on Systems, Man and Cybernetics. Conference Theme-System Security and Assurance (Cat. No.03CH37483), Washington, DC, USA, 5–8 October 2003. [Google Scholar]
  22. Qian, B.; Rasheed, K. Hurst exponent and financial market predictability. In Proceedings of the 2nd IASTED International Conference on Financial Engineering and Applications, Berkeley, CA, USA, 24–26 September 2004; pp. 203–209. [Google Scholar]
  23. Selvaratnam, S.; Kirley, M. Predicting Stock Market Time Series Using Evolutionary Artificial Neural Networks with Hurst Exponent Input Windows. In Australasian Joint Conference on Artificial Intelligence; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germnay, 2006; Volume 4304, pp. 617–626. [Google Scholar] [CrossRef]
  24. Qian, B.; Rasheed, K. Stock market prediction with multiple classifiers. Appl. Intell. 2007, 26, 25–33. [Google Scholar] [CrossRef]
  25. Diaconescu, E. The use of NARX neural networks to predict chaotic time series. WSEAS Trans. Comput. Res. 2008, 3, 182–191. [Google Scholar]
  26. Ni, L.P.; Ni, Z.W.; Gao, Y.Z. Stock trend prediction based on fractal feature selection and support vector machine. Expert Syst. Appl. 2011, 38, 5569–5576. [Google Scholar] [CrossRef]
  27. Jurado, S.; Nebot, À.; Mugica, F.; Avellana, N. Hybrid methodologies for electricity load forecasting: Entropy-based feature selection with machine learning and soft computing techniques. Energy 2015, 86, 276–291. [Google Scholar] [CrossRef] [Green Version]
  28. Ghosh, I.; Sanyal, M.K.; Jana, R.K. Fractal Inspection and Machine Learning-Based Predictive Modelling Framework for Financial Markets. Arab. J. Sci. Eng. 2018, 43, 4273–4287. [Google Scholar] [CrossRef]
  29. De Mendonça Neto, J.N.; Lopes Fávero, L.P.; Takamatsu, R.T. Hurst exponent, fractals and neural networks for forecasting financial asset returns in Brazil. Int. J. Data Sci. Anal. 2018, 3, 29–49. [Google Scholar] [CrossRef]
  30. Karaca, Y.; Baleanu, D. A Novel R/S Fractal Analysis and Wavelet Entropy Characterization Approach for Robust Forecasting Based on Self-Similar Time Series Modelling. Fractals 2020, 28, 2040032. [Google Scholar] [CrossRef]
  31. Karaca, Y.; Zhang, Y.D.; Muhammad, K. Characterizing Complexity and Self-Similarity Based on Fractal and Entropy Analyses for Stock Market Forecast Modelling. Expert Syst. Appl. 2020, 144, 113098. [Google Scholar] [CrossRef]
  32. Kim, S.; Ku, S.; Chang, W.; Song, J.W. Predicting the Direction of US Stock Prices Using Effective Transfer Entropy and Machine Learning Techniques. IEEE Access 2020, 8, 111660–111682. [Google Scholar] [CrossRef]
  33. Raubitzek, S.; Neubauer, T. Taming the Chaos in Neural Network Time Series Predictions. Entropy 2021, 23, 1424. [Google Scholar] [CrossRef]
  34. Hurst, H.; Black, R.; Sinaika, Y. Long-Term Storage in Reservoirs: An experimental Study; Constable: London, UK, 1965. [Google Scholar]
  35. Di Matteo, T. Multi-scaling in finance. Quant. Financ. 2007, 7, 21–36. [Google Scholar] [CrossRef]
  36. Feller, W. An Introduction to Probability Theory and Its Applications; Wiley-Blackwell: Hoboken, NJ, USA, 1971. [Google Scholar]
  37. Feder, J. Fractals (Physics of Solids and Liquids); Springer: New York, NY, USA, 1988. [Google Scholar]
  38. Pang, J.; North, C.P. Fractals and their applicability in geological wireline log analysis. J. Pet. Geol. 1996, 19, 339–350. [Google Scholar] [CrossRef]
  39. Breslin, M.; Belward, J. Fractal dimensions for rainfall time series. Math. Comput. Simul. 1999, 48, 437–446. [Google Scholar] [CrossRef]
  40. Katz, M.J. Fractals and the analysis of waveforms. Comput. Biol. Med. 1988, 18, 145–156. [Google Scholar] [CrossRef]
  41. Higuchi, T. Approach to an irregular time series on the basis of the fractal theory. Phys. D Nonlinear Phenom. 1988, 31, 277–283. [Google Scholar] [CrossRef]
  42. Petrosian, A. Kolmogorov complexity of finite sequences and recognition of different preictal EEG patterns. In Proceedings of the Eighth IEEE Symposium on Computer-Based Medical Systems, Lubbock, TX, USA, 9–10 June 1995; pp. 212–217. [Google Scholar] [CrossRef]
  43. Mandelbrot, B.B.; Wallis, J.R. Noah, Joseph, and Operational Hydrology. Water Resour. Res. 1968, 4, 909–918. [Google Scholar] [CrossRef] [Green Version]
  44. Halsey, T.C.; Jensen, M.H.; Kadanoff, L.P.; Procaccia, I.; Shraiman, B.I. Fractal measures and their singularities: The characterization of strange sets. Nucl. Phys. B Proc. Suppl. 1987, 2, 501–511. [Google Scholar] [CrossRef]
  45. Vörös, Z. On multifractality of high-latitude geomagnetic fluctuations. Ann. Geophys. 2000, 18, 1273–1282. [Google Scholar] [CrossRef]
  46. Schreiber, T. Measuring Information Transfer. Phys. Rev. Lett. 2000, 85, 461–464. [Google Scholar] [CrossRef] [Green Version]
  47. Marschinski, R.; Kantz, H. Analysing the information flow between financial time series. T Eur. Phys. J. B Condens. Matter Complex Syst. 2002, 30, 275–281. [Google Scholar] [CrossRef]
  48. Zmeskal, O.; Dzik, P.; Vesely, M. Entropy of fractal systems. Comput. Math. Appl. 2013, 66, 135–146. [Google Scholar] [CrossRef]
  49. König, R.; Renner, R.; Schaffner, C. The Operational Meaning of Min- and Max-Entropy. IEEE Trans. Inf. Theory 2009, 55, 4337–4347. [Google Scholar] [CrossRef] [Green Version]
  50. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  51. Blanco, S.; Figliola, A.; Quiroga, R.Q.; Rosso, O.A.; Serrano, E. Time-frequency analysis of electroencephalogram series. III. Wavelet packets and information cost function. Phys. Rev. E 1998, 57, 932–940. [Google Scholar] [CrossRef]
  52. Rosso, O.A.; Blanco, S.; Yordanova, J.; Kolev, V.; Figliola, A.; Schürmann, M.; Başar, E. Wavelet entropy: A new tool for analysis of short duration brain electrical signals. J. Neurosci. Methods 2001, 105, 65–75. [Google Scholar] [CrossRef]
  53. Addison, P. Wavelet transforms and the ECG: A review. Physiol. Meas. 2005, 26, R155-99. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  54. Eckmann, J.P.; Kamphorst, S.; Ciliberto, S. Liapunov exponents from time series. Phys. Rev. A 1987, 34, 4971–4979. [Google Scholar] [CrossRef]
  55. Zengrong, L.; Liqun, C.; Ling, Y. On properties of hyperchaos: Case study. Acta Mech. Sin. 1999, 15, 366–370. [Google Scholar] [CrossRef]
  56. Wolf, A.; Swift, J.B.; Swinney, H.L.; Vastano, J.A. Determining Lyapunov exponents from a time series. Phys. D Nonlinear Phenom. 1985, 16, 285–317. [Google Scholar] [CrossRef] [Green Version]
  57. Rosenstein, M.T.; Collins, J.J.; De Luca, C.J. A practical method for calculating largest Lyapunov exponents from small data sets. Phys. D Nonlinear Phenom. 1993, 65, 117–134. [Google Scholar] [CrossRef]
  58. Briggs, K. An improved method for estimating Liapunov exponents of chaotic time series. Phys. Lett. A 1990, 151, 27–32. [Google Scholar] [CrossRef]
  59. Brown, R. Calculating Lyapunov exponents for short and/or noisy data sets. Phys. Rev. E 1993, 47, 3962–3969. [Google Scholar] [CrossRef] [PubMed]
  60. Sano, M.; Sawada, Y. Measurement of the Lyapunov Spectrum from a Chaotic Time Series. Phys. Rev. Lett. 1985, 55, 1082–1085. [Google Scholar] [CrossRef] [PubMed]
  61. Zeng, X.; Eykholt, R.; Pielke, R.A. Estimating the Lyapunov-exponent spectrum from short time series of low precision. Phys. Rev. Lett. 1991, 66, 3229–3232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  62. Fraser, A.M.; Swinney, H.L. Independent coordinates for strange attractors from mutual information. Phys. Rev. A 1986, 33, 1134–1140. [Google Scholar] [CrossRef]
  63. Krakovská, A.; Mezeiová, K.; Budáčová, H. Use of False Nearest Neighbours for Selecting Variables and Embedding Parameters for State Space Reconstruction. J. Complex Syst. 2015, 2015, 932750. [Google Scholar] [CrossRef] [Green Version]
  64. Mayer, A.L.; Pawlowski, C.W.; Cabezas, H. Fisher Information and dynamic regime changes in ecological systems. Ecol. Model. 2006, 195, 72–82. [Google Scholar] [CrossRef]
  65. Klema, V.; Laub, A. The singular value decomposition: Its computation and some applications. IEEE Trans. Autom. Control. 1980, 25, 164–176. [Google Scholar] [CrossRef] [Green Version]
  66. Caraiani, P. The predictive power of singular value decomposition entropy for stock market dynamics. Phys. A Stat. Mech. Appl. 2014, 393, 571–578. [Google Scholar] [CrossRef]
  67. Gu, R.; Shao, Y. How long the singular value decomposed entropy predicts the stock market?—Evidence from the Dow Jones Industrial Average Index. Phys. A Stat. Mech. Appl. 2016, 453. [Google Scholar] [CrossRef]
  68. Roberts, S.J.; Penny, W.; Rezek, I. Temporal and spatial complexity measures for electroencephalogram based brain–computer interfacing. Med. Biol. Eng. Comput. 1999, 37, 93–98. [Google Scholar] [CrossRef]
  69. Sánchez, R.; Newman, D. A Primer on Complex Systems with Applications to Astrophysical and Laboratory Plasmas; Springer: New York, NY, USA, 2018. [Google Scholar]
  70. Pincus, S.M. Approximate entropy as a measure of system complexity. Proc. Natl. Acad. Sci. USA 1991, 88, 2297–2301. [Google Scholar] [CrossRef] [Green Version]
  71. Pincus, S.M. Irregularity and asynchrony in biologic network signals. Methods Enzymol. 2000, 321, 149–182. [Google Scholar]
  72. Delgado-Bonal, A.; Marshak, A. Approximate Entropy and Sample Entropy: A Comprehensive Tutorial. Entropy 2019, 21, 541. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  73. Richman, J.S.; Moorman, J.R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiol. Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef] [Green Version]
  74. Omidvarnia, A.; Mesbah, M.; Pedersen, M.; Jackson, G. Range Entropy: A Bridge between Signal Complexity and Self-Similarity. Entropy 2018, 20, 962. [Google Scholar] [CrossRef] [Green Version]
  75. Fix, E.; Hodges, J.L. Discriminatory Analysis. Nonparametric Discrimination: Consistency Properties. Int. Stat. Rev. Rev. Int. Stat. 1989, 57, 238–247. [Google Scholar] [CrossRef]
  76. Fürnkranz, J. Decision Tree. In Encyclopedia of Machine Learning; Sammut, C., Webb, G.I., Eds.; Springer: Boston, MA, USA, 2010; pp. 263–267. [Google Scholar] [CrossRef]
  77. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  78. Friedman, J.H. Greedy Function Approximation: A Gradient Boosting Machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  79. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; Association for Computing Machinery: New York, NY, USA, 2016; pp. 785–794. [Google Scholar] [CrossRef] [Green Version]
  80. Cox, D.R. The Regression Analysis of Binary Sequences. J. R. Stat. Soc. Ser. B (Methodol.) 1958, 20, 215–242. [Google Scholar] [CrossRef]
  81. Cristianini, N.; Shawe-Taylor, J. An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods; Cambridge University Press: Cambridge, UK, 2000. [Google Scholar] [CrossRef]
  82. Awad, M.; Khanna, R. Support Vector Regression. In Efficient Learning Machines: Theories, Concepts, and Applications for Engineers and System Designers; Awad, M., Khanna, R., Eds.; Apress: Berkeley, CA, USA, 2015; pp. 67–80. [Google Scholar] [CrossRef] [Green Version]
  83. Boser, B.E.; Guyon, I.M.; Vapnik, V.N. A training algorithm for optimal margin classifiers. In Proceedings of the 5th Annual ACM Workshop on Computational Learning Theory, Pittsburgh, PA, USA, 27–29 July 1992; pp. 144–152. [Google Scholar]
  84. Su, Y.; Gao, X.; Li, X.; Tao, D. Multivariate Multilinear Regression. IEEE Trans. Syst. Man, Cybern. Part B (Cybern.) 2012, 42, 1560–1573. [Google Scholar] [CrossRef]
  85. Sopipan, N.; Kanjanavajee, W.; Sattayatham, P. Forecasting SET50 Index with Multiple Regression based on Principal Component Analysis. J. Appl. Financ. Bank. 2012, 2, 271–294. [Google Scholar]
  86. Jang, J.S.; Sun, C.T.; Mizutani, E. Neuro-Fuzzy and Soft Computing-A Computational Approach to Learning and Machine Intelligence [Book Review]. Autom. Control. IEEE Trans. 1997, 42, 1482–1484. [Google Scholar] [CrossRef]
  87. Kasabov, N.; Song, Q. DENFIS: Dynamic Evolving Neural-Fuzzy Inference System and Its Application for Time-Series Prediction. Fuzzy Syst. IEEE Trans. 2002, 10, 144–154. [Google Scholar] [CrossRef] [Green Version]
  88. Klir, G.; Elias, D. Architecture of Systems Problem Solving, 2nd ed.; IFSR International Series in Systems Science and Systems Engineering; Springer: New York, NY, USA, 2003. [Google Scholar] [CrossRef]
  89. Singh, P. A brief review of modeling approaches based on fuzzy time series. Int. J. Mach. Learn. Cybern. 2017, 8, 397–420. [Google Scholar] [CrossRef]
  90. Jain, A.; Mao, J.; Mohiuddin, K. Artificial neural networks: A tutorial. Computer 1996, 29, 31–44. [Google Scholar] [CrossRef] [Green Version]
  91. Abiodun, O.I.; Jantan, A.; Omolara, A.E.; Dada, K.V.; Mohamed, N.A.; Arshad, H. State-of-the-art in artificial neural network applications: A survey. Heliyon 2018, 4, e00938. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  92. Kelley, H.J. Gradient theory of optimal flight paths. ARS J. 1960, 30, 947–954. [Google Scholar] [CrossRef]
  93. Warsito, B.; Santoso, R.; Suparti; Yasin, H. Cascade Forward Neural Network for Time Series Prediction. J. Phys. Conf. Ser. 2018, 1025, 012097. [Google Scholar] [CrossRef]
  94. Kohonen, T. Self-Organizing Maps, 3rd ed.; Springer Series in Information Sciences; Springer: Berlin/Heidelberg, Germany, 2001. [Google Scholar] [CrossRef]
  95. Brownlee, J. Clever Algorithms: Nature-Inspired Programming Recipes, Revision, 2nd ed.; LuLu.com: Morrisville, NC, USA, 2012. [Google Scholar]
  96. Tealab, A. Time series forecasting using artificial neural networks methodologies: A systematic review. Future Comput. Inf. J. 2018, 3, 334–340. [Google Scholar] [CrossRef]
  97. Jordan, M.I. Attractor Dynamics and Parallelism in a Connectionist Sequential Machine. In Artificial Neural Networks: Concept Learning; IEEE Press: Hoboken, NJ, USA, 1990; pp. 112–127. [Google Scholar]
  98. Hochreiter, S.; Schmidhuber, J. Long Short-term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  99. Cho, K.; van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1724–1734. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Raubitzek, S.; Neubauer, T. Combining Measures of Signal Complexity and Machine Learning for Time Series Analyis: A Review. Entropy 2021, 23, 1672. https://0-doi-org.brum.beds.ac.uk/10.3390/e23121672

AMA Style

Raubitzek S, Neubauer T. Combining Measures of Signal Complexity and Machine Learning for Time Series Analyis: A Review. Entropy. 2021; 23(12):1672. https://0-doi-org.brum.beds.ac.uk/10.3390/e23121672

Chicago/Turabian Style

Raubitzek, Sebastian, and Thomas Neubauer. 2021. "Combining Measures of Signal Complexity and Machine Learning for Time Series Analyis: A Review" Entropy 23, no. 12: 1672. https://0-doi-org.brum.beds.ac.uk/10.3390/e23121672

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop