Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Emerging interdependence between stock values during financial crashes

Abstract

To identify emerging interdependencies between traded stocks we investigate the behavior of the stocks of FTSE 100 companies in the period 2000-2015, by looking at daily stock values. Exploiting the power of information theoretical measures to extract direct influences between multiple time series, we compute the information flow across stock values to identify several different regimes. While small information flows is detected in most of the period, a dramatically different situation occurs in the proximity of global financial crises, where stock values exhibit strong and substantial interdependence for a prolonged period. This behavior is consistent with what one would generally expect from a complex system near criticality in physical systems, showing the long lasting effects of crashes on stock markets.

Introduction

Financial markets in general and the drivers for market crashes in particular have been intensively investigated by the physics community in recent years. The interest in markets’ behavior stems from a number of different reasons relating both to the narrow financial interests of shareholders and investors and, arguably more importantly, to the devastating impact financial turmoil may have on national economies, leading to harsh social consequences and societal unrest. Moreover, the observation of a sudden and dramatic collapse in complex systems intrigues the scientific community due to its resemblance to collective rearrangement in physical systems at critical points. This problem is at the core of the emerging field of econophysics [13], which borrows mathematical and physical tools such as random matrix theory [4, 5], clustering analysis [6], extreme and rare events [7], agent based models [8] and network theory [9], to name a few, to tackle the complexity of economical and financial systems. Network theory provided a framework for analyzing economic structures [912] from the perspective of complex systems [13], and is rooted in a much earlier search for structures in financial markets [14]. The success of this approach is partly due to shortcomings of existing economics-based theories in addressing the complexity of financial systems, and partly because many phenomena such as financial bubbles, herding, contagion and crashes found a natural interpretation in physical models that involve multi agents, collective behavior, influence spreading and phase transitions.

Phase transitions are sudden reorganizations of the system occurring when an external parameter, such as the temperature, is tuned to a critical value. Moreover, at this critical point, the system is scale-invariant, leading to a power law behavior of the observables whose critical exponents may be studied using renormalization group and scaling theory [15]. A similar symmetry, discrete scale invariance [16], has been shown to give rise to log-periodicity of prices in the proximity of financial collapses [17]. Switching trends in financial markets have been analysed in [18], where first order phase transitions, like those occurring in low temperature Ising models when tuning an external magnetic field, have been proposed as for explaining herding behaviour in the proximity of crises. Important contributions to the analogy between financial crashes and phase transition came also from the study of ecological and climate systems. In particular, a large effort has been devoted to studying the precursors of collapses [1921], showing that crashes are usually anticipated by a loss of resilience in the system; in other words, when approaching a critical point, perturbations take more time to be reabsorbed and are more likely to propagate. This picture is consistent with that of physical systems near the critical point of a phase transition, where the cross correlation between the fundamental degrees of freedom of the systems is very large and the system exhibits high susceptibility.

The study of correlations in financial networks has started a couple of decades ago and led to the development of several effective algorithms for extracting the underlying network topology [9, 10, 2225]. Their hierarchical structure could then be used to identify groups of stocks in terms of the corresponding economical sectors. These works originally focused on same-time correlations and only recently have been generalized to deal with the concept of causality [26] in financial data [27]. The study of directed influences, aiming at predicting future prices, is interesting for investors for maximizing their returns; however, it also has the potential to forecast the macroscopic behavior of markets. This objective is highly ambitious, but predicting the behavior of macroscopic properties from observation and modeling is at the heart of statistical physics and is of theoretical interest on its own right. Attempts in this direction has been made using Granger causality [28] and information theoretical methods such as mutual information and Transfer Entropy (TE) [29], which outperform simple retarded correlations in capturing non-linear influences [30, 31].

In this paper we generalize existing information theoretical approaches to study temporal interdependencies between financial indices in a time period of several years. More precisely, we analyze interdependencies between stocks of the FTSE 100 companies, which includes the 100 largest companies (in terms of market capitalization) listed on the London Stock Exchange, from 2000 to 2015. Exploiting the physical intuition that financial crashes may be anticipated by periods of large susceptibility, we use this method to investigate directed influences among the corresponding index constituents looking for a similar behavior, with the limited resolution of daily stock values. Most of the literature in this area relies on intra-days influences since it is commonly believed that traders are well informed when making their decision, so that directed correlations and influences can be detected only at short time scales. This belief is an interpretation of the “efficient market hypothesis” (EMH) [32], according to which there is no hope to predict and outperform the market. Strictly speaking, this is called “strong market efficiency” and originated from the faith in the ability of the market to quickly process all the available information and to push prices towards their fundamental values [33]. As we will discuss, it is not always true and these situations are usually described in terms of “weakly” efficient markets. In fact, analyses of daily stock prices reveal that they do not behave as martingales [30, 34] and that during the crash of 2008 price movements have been highly correlated [35, 36]. Similar problems have been studied also in [37, 38] where it has been noticed that the aggregated correlation between stocks sharply increases after a period of market downtrend. We investigate this behavior analyzing moving periods of roughly two years each and find results that support the common view according to which little information can be extracted from the past to predict future values at daily time scales. However, employing directed information theoretical methods inspired by TE, we measure a boost in the overall information flow between index components in periods corresponding to the crash of 2008 and the Eurozone debt crisis of 2010-2012.

The paper is organized in the following way. The second section focuses on the materials and methods used, and is divided to three subsections. The first introduces the information theoretical methods, which are commonly used for extracting direct influences, the second provides details on the dataset and the third explains the null model used to validate our analysis. The third section contains the results and the following ones provide a discussion and conclusion. Further details on the methods are provided in the Supporting Information.

Materials and methods

Measuring influences

In order to measure the influences among stocks we used an Information theoretical tool, which follows from a generalization of the Symbolic Transfer Entropy (STE) [39]. TE and STE are powerful methods able to measure the amount of information flow between time series and thus can be used to reconstruct the network of influences between components of a complex systems. Some of the most interesting cases where they have been successfully employed include network reconstruction of functional areas of the brain [40, 41], the study of social phenomena [42] and the influence of social networks on financial markets [43]. They have also been used in finance to analyze the relations between indices [44, 45] and components of indices [31].

Transfer entropy evaluates the information gained on future values of a time series X(t) by observing past values of another time series Y(t) in addition to the past values of X(t), relying on estimating the probabilities of occurrences of time series values. For real valued time series this is more difficult but STE, making use of symbolization, provides an effective solution to this problem. A symbol of k−literals of the time series X(t) at time s is obtained by reordering the last k values of the time series at time s (i.e. {xsk, …, xs−1}) in an ascending order. By generalizing the time step from 1 to an integer δ we transform the data to a set of k−dimensional symbols at times s + δ, denoted by . The role of the time scale δ, which reflects the inherent effective delayed interaction between components will be investigated later. A more formal definition as well as further details on the corresponding information theoretical measures are provided in the Supplementary Information.

In this work, we evaluate the influence of time series Y(t) on X(t) by computing the following quantity (1) This is the Kullback-Leibler divergence between the probabilities and , averaged over past symbols. Alternatively it can be viewed as the difference between the conditional entropies of the two probabilities. If Y(t) contains no information about X(t) this measure is zero. Practically, due to the noisy nature of the data this never happens and one finds non zero values even when the two systems do not interact. This calls for the introduction of a null model to extract the genuine underlying behavior from the dataset; this will be discussed in detail later on. Finally we would like to emphasize that our measure aims at predicting (k+1)-dimensional symbols by looking at k-dimensional historic symbols. This is a slight modification of the STE measure, where k-dimensional symbols are predicted by looking at k-dimensional historic symbols. Further details are provided in the Supplementary Information.

Dataset

We collected financial time series data of the component stocks of FTSE 100 from 3 January, 2000 to 15 May, 2015 (around 4000 trading days), available from Yahoo! Finance [46]. Data is freely available at [47]. Clearly, the FTSE 100 index is continuously changing over time as companies migrate in and out of the index [48]. To mitigate the effect, rather than taking into account individual component changes, we considered the list of the FTSE 100 constituents in May 2015, and analyzed the related time series, tracing them back to January 2000. About 60% of the stocks that have been included in the May 2015 FTSE 100 index were already included in the index in 2000. Discarding short-lived stocks, we labeled the remaining N = 97 stocks from 1 to 97 in ascending alphabetical order of their ticker symbol. These time series have a time resolution of one (trading) day and we focused on the closing prices. Instead of looking at the whole time series, we analyze time windows of Ω = 500 days. This time window is being shifted by ω = 25 days across about 4000 trading days for which we have data, investigating the time evolution of the network structure. Since the stationarity hypothesis, useful in estimating the probabilities Eq (1), is unlikely to hold for long periods, studying shorter time windows would help in getting more reliable estimates. We process the time series linked to each stock in order to obtain the geometric returns r(t) at the time scale δ: (2) where p(t) is the price at time t (closing price on day t). In each time window, we compute the information flow between time series at different δ values using Eq (1); while log-ratios with large δ values are expected to carry little or no information we will show that also log-ratios with small δ values do not, in certain periods. To further control errors and ensure that the stocks considered had existed for long enough to give rise to meaningful influences, we restrict the computation of TYX to cases where the number of days the considered pair of stocks {X(t), Y(t)} have in common is at least 80% of the time window Ω.

Surrogate dataset

To validate our results and eliminate spurious instances of entropy transfer, we construct a null model of non interacting components. This may be done in several ways [31, 42, 44, 45], ranging from a random reshuffle of the original time series to more refined methods [49, 50]. A simple reshuffling of data, while clearly destroying the interdependence among different time series, also destroys the single time series structure. The null model we use, based on the theory of surrogate data [51], does allow one to preserve the spectral properties of the original spectrum in spite of the randomization. Under the assumption that the single time series structure can be effectively represented by the power spectrum of the signal, a general time series X(t) can be randomized via the generation of the time series (3) where X(k) is the Fourier transform of the original signal, , and ϕ(k) is a random phase attached to each Fourier component such that ϕ(−k) = −ϕ(k), so that is real. The series is thus a randomized version of X(t) but having the same power spectrum. We construct the null model by randomizing the original time series of the closing prices. Then, for a given δ, we process these time series to obtain random returns using Eq (2) and compute the influences between the surrogate time series using Eq (1). We compared the information flow in our original dataset and the null model to identify true information from noise since the latter does not contain genuine information flow between series.

Results

We analyzed the evolution of the network of influences between stocks in each of the (about) 140 time windows indexed by w. We associate a value I(X, Y) ∈ [0, 1] to each directed link {XY} in order to measure the amount of genuine information flow from X(t) to Y(t), (4) where the parameters a and r* are set to 100 and 0.03, respectively. From a qualitative point of view, r(x) represents the probability that the information flow x = TXY value calculated from the real dataset has been obtained at random. Further details on this definition are provided in S1 File, see Eq (13). The quantities I(X, Y) are supposed to vary slowly from one time window to the next, say w to w + 1; conversely, the parameter ω = 25 that controls the shift between consecutive time-windows, may be too large or our results may not be sufficiently stable with respect to small changes. To address this issue we introduced the quantity (5) where, the expression within the absolute value sign is the change in the genuine information flow originating from stock i in two consecutive time windows.

To estimate the total information flow, we introduce the quantity (6) and study its behavior as a function of δ and w. While the results for large δ confirm our expectations that no information flow can be detected, the results obtained at small δ as a function of w are much more interesting. In particular, analyzing geometric return time series by setting k = 2 in Eq (1), we obtain the results shown in Fig 1. At small values of δ this quantity sharply peaks around the financial crisis of 2008, while this effect fades away as δ increases. We also notice that a similar behavior is observed during the period of the Eurozone debt crisis between 2010 and early 2013, and then again fades away as δ increases. The robustness of our results can be checked analyzing the behavior of . As can be seen in Fig 2, this quantity, representing the information flow in two consecutive time windows, is close to zero most of the time and peaks at a value smaller than 1. This confirms our assumption that the structure of influences is evolving smoothly. Moreover, it drops to zero when δ increases, since there is no information flow in any windows.

thumbnail
Fig 1. Total information flow as a function of time.

The behavior of the total information flow , defined in Eq (6), at different time scales δ. Each time window w is 500 days long. The date associated to each w is the middle of the time window considered. The x-axis tick marks represent the first of March of every year. While at short time scales (less than 3 days) we observe a peak around the two major financial crises of the last decades, this effect fades away as δ increases. Interestingly, the results at δ = 2 carries much more information than those at δ = 1.

https://doi.org/10.1371/journal.pone.0176764.g001

thumbnail
Fig 2. Network evolution.

Plot of , the information flow in two consecutive time windows, defined in Eq (5), at time differences δ = 2 and δ = 3. Each time window w is 500 days long. The date associated to each w is the middle of the time window considered. The x-axis tick marks represent the first of March of every year. This quantity measures the evolution of the detected structure of influences. We observe a smooth behavior, meaning that structures in consecutive time windows are similar, except for during crises where more pronounced market readjustment take place.

https://doi.org/10.1371/journal.pone.0176764.g002

Finally, by manipulating the matrix I(X, Y), it is possible to probe a more detailed structure of the influences, introducing the information directionality flow (7) The first term is a summation across columns, measuring the total information flow originating from the stock n, while the second is a summation across rows, measuring the total information flow directed toward n, originated from other stocks. The difference between these two terms provides a measure of whether the stock n is influencing the market (Δn > 0) or is being influenced by the market (Δn < 0), and by how much. A plot of the stock’s directionality measures with time provides knowledge about how the stock’s role in the market has been evolving, as shown in Figs 35. This picture also provides another check on the slow evolution of the matrices I(Xi, Xj) during several time windows.

thumbnail
Fig 3. Information directionality flow.

For each component n, we evaluate the information directionality flow Δn, defined in Eq (7), measuring how much the component has influenced (or has been influenced by) the market. Positive values are associated to lead effects. The horizontal axis refers to the window time index w. The vertical axis refers to the component index. It is interesting to see how strength and directionality of influences become clearer and more emphasized at time of financial crises. A closer look at these values is provided in Figs 4 and 5.

https://doi.org/10.1371/journal.pone.0176764.g003

thumbnail
Fig 4. Directionality leaders.

To identify more clearly leading stocks in terms of their effect on others, we present the information of Fig 3, but focussing on the 30 components with the largest directionality flow values. For the sake of clarity each time tick has been obtained by averaging three consecutive time windows. So we have about 45 different ticks rather than the original 140 time windows.

https://doi.org/10.1371/journal.pone.0176764.g004

thumbnail
Fig 5. Directionality followers.

To identify more clearly stocks led by the market, we present the information of Fig 3, but focussing on the 30 components with the smallest directionality flow values. For the sake of clarity each time tick has been obtained by averaging three consecutive time windows. So we have about 45 different ticks rather than the original 140 time windows.

https://doi.org/10.1371/journal.pone.0176764.g005

Discussion

To validate the results obtained using a different approach we employed another method to investigate the data. This method is based on extracting all the pairs {X, Y} for which TXY is larger than a given threshold and then, using the same threshold, extracting all pairs of a surrogate dataset of time series for which the same condition is satisfied. Finally, we compare the number of links extracted in the two cases. The surrogate dataset, obtained following the protocol outlined above, does not include genuine interactions between its degrees of freedom at any of the time windows considered and we do not expect it to reproduce the patterns found in Fig 1. More specifically, we do not expect to observe an abrupt increase in the number of links around crises. Fig 6 shows that such significant increases are not observed for the surrogate data and therefore these are genuine phenomena of the real dataset.

thumbnail
Fig 6. Thresholding method.

We compute the ratio of the number of directed links detectable in the real and surrogate datasets. We define this ration as 1/rt. The threshold used is 0.03 but similar qualitative results are obtained for other values. We see that as δ increases this ratio approaches 1, while at short time scales it resembles the results of Fig 1.

https://doi.org/10.1371/journal.pone.0176764.g006

We would like to point out that while we mainly discuss results obtained for k = 2, we also carried out the analysis for larger k values. Computing the matrices TYX for larger k values requires much more time; additionally we were unsuccessful in extracting meaningful information I(X, Y) even for k = 3. This is disappointing but is confirmed by other works [44] studying the TE measure Eq (1) by varying the parameter k, where it has been shown that large values of k provide uninformative results. A possible reason is the signal-to-noise ratio, which is very small already for k = 2 and is presumably lost in more complicated models of influences. Moreover, when increasing k at a fixed Ω, we notice that the quality of the probability estimation in Eq (1) decreases.

We repeated the analysis presented in the Result section by using the row prices rather than the returns, even if they are usually not studied in this context. As explained in the Supporting information, there are more or less conservative link extraction protocols, where, roughly speaking, the first ones lead to extract (maybe too) few trustworthy links and the last ones lead to extract (maybe too) many links at the price of considering many spurious links. Employing a less conservative protocol when considering time series of prices we obtained results that resemble those presented above while the exact same protocol used before leads to uninteresting results where no information can be extracted in any window, as shown in Fig 7. This is consistent with the common belief that returns, in general, carry more useful information than prices.

thumbnail
Fig 7. Total information flow as a function of time—The prices time series.

Behavior of the total information flow Eq (6) at different time differences δ, when computed for daily prices rather than returns. Each time window w is 500 days long. The date associated with each w is the middle of the time window considered. The x-axis tick marks correspond to the dates of the first of March each year. No meaningful information can be obtained from this analysis in contrast to the results presented in Fig 1.

https://doi.org/10.1371/journal.pone.0176764.g007

Finally, we would like to comment on the composition of the list of companies considered. As previously stated, the index composition is constantly changing and rather than taking into account these effects we studied the price time series of companies that were part of the index in May 2015. While theoretically the dynamics of the index may be influenced by this choice, the explosion in information flow detected at the onset of the 2008 financial crisis is unlikely to be sensitive to this selection. In fact, as seen in Fig 1, the value of in 2008 is larger than the background value of the previous year by a factor of about 100. At the same time, the index has a solid backbone, since about 60% of the companies that were part of the index in May 2015 were already included in it in January 2000 [48].

Conclusions

Market efficiency, or more precisely market unpredictability, has been studied using retarded correlations [27], Granger causality [31] and information theoretical methods [30]. Looking at short time scales, these studies find that log-returns are strongly correlated up to time scales of hours. Longer time periods had been analyzed in [35], using mutual information on daily time series of the closing prices of the components of the US stock market, that was found to be very large during the financial crash of 2008. In our work we develop an information theoretical measure inspired by TE to compute the information flow among daily time series of the closing prices of the components of the FTSE 100 index. Respect to mutual information, TE-like measures are able to detect temporal and directed influences. In fact, they are able to disentangle information exchanged between two inputs from that due to a common driving source, as can be seen in Eq (1). Moreover, they improve on the standard linear methods such as retarded correlations or the Granger causality measure [30, 43], allowing one to extract influences that would be otherwise hardly detectable. Our analysis describes the structure of the market in a long time period spanning from 2000 to 2015, exposing a critical behavior at times of financial crises, as suggested in [35]. Moreover, the analysis supports the dominant viewpoint according to which no information can be extracted at long time scales (of days) at normal day-to-day operations, but sheds light on the emergence of interlinked information flow at times that are close to critical events. We clearly observe that information flow between stocks during non-crises periods have short-lived effects on the market (the market is informationally efficient) whereas during crises they exert their influence over a larger range of time scales, having longer-lasting effects. Near such events the market is thus highly inefficient. Interestingly, we notice that early signals of crashes could already be detected in the markets months before the full manifestation of the crises. The Lehman Brothers bankruptcy, that arguably marks the onset of the 2008 financial crisis, is dated to September 15, 2008, and while this event is clearly recognizable in Fig 1, we also notice that the cross-stock information flow started to grow months in advance, when the American subprime mortgage market started to unfold. Whether or not this method can predict catastrophic events is unclear, but it can definitely measure the susceptibility of financial markets and their robustness to volatility as it exposes the strengthening of long-range correlations, in analogy to complex systems close to a phase transition. The reasons for the inefficiency of financial markets during crashes may have several explanations. The main assumption behind EMH is that traders have rational expectations and that information is easily available to different traders. It is likely that none of the two remain valid during these periods, when herding behavior is known to play a very important role [52]. Moreover, if information were not easily available or if it were processed in different ways by different traders, price movements could become informative and the market become informationally inefficient [33]. According to this point of view, financial crashes can be viewed as collective reorganizations of the market from preexisting equilibria, or possibly out of equilibrium states, to new equilibrium states, taking place on time scales of a few days. The reason why markets may not be at equilibrium before crises, i.e. the reason why prices may not reflect their fundamental values and, thus, markets may not be Pareto efficient, need to be carefully considered. Future work will apply this method to probe information flows at shorter time scales in emergent markets. Looking at finer time scale would provide useful insights in order to characterize the state of markets even far from global crises in weakly efficient markets.

Supporting information

S1 File. Supporting information.

Here we give further information about the definition of the transfer entropy, the symbolisation technique and the methods used to filter genuine information from noise.

https://doi.org/10.1371/journal.pone.0176764.s001

(PDF)

S1 Fig. Transfer entropy values—Real and surrogate data.

Histograms of values found in the sets and for w = 80, i.e. the period of November 2008, using the time scale of Fig 1. The inset shows the same quantities computed at w = 45, i.e. September 2005. While in the second case no information flows can be detected, in the first, using the protocol discussed in this section, many directed influenced can be obtained. This matrix values refer to δ = 2, for which the amount of information is maximized.

https://doi.org/10.1371/journal.pone.0176764.s002

(EPS)

S2 Fig. Correlations between transfer entropy values—Real and surrogate data.

Scatter plot of the values found in at δ = 2 in w = 80 versus those found in the surrogate dataset at the same w and δ. The values do not appear to show any correlation between the two.

https://doi.org/10.1371/journal.pone.0176764.s003

(EPS)

Acknowledgments

We would like to thank David Lowe for interesting discussions. This work is supported by The Leverhulme Trust grant RPG-2013-48. E.T. also wishes to thank the Lee Hysan Foundation of Hong Kong for funding.

Author Contributions

  1. Conceptualization: DS JR.
  2. Data curation: JR DS EYLT.
  3. Formal analysis: JR DS EYLT.
  4. Funding acquisition: DS.
  5. Investigation: JR DS EYLT.
  6. Methodology: JR DS EYLT.
  7. Project administration: JR DS.
  8. Resources: DS.
  9. Software: JR EYLT.
  10. Supervision: JR DS.
  11. Validation: JR EYLT.
  12. Visualization: JR.
  13. Writing – original draft: JR DS EYLT.
  14. Writing – review & editing: JR DS.

References

  1. 1. Mantegna RN, Stanley HE. An introduction to econophysics: correlation and complexity in finance. Cambridge, UK: Cambridge University. 2000; https://doi.org/10.1063/1.1341926
  2. 2. Plerou V, Gopikrishnan P, Rosenow B, Amaral LA, Stanley HE. Econophysics: financial time series from a statistical physics point of view. Physica A: Statistical Mechanics and its Applications. 2000;279(1):443–456.
  3. 3. Bouchaud JP, Potters M. Theory of financial risk and derivative pricing: from statistical physics to risk management. Cambridge university press; 2003. https://doi.org/10.1017/CBO9780511753893
  4. 4. Plerou V, Gopikrishnan P, Rosenow B, Amaral LAN, Stanley HE. Universal and nonuniversal properties of cross correlations in financial time series. Physical Review Letters. 1999;83(7):1471.
  5. 5. Laloux L, Cizeau P, Potters M, Bouchaud JP. Random matrix theory and financial correlations. International Journal of Theoretical and Applied Finance. 2000;3(03):391–397.
  6. 6. Fenn DJ, Porter MA, Williams S, McDonald M, Johnson NF, Jones NS. Temporal evolution of financial-market correlations. Physical review E. 2011;84(2):026109.
  7. 7. Sornette D, Ouillon G. Dragon-kings: mechanisms, statistical methods and empirical evidence. The European Physical Journal Special Topics. 2012;205(1):1–26.
  8. 8. Gualdi S, Tarzia M, Zamponi F, Bouchaud JP. Tipping points in macroeconomic agent-based models. Journal of Economic Dynamics and Control. 2015;50:29–61.
  9. 9. Mantegna RN. Hierarchical structure in financial markets. The European Physical Journal B-Condensed Matter and Complex Systems. 1999;11(1):193–197.
  10. 10. Tumminello M, Aste T, Di Matteo T, Mantegna RN. A tool for filtering information in complex systems. Proceedings of the National Academy of Sciences of the United States of America. 2005;102(30):10421–10426. pmid:16027373
  11. 11. Allen F, Babus A. Networks in finance. Wharton Financial Institutions Center Working Paper; 2008.
  12. 12. Haldane AG. Rethinking the financial network. In: Fragile stabilität–stabile fragilität. Springer; 2013. p. 243–278. https://doi.org/10.1007/978-3-658-02248-8_17
  13. 13. Strogatz SH. Exploring complex networks. Nature. 2001;410(6825):268–276. pmid:11258382
  14. 14. Elton EJ, Gruber MJ. Improved forecasting through the design of homogeneous groups. The Journal of Business. 1971;44(4):432–450.
  15. 15. Wilson KG. Renormalization group and critical phenomena. I. Renormalization group and the Kadanoff scaling picture. Physical review B. 1971;4(9):3174.
  16. 16. Sornette D. Discrete-scale invariance and complex dimensions. Physics reports. 1998;297(5):239–270.
  17. 17. Sornette D, Johansen A, et al. Significance of log-periodic precursors to financial crashes. Quantitative Finance. 2001;1(4):452–471.
  18. 18. Preis T, Schneider JJ, Stanley HE. Switching processes in financial markets. Proceedings of the National Academy of Sciences. 2011;108(19):7674–7678.
  19. 19. Dakos V, Scheffer M, van Nes EH, Brovkin V, Petoukhov V, Held H. Slowing down as an early warning signal for abrupt climate change. Proceedings of the National Academy of Sciences. 2008;105(38):14308–14312. pmid:18787119
  20. 20. Scheffer M, Bascompte J, Brock WA, Brovkin V, Carpenter SR, Dakos V, et al. Early-warning signals for critical transitions. Nature. 2009;461(7260):53–59. pmid:19727193
  21. 21. Scheffer M, Carpenter SR, Lenton TM, Bascompte J, Brock W, Dakos V, et al. Anticipating critical transitions. science. 2012;338(6105):344–348. pmid:23087241
  22. 22. Bonanno G, Caldarelli G, Lillo F, Mantegna RN. Topology of correlation-based minimal spanning trees in real and model markets. Physical Review E. 2003;68(4):046130.
  23. 23. Onnela JP, Chakraborti A, Kaski K, Kertesz J, Kanto A. Dynamics of market correlations: Taxonomy and portfolio analysis. Physical Review E. 2003;68(5):056110.
  24. 24. Bonanno G, Caldarelli G, Lillo F, Miccichè S, Vandewalle N, Mantegna RN. Networks of equities in financial markets. The European Physical Journal B-Condensed Matter and Complex Systems. 2004;38(2):363–371.
  25. 25. Tumminello M, Di Matteo T, Aste T, Mantegna R. Correlation based networks of equity returns sampled at different time horizons. The European Physical Journal B. 2007;55(2):209–217.
  26. 26. Granger CW. Some recent development in a concept of causality. Journal of econometrics. 1988;39(1):199–211.
  27. 27. Curme C, Tumminello M, Mantegna RN, Stanley HE, Kenett DY. Emergence of statistically validated financial intraday lead-lag relationships. Quantitative Finance. 2015;15(8):1375–1386.
  28. 28. Huth N, Abergel F. High frequency lead/lag relationships-Empirical facts. Journal of Empirical Finance. 2014;26:41–58.
  29. 29. Schreiber T. Measuring information transfer. Physical review letters. 2000;85(2):461. pmid:10991308
  30. 30. Fiedor P. Information-theoretic approach to lead-lag effect on financial markets. The European Physical Journal B. 2014;87(8):1–9.
  31. 31. Fiedor P. Granger-causal nonlinear financial networks. Journal of Network Theory in Finance. 2015;1(2):53–82.
  32. 32. Fama EF. The behavior of stock-market prices. The journal of Business. 1965;38(1):34–105.
  33. 33. Bouchaud JP, Farmer JD, Lillo F. How markets slowly digest changes in supply and demand. 2008;.
  34. 34. Fiedor P. Frequency effects on predictability of stock returns. In: 2014 IEEE Conference on Computational Intelligence for Financial Engineering & Economics (CIFEr). IEEE; 2014. p. 247–254.
  35. 35. Harré M, Bossomaier T. Phase-transition–like behaviour of information measures in financial markets. EPL (Europhysics Letters). 2009;87(1):18009.
  36. 36. Bossomaier T, Barnett L, Harré M. Information and phase transitions in socio-economic systems. Complex Adaptive Systems Modeling. 2013;1(1):1.
  37. 37. Reigneron PA, Allez R, Bouchaud JP. Principal regression analysis and the index leverage effect. Physica A: Statistical Mechanics and its Applications. 2011;390(17):3026–3035.
  38. 38. Preis T, Kenett DY, Stanley HE, Helbing D, Ben-Jacob E. Quantifying the behavior of stock correlations under market stress. Nature Scientific Report. 2012;2(752). pmid:23082242
  39. 39. Staniek M, Lehnertz K. Symbolic transfer entropy. Physical Review Letters. 2008;100(15):158101. pmid:18518155
  40. 40. Rubinov M, Sporns O. Complex network measures of brain connectivity: uses and interpretations. Neuroimage. 2010;52(3):1059–1069. pmid:19819337
  41. 41. Vicente R, Wibral M, Lindner M, Pipa G. Transfer entropy-a model-free measure of effective connectivity for the neurosciences. Journal of computational neuroscience. 2011;30(1):45–67. pmid:20706781
  42. 42. Borge-Holthoefer J, Perra N, Gonçalves B, González-Bailón S, Arenas A, Moreno Y, et al. The dynamics of information-driven coordination phenomena: A transfer entropy analysis. Science advances. 2016;2(4):e1501158. pmid:27051875
  43. 43. Souza TT, Aste T. A nonlinear impact: evidences of causal effects of social media on market prices. arXiv preprint arXiv:160104535. 2016;.
  44. 44. Marschinski R, Kantz H. Analysing the information flow between financial time series. The European Physical Journal B-Condensed Matter and Complex Systems. 2002;30(2):275–281.
  45. 45. Kwon O, Yang JS. Information flow between stock indices. EPL (Europhysics Letters). 2008;82(6):68003.
  46. 46. Yahoo! Finance website;. http://finance.yahoo.com.
  47. 47. Research Data Aston University website;. http://researchdata.aston.ac.uk/212/.
  48. 48. FTSE 100 History website;. http://www.ftse.com/products/downloads/FTSE_100_Constituent_history.pdf.
  49. 49. Theiler J, Prichard D. Constrained-realization Monte-Carlo method for hypothesis testing. Physica D: Nonlinear Phenomena. 1996;94(4):221–235.
  50. 50. Schreiber T. Constrained randomization of time series data. Physical Review Letters. 1998;80(10):2105.
  51. 51. Theiler J, Eubank S, Longtin A, Galdrikian B, Farmer JD. Testing for nonlinearity in time series: the method of surrogate data. Physica D: Nonlinear Phenomena. 1992;58(1-4):77–94.
  52. 52. Curty P, Marsili M. Phase coexistence in a forecasting game. Journal of Statistical Mechanics: Theory and Experiment. 2006;2006(03):P03013.