Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

High-frequency enhanced VaR: A robust univariate realized volatility model for diverse portfolios and market conditions

  • Wei Kuang

    Roles Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing

    kuangwei8355@hotmail.com

    Affiliation School of Sport Business, Guangzhou Sport University, Guangzhou, Guangdong, China

Abstract

In the field of financial risk management, the accuracy of portfolio Value-at-Risk (VaR) forecasts is of critical importance to both practitioners and academics. This study pioneers a comprehensive evaluation of a univariate model that leverages high-frequency intraday data to improve portfolio VaR forecasts, providing a novel contrast to both univariate and multivariate models based on daily data. Existing research has used such high-frequency-based univariate models for index portfolios, it has not adequately studied their robustness for portfolios with diverse risk profiles, particularly under changing market conditions, such as during crises. Our research fills this gap by proposing a refined univariate long-memory realized volatility model that incorporates realized variance and covariance metrics, eliminating the necessity for a parametric covariance matrix. This model captures the long-run dependencies inherent in the volatility process and provides a flexible alternative that can be paired with appropriate return innovation distributions for VaR estimation. Empirical analyses show that our methodology significantly outperforms traditional univariate and multivariate Generalized AutoRegressive Conditional Heteroskedasticity (GARCH) models in terms of forecasting accuracy while maintaining computational simplicity and ease of implementation. In particular, the inclusion of high-frequency data in univariate volatility models not only improves forecasting accuracy but also streamlines the complexity of portfolio risk assessment. This research extends the discourse between academic research and financial practice, highlighting the transformative impact of high-frequency data on risk management strategies within the financial sector.

Introduction

Value at Risk (VaR) has become one of the most widely used risk measures to control and manage market risk since its adoption by the Basel Committee on Banking Supervision in 1996. VaR is defined as the maximum loss that may be incurred by a trading portfolio, over a given time horizon with a specified probability. Effective VaR forecasts are critical not only for meeting regulatory capital requirements to maintain financial stability but also for making optimal capital allocation and investment decisions to promote financial vitality [1]. There has been a surge in VaR study since the 2008 financial crisis. While large-scale research has focused on developing VaR methodology for a single asset or index, far less attention has been dedicated to exploring its implications for portfolio VaR forecast. This is even though risks are managed on a portfolio basis by financial institutions, with diversification or tail contagion effects being considered. The banks’ minimum market risk capital requirements apply to the consolidated trading books of their financial entities [2].

To forecast the VaR of a portfolio, there are two approaches: 1) fitting a univariate volatility model to the portfolio returns; and 2) using a multivariate model to capture the dynamic covariances or correlations between asset returns. One significant advantage of the univariate model is that it does not require the modelling of asset correlations and other interdependencies. However, the univariate model’s reliance on portfolio weights presents a significant limitation [3]. The weights of the assets within the portfolio change significantly or when the portfolio is rebalanced. These changes can alter the portfolio’s risk profile, and the univariate model may not adequately adjust to these shifts without recalibration, leading to potential inaccuracies in VaR estimation. The question of which model is preferable for portfolio VaR forecasting given portfolio weights has attracted research interest. On the one hand, using a multivariate model to predict the combined dynamics of the assets in the portfolio might lead to better forecasting due to the usage of more information. The additional information, on the other hand, may be jeopardized by the greater uncertainty caused by the large number of parameters to be estimated in the multivariate volatility models [4]. [5] find that higher-dimensional information is irrelevant or even misleading in some situations for portfolio risk management.

[6] find that the benefits of using a multivariate model for forecasting portfolio volatility and VaR are insignificant. [5] suggest that VaR forecast based on the univariate GARCH model is at least as good as the forecasts from the multivariate counterparts within identical innovation distribution families. The complex multivariate models tend to overfit the data [7]. The relative performance of multivariate volatility in VaR forecasts depends more on the distributional assumptions than on the parametric specification of the volatility models [8, 9]. Indeed, a model with skew student t innovations outperforms those with symmetric distributions in both univariate and multivariate contexts [1012]. Moreover, capturing time-varying conditional correlations is also important [13]. Some studies further investigate the value of using Copula functions to allow for a flexible joint distribution for portfolio VaR forecasts [14, 15]. On the other hand, [16] finds mixed evidence about the VaR performance of univariate and multivariate models. [4] show that multivariate models outperform their univariate counterparts to forecast portfolio VaR for relatively large diversified portfolios.

Clearly, there is no conclusive evidence for the superiority of a univariate or multivariate volatility model. In addition, there are some limitations of previous research. First, most studies are based on a portfolio with a few assets, and the findings of any superior multivariate models are difficult to adapt to larger portfolios in practical applications. Second, the reliance on index data or the assumption of equally weighted portfolios in previous studies oversimplifies the intricate nature of individual portfolio risk profiles. Each portfolio’s risk profile is shaped by the specific assets it contains, their respective weights, and the interdependencies among them. By ignoring these nuanced factors, researchers may fail to recognize the significant influence that a portfolio’s unique risk profile has on the selection and performance of the most appropriate risk assessment model. Third, much of the existing analytical work has concentrated on comparing the efficacy of univariate and multivariate GARCH models. This focus has overlooked the potential advantages of integrating high-frequency intraday returns for portfolio VaR forecasting. The inclusion of such detailed frequency data can offer a more refined understanding of market fluctuations, potentially improving the precision of volatility estimates and, by extension, the accuracy of VaR forecasts.

The GARCH models introduced by [17, 18] are based on the daily returns. Despite being unbiased, squares and cross-products of daily returns are inefficient estimators of “integrated” volatility because they only employ one price measurement each period and so contain no information about the price trajectory between measurements. The realized volatility based on the intraday returns gives a more precise and quick-adapting estimate of current volatility than models based on daily data [1921]. Conditional realized volatility models are shown to outperform GARCH models in volatility forecasting [2023]. The improvement applies not just to univariate cases, but also multivariate extensions [2426]. In view of the superior statistical performance, the conditional realized volatility models have been explored in VaR forecasts. [27] were the first to compare the conditional realized volatility model with the GARCH models and find that they deliver equivalent results for VaR forecast. More recent research suggests that realized volatility improves VaR forecasting [2833]. However, these studies have primarily focused on a single asset or index VaR forecast leaving the significance of intraday data in portfolio VaR forecasting unexplored.

Given the ongoing debate on the efficacy of univariate versus multivariate models in portfolio VaR forecasting and the limitations of existing research, this study aims to address the research question: Can a refined univariate model, enhanced by high-frequency intraday data, provide a more accurate and computationally efficient alternative for portfolio VaR estimation across diverse market conditions and risk profiles? Specifically, we propose extending univariate conditional realized volatility models for portfolio variance estimation, which can then be used for VaR forecasting. The suggested method effectively harnesses realized volatility and correlation information across assets while not explicitly modelling the covariance matrix in a parametric manner. Therefore, it avoids forecasting entire covariance matrices using realized measures, which may suffer from the positive-definite restriction or curve-of-dimensionality problem. This is built on recent studies on utilizing realized volatility for single index VaR forecasting, but we intend to investigate this univariate modelling technique in a portfolio risk management scenario with more granular asset information. The proposed method is motivated by the findings of approximate log-normality and high persistence on realized volatilities [34]. The selection of univariate models draws on insights from [20] which show that for three exchange rates, univariate long-memory models can explain a considerable percentage of the volatility variation, with only modest incremental value from modelling by a vector autoregressive process. The results of this study could be generalized to other assets due to the remarkable similarities in realized volatility across asset and asset classes such as similar unconditional distributions, highly persistent auto-correlation, and strong co-movement [35]. Moreover, the ease of implementation holds promise for high-dimensional applications.

The data include daily returns and realized covariance matrices for a portfolio of ten Down Jones Industrial Average stocks from February 01, 2001 to December 31, 2009, provided by the [24] paper. We investigate the out-of-sample portfolio volatility and VaR forecast of the long memory realized volatility models which include AutoRegressive Fractionally Integrated Moving Average (ARFIMA), heterogeneous autoregressive-realized volatility (HAR-RV) [36], and one of its extension asymmetric HAR-RV [37]. As a benchmark, we report results for the univariate and multivariate GARCH models that explicitly allow for time-varying volatilities and correlations. The evaluation is based on a range of statistical and risk management criteria including VaR exception frequency, loss magnitude, and efficiency. The results show that the realized volatility models produce more informative and accurate portfolio conditional volatility forecasts, resulting in more accurate VaR predictions. The findings are robust against portfolio risk profiles and market conditions.

The remainder of the article is organized as follows. Section Theoretical background provides the theoretical framework for volatility and VaR forecasting. Section Models and methods presents the conditional volatility models and the innovation distributions assumptions. Section Data and Evaluation describes the data, estimation procedure, and evaluation criteria. Section Results and Discussion discuss the empirical results. Section Conclusion concludes.

Theoretical background

Denote the vector of returns of n assets in the portfolio at time t by Rt = (r1,t, …, rN,t)′ and portfolio return by where is the vector of portfolio weights. Portfolio VaR is defined as the maximum loss that can be experienced over a given time horizon with a certain confidence level θ, i.e. where α = 1 − θ and rp,t is the portfolio’s return over the period. The day ahead portfolio VaR is given by (1) where and are the portfolio conditional mean and conditional standard deviation forecasts at time t, respectively and is the inverse cdf of the standardized returns or innovations, i.e. . In order to compute the portfolio VaR, , and need to be estimated, which can be produced using two alternative conditional sets assessible at time t − 1: previous portfolio returns, i.e. or the full vector of past asset returns, i.e. Rth = (r1,th, …, rn,th)′. The former instance results in a univariate model for portfolio returns, while the latter results in a multivariate model [4].

Although conditional mean returns are difficult to forecast due to weak dependence, conditional volatility is both time-varying and highly predictable. Empirical research on conditional volatility modelling has exploded since the GARCH model was introduced by [17, 18] to capture the volatility clustering effect. See [3842], and among others for a survey of ARCH models including the multivariate extensions. The finding of time variation in conditional covariance and correlations leads to the development of multivariate GARCH models. For portfolio VaR forecast, conditional volatility of portfolio returns can be estimated by either: (1) fitting a univariate GARCH model to the portfolio returns; (2) using a multivariate GARCH volatility model to forecast the conditional variance of each asset in the portfolio, as well as the conditional correlations between assets returns [16].

Consider an N × 1 vector of continuous logarithmic prices, p(t) which follows the N-dimensional continuous-time diffusion, (2) where M(t) is an N × 1 instantaneous drift, Ω(t) denotes the N × N instantaneous diffusion matrix, and W(t) is an N × 1 dimensional vector of independent standard Brownian motions. The stochastic process governing the discretely observed N × 1 daily logarithmic return vector, Rt = p(t) − p(t − 1), is given by (3) where zt is an N × 1 vector of standard normally distributed, serially uncorrelated random variables and Ωt is an N × N integrated covariance matrix defined as (4)

The integrated covariance matrix given by Eq (4) is unobservable. An unbiased estimator of Ωt is given by the squared daily returns (5)

The estimator of the integrated covariance matrix based on daily returns forms the basis for a wide range of interdaily conditional volatility models such as the GARCH models. Despite being unbiased, the estimator using squared daily returns is inefficient because the signal that it contains about Ωt is dominated by noise [43]. [44] show that realized covariance, the sum of intraday cross products of returns, converges in probability to the quadratic covariation under mild regularity conditions. Let R(t, Δ) denote the n × 1 vector of logarithmic returns over the [t − Δ, t] time interval, R(t, Δ) = p(t) − p(t − Δ). The n × n realized covariance matrix for the unit time interval [t − 1, t] can be defined as (6) where Σt will converge to the integrated covariance matrix Ωt for finer and finer sampled high-frequency returns. Σt generally satisfies the positive definiteness condition as long as n < 1/Δ [20]. The accuracy of this approach, however, is limited by the market microstructure effects which distort the measurement of returns at high frequencies, making measured returns no longer satisfy the regularity conditions required for the consistency properties of realized volatility. Returns measured at intervals of five to thirty minutes are shown to provide the optimal balance of precision and market microstructure frictions [20, 34].

[19] show that the realized variance and covariance of stock returns are approximately log-normally distributed and highly persistent, which are well represented by long-memory processes. This implies that the realized volatility of a portfolio which is a linear combination of asset returns may have a similar attribute, namely approximating a log-normal distribution with a long memory. [20] find similar results for currencies and their cross rate, and suggest using a fractionally integrated model to forecast each element of the realized covariance matrix or its Cholesky decomposition. This is because the continuously compounded return on the cross rate should be equal to the difference between the two pairs of currency returns in the absence of triangle arbitrage.

Rather than forecasting Σt, this paper proposes to extend long memory realized volatility models to forecast portfolio variance, which can directly feed into VaR forecasts. Specifically, given the vector of portfolio weights Wt, the portfolio realized variance can be specified as (7) where RVp.t is a linear combination of individual elements of Σt. For example, the realized variance of a bivariate portfolio can be calculated as (8) where wi,t is the weight for asset i at time t; σij,t is the realized variance of an individual asset for (i = j) and realized covariance between assets for (ij). This is essentially the realized volatility based on the aggregated returns of all the assets in the portfolios. The proposed method takes into account: 1) the complexity of forecasting realized covariance matrix Σt with multivariate modelling constraint; and 2) the long-run dependency of realized volatility with univariate modelling flexibility.

There are three challenges to incorporating realized measures into multivariate volatility models for tail risk forecasting. First, the conditional realized volatility models must be parameterized in such a way that the forecasted covariance matrices are positive definite in high-dimensional settings. Second, the multivariate volatility models, particularly those based on the GARCH specification have a large number of parameters, resulting in the curse-of-dimensionality problem. Third, the conditional distributions of innovations are analytically tractable only when the return distribution is closed under linear transformation, i.e. the linear combinations of ri,t have the same distribution as the marginal distribution of returns ri,t [3]. The proposed method, however, shares the benefit of univariate modelling in that it is parsimonious and allows for flexible distribution to capture skewness and excess kurtosis in the innovation process without adding significant computational costs. The details of the models are discussed in the following section.

Models and methods

The conditional mean return is assumed to be constant. To forecast portfolio variance, three intradaily volatility models are considered, including the ARFIMA model suggested by [19, 20] for realized volatility modelling, the HAR-RV model [36] based on the Heterogeneous Market Hypothesis of [45], and one of its extension asymmetric HAR-RV model [37] to allow for leverage effect. The conventional HAR-RV model is well known for capturing the long memory features observed in financial time series data [36]. Furthermore, the inclusion of asymmetry extensions in the HAR-RV model is consistent with empirical evidence suggesting that volatility is frequently subject to the leverage effect, in which negative market movements generate larger changes in volatility than positive market movements of the same magnitude. While alternative models, such as those incorporating jumps certainly offer additional lenses through which to analyze volatility, our objective was to strike a balance between model complexity and interpretability. The model description is provided in section. Some representative GARCH-type models are included for comparison. The univariate GARCH models include the standard GARCH [18] to capture volatility clustering, as well as various extensions to address leverage effects such as the Glosten-Jagannathan-Runkle (GJR)GARCH [46] and the Asymmetric Power Autoregressive Conditional Heteroskedasticity (APARCH) [47], and long memory property such as the Fractionally Integrated Generalized Autoregressive Conditional Heteroskedasticity (FIGARCH) [48]. The multivariate GARCH models include the RiskMetrics Exponentially Weighted Moving Average (EWMA) [49] and the Dynamic Conditional Correlation (DCC) model [50]. Because these GARCH models are standard, their specifications are presented in Table 1 for reference.

Two innovation distributions are considered in univariate models: the standard normal and the skewed student distribution [51] that addresses the skewness and excess kurtosis commonly observed for financial return data [10, 52, 53]. We also include the filter historical simulation (FHS) approach [54] which combines the standard GARCH with the empirical distribution of innovation for VaR forecast. For the multivariate models, we follow [4] to consider two alternative multivariate distributions: the multivariate normal distribution for the EWMA model and the multivariate student t distribution for the DCC model. We intend to keep the innovation distribution tractable so that it can be applied to higher-dimensional settings without raising estimation issues or computation costs. Some more sophisticated distributions such as the multivariate skewed Student t distribution of [11] do not have an analytical formula for computing VaR using conditional volatilities which require Monte-Carlo simulation. The details of the innovation distributions are covered in section.

Volatility models

ARFIMA model.

The long memory property of realized volatility is captured by the fractionally differencing parameter d. (9) where is the logarithm of the daily portfolio realized variance defined in Eq (7).

HAR-RV model.

The persistence of realized volatility is approximated by an autoregressive structure using daily, weekly, and monthly volatility components. (10) where is the logarithm of the daily portfolio realized variance defined in Eq (7); with h = w = 5 and h = m = 22 being the weekly and monthly components respectively.

Asymmetric HAR-RV model.

The model contains additional lagged (absolute) standardized returns occurring at different time horizons: daily, weekly, and monthly. (11) where are the daily (d), weekly (w), and monthly (m) standardized returns.

The heteroscedasticity in the residuals is expected to persist due to the variance of the realized volatility estimator [55]. Therefore, the GARCH(1,1) with a normally distributed innovation is employed to account for the conditional heteroscedasticity of the residuals in the conditional realized volatility models. (12) (13)

By definition when . Therefore, the conditional realized volatility can be computed as (14) where denotes the estimated value of ut by the respective realized volatility models and is the estimated variance defined in Eq (13).

Innovation distribution

Normal distribution.

If the innovation follows a standard normal distribution, i.e. zti.i.dN(0, 1), there is no additional parameter to be estimated. The day-ahead VaR forecast is given by (15) where Φ−1(α) is the α quantile of the standard normal distribution.

Skewed student distribution.

The quantile function is derived by [56] as follows (16) where is the αth quantile of the unit variance skewed student distribution with ν > 2 degrees of freedom and asymmetric parameter ξ > 0; denotes the quantile function of the standardized Student-t density function; and are the mean and standard deviation of the non-standardized skewed student distribution, respectively. The day-ahead VaR forecast is then given by (17)

Filtered historical simulation.

The filtered historical simulation (FHS) introduced by [54] makes no distribution assumptions regarding the innovation. It employs innovation to develop hypothetical future possibilities, which can be accomplished by resampling or bootstrapping. We use the empirical distribution function (EDF) with the historical cumulative distribution of standardized returns in this paper. As a result, the α quantile is based on the empirical data presented below. (18)

The day-ahead VaR forecast is then given by (19)

Multivariate distribution.

The portfolio conditional variance is denoted by where is the conditional covariance matrix forecast. The day-ahead VaR forecast under the multivariate normal distribution assumption is given by (20) where zα is the α quantile of the normal distribution. Under the assumption of multivariate Student t distribution, the zα is replaced by tα,ν where ν is the degrees of freedom.

Data and evaluation

Data and estimation

To build an equity portfolio, we use the data of 10 Dow Jones Industrial Average (DJIA) stocks from the [24] paper. The data are available from the website of Oxford Man Institute of Quantitative Finance. The underlying high-frequency data came from the Trade and Quote database of the New York Stock Exchange. The realized covariance matrix was constructed using intraday returns sampled at a 10-minute frequency. The data set includes open-to-close and close-to-close daily returns, as well as the vech of the 10 × 10 realized covariance matrix for the following stocks: Bank of America (BAC), JP Morgan (JPM), International Business Machines (IBM), Microsoft (MSFT), Exxon Mobil (XOM), Alcoa (AA), American Express (AXP), Du Pont (DD), General Electric(GE) and Coca Cola (KO) from February 01, 2001 to December 31, 2009 for a total of 2,242 observations. We use open-to-close daily returns for all interdaily, i.e. GARCH models, given that the realized covariance matrix does not account for overnight returns. The sample is divided into an initial estimation period of 1,000 observations from February 01, 2001 to January 26, 2005 and 1,242 out-of-sample observations from January 27, 2005 to December 31, 2009. The out-of-sample period covers calm and turbulent times, including the 2008 financial crisis. The models are initialized using the estimating period and the volatility and VaR forecasts are created for Day 1,001. After that, the estimation window is then advanced by one day. The models are re-evaluated and used to generate forecasts for Day 1,002, and so on until the sample is completed. The key advantage of using a rolling window of 1000 observations is that it combines the most recent market data while discarding out-of-date observations.

Table 2 reports the summary statistics for each return series as well as the equally weighted portfolio return and its logarithmic realized standard deviation. The mean returns on individual stocks and the portfolio are close to zero. All returns show skewness and excess kurtosis with the normality assumption being rejected in all cases. The ARCH-LM test of [17] for up to twentieth-order serial correlation in squared returns reveals evidence of considerable volatility clustering. The decision to consider autocorrelation and conditional heteroskedasticity of order up to 20 was based on a comprehensive review of the literature [27, 57, 58], which allows for the detection of both short-term and moderately long-term dependencies in the data. The logarithmic realized standard deviation (LRV), on the other hand, follows more closely to the normal distribution but has a strong serial correlation and ARCH effect. Moreover, it should be noted that the Ljung-Box statistics of LRV is an order of magnitude larger than the ARCH-LM statistics of daily squared returns. This could be explained by the fact that daily squared returns are relatively noisy volatility proxies when compared to the daily realized volatilities [43]. The noise masks the significant persistence in the underlying volatility dynamics [20]. The upper panel of Fig 1 provides graphical illustration of the sample autocorrelation of the portfolio realized logarithmic standard deviation out to over 30 days. The autocorrelation is significantly above the 95% confidence band and decays slowly, indicating a long-memory mechanism. The lower panel compares the one-day ahead volatility forecast of HAR-RV model and the DCC model against the realized volatility over the out-of-sample period. Realized portfolio volatility fluctuates dramatically over time. The HAR-RV model generates volatility forecasts that respond to changes in the underlying unobserved volatility levels faster than the DCC model, although the DCC model captures the volatility and correlation dynamics using stock daily returns.

thumbnail
Fig 1. Realized volatility.

Notes: The upper graph shows the autocorrelation of portfolio logarithm realized standard deviation over the full sample period from February 01, 2001 to December 31, 2009 (2,242 observations). The lower graph shows the portfolio realized standard deviation against forecasts from HAR-RV and DCC models over the out-of-sample period from January 27, 2005 to December 31, 2009 (1,242 observations).

https://doi.org/10.1371/journal.pone.0303962.g001

The estimation results of the univariate ARMA(0,0)-GARCH models using daily returns and the ARFIMA and HAR-RV type models using intraday data are presented in Tables 3 and 4, respectively. The estimation is based on the entire sample period for illustration. We first examine Table 3. The conditional mean returns μ are not statistically significantly different from zero in all the models. The parameter β of all GARCH models is close to 1, indicating the persistence of volatility. The FIGARCH model’s fractional differencing parameter d shows a moderate degree of long memory. The GJRGARCH model’s leverage parameter γ is positive and significant, suggesting a leverage effect. Moreover, it should be noted that GARCH filtering removes the time-series dynamics from the return series but not all of the excess kurtosis. The estimates for the skewed student distribution parameters (ξ and ν) indicate negative skewness and fat tails. We then proceed to Table 4. The fractional differencing parameter d in the ARFIMA model, as well as the highly significant and positive αd, αw, and αm in the HAR-RV and the Asym.HAR-RV model demonstrates the persistence of the realized volatility series. The leverage parameters ϑ(d), ϑ(w), ϑ(m) are all negative and statistically significant, indicating that daily, weekly and monthly shocks all increase volatility. Positive and statistically significant estimates for γ(d), γ(w), and γ(m) suggest heterogeneous size effects. The insignificant ARCH-LM statistics show that the GARCH(1,1) process removes the “volatility of volatility” clustering. Furthermore, the diagnostic statistics of standardized return residuals indicate volatility dynamics have been reduced to a substantial extent with minimal autocorrelation. However, the standard residuals still exhibit significant non-normality.

Forecast evaluation

The relative performances of the models are assessed using both statistical and economic criteria. Although there are many alternative backtesting approaches available, we focus on a sample selection of classic backtesting processes that should provide sufficient evidence for the paper’s results. The statistical criteria examine the accuracy of portfolio volatility forecasting using Root Mean Square Error (RMSE) and Mean Absolute Error (MAE) metrics, as well as the efficiency and information contents using the Mincer-Zarnowitz regression. The portfolio realized volatility serves as the evaluation benchmark. The risk management criteria examine the VaR forecast performance in terms of exception frequency (unconditional and conditional coverage tests), exception magnitude (Berkowitz test), joint frequency and magnitude (tick loss and quadratic loss), and VaR efficiency (based on mean and standard deviations).

Root Mean Square Error (RMSE).

(21)

Mean Absolute Error (MAE).

(22)

Mincer-Zarnowitz (MZ) Regression.

(23)

A forecast is conditionally unbiased (i.e., weak-form efficient) if and only if α = 0 and β = 1 in the MZ regression. The R-squared coefficient indicates the explanatory power of each of the model’s forecasts, irrespective of any bias or inefficiency.

Unconditional coverage.

The unconditional coverage test of [59] examines the null hypothesis that the exception rate is statistically equal to the expected value given the confidence level. The test statistic follows an asymptotic χ2 distribution with one degree of freedom (24) where n1 represents the number of exceptions, n0 represents the number of non-exception, is the estimated proportion of exceptions, and α = 1 − p is the probability of an exception for a given confidence level p.

Conditional coverage.

The conditional coverage test LRcc of [60] jointly examines whether total exceptions are equal to the expected and whether exceptions are distributed independently. The statistic follows an asymptotic χ2 distribution with two degrees of freedom. (25) where πij denotes the number of times that state j follows state i for i, j = 0, 1 and ; state 0 denotes that no exception of VaR forecast occurs and state 1 denotes that exception occurs.

Berkowitz test.

The tail test of [61] examines whether the magnitude of observed VaR exceptions is consistent with the underlying VaR model. Suppose rp,t is the ex-pose portfolio returns and is the ex-ante forecasted loss density with cdf of . is i.i.d and distributed uniformly on (0, 1) according to [62]. The empirical quantile series can be transformed into a standard normal series using the inverse Normal cdf, i.e. . Then zt is treated as a censored normal random variable with the censoring tied to the target confidence level of the VaR estimates. Specifically, zt is transformed into as follows (26)

If the empirical quantiles are generated correctly by the VaR model, the zt series should be identically distributed with zero unconditional mean μ and unit unconditional standard deviation σ. The log-likelihood function is used to estimate μ and σ jointly (27)

The relevant LR statistic is defined as (28)

Tick loss.

The tick loss function (TLF) of [63] is an asymmetric loss function that penalizes observations that have a VaR exception more heavily. (29) where rp,t is the realized portfolio return on day t and It takes a value of one if exceptions occur and zero otherwise.

Quadratic loss.

The quadratic loss function (QLF) of [64] penalizes a large loss more than a smaller loss at exceptions using a quadratic term. (30) where rp,t is the realized portfolio return on day t.

Average VaR.

The TLF and QLF examine the cost of capital deficit to cover a loss greater than the portfolio’s predicted VaR while ignoring the opportunity cost of keeping superfluous capital when the exception does not occur. Therefore, we follow [65] to assess the efficiency of a VaR model by utilizing the average predicted portfolio VaR. (31)

Standard deviation of VaR.

The lower the standard deviation of predicted VaR, the less uncertain the amount of capital required to cover against unexpected portfolio losses. The final measure is the standard deviation of the portfolio predicted VaR proposed by [65]. (32)

1 Results and discussion

1.1 Statistical evaluation

We first perform a statistical assessment of portfolio volatility forecast. Table 5 represents the Mincer-Zarnowitz regression results, as well as the RMSE and MAE of competing volatility models over the out-of-sample period. The assumption of innovation distribution has no substantial impact on the performance of the univariate GARCH models. The coefficient estimates of the conditional realized volatility models are closer to zero for the intercept and near unity for the slope with a higher R2 when compared to the univariate GARCH models. The non-significant intercepts suggest that there is no systematic bias in the volatility forecasts over the out-of-sample period, indicating that the models are well-calibrated and do not consistently over- or under-predict the volatility levels. It can been that the forecasts produced by intradaily volatility models are less biased, more efficient, and more informative. Moreover, their RMSE and MAE are lower than those produced by the GARCH models. Specifically, the ARFIMA model performs the best, followed by the ASYHAR and then the HAR models. Their performance, however, is generally indistinguishable. It is interesting to note that the DCC model outperforms all univariate GARCH models and EWMA with the weak-form efficient hypothesis not being rejected at the 5% significance level. However, the forecast of the DCC model is neither more informative nor more accurate than the forecasts produced by the intradaily volatility models.

Next, the fluctuation test of [66] is employed to provide insight into the models’ relative forecast accuracy over time. Fig 2 displays a sequence of differences between the MSE of the standard HAR-RV model and the univariate GARCH models with skewed student innovations over a rolling window of 124 observations (approximately 10% of the total out-of-sample observations). It also depicts the two-sided 95% critical values for testing the null hypothesis that the two models have equal out-of-sample performance at each point in time. Negative (positive) values indicate that the first model in the pair of comparisons produces better volatility forecast than the second model, i.e. the standard HAR-RV model. It is noted that all the GARCH models are less accurate than the HAR-RV model, with the upper 95% critical values being exceeded at least one point in time. The difference between ARFIMA (or ASYHAR-RV) and the standard HAR-RV is insignificant, although the ASYHAR-RV forecast appears to be more accurate. Fig 3 compares the univariate models to the multivariate EWMA and the DCC models. The relative performance of the univariate and multivariate EWMA or GARCH models varies over time, with neither consistently surpassing the other. The HAR-RV model, however, consistently delivers more accurate forecasts than the EWMA and the DCC, with the lower 95% critical values crossing frequently. Overall, the fluctuation test results show that when integrated with intraday information, the univariate models give more accurate volatility forecasts than both the univariate and multivariate GARCH models.

thumbnail
Fig 2. Fluctuation test I.

Notes: The graphs show the fluctuation test statistics with the 95% critical values for the MSE differences between competing conditional volatility models. The MSE statistics are calculated using rolling windows of 124 observations over the out-of-sample evaluation period from January 27, 2005 to December 31, 2009 (1,242 observations).

https://doi.org/10.1371/journal.pone.0303962.g002

thumbnail
Fig 3. Fluctuation test II.

Notes: The graphs show the fluctuation test statistics with the 95% critical values for the MSE differences between competing conditional volatility models. The MSE statistics are calculated using rolling windows of 124 observations over the out-of-sample evaluation period from January 27, 2005 to December 31, 2009 (1,242 observations).

https://doi.org/10.1371/journal.pone.0303962.g003

1.2 VaR evaluation

The VaR forecasts are assessed at four confidence levels p ∈ {0.95, 0.975, 0.99, 0.995} for both long and short trading positions. These levels have been chosen to provide a comprehensive evaluation of the models across a range of risk thresholds that are commonly used in financial risk management. Specifically, the 95% level is frequently used for standard internal risk management. It provides insight into the moderate tail risks that financial institutions might face under normal market conditions. The 99% level is a critical threshold commonly used by regulators, including the Basel Committee on Banking Supervision, for calculating market risk and determining capital requirements. It reflects significant financial stress but not the most extreme conditions. By extending the evaluation to include confidence levels of 97.5%, and 99.5%, we aim to assess the model’s performance under various extreme market scenarios and for investors with different risk appetites. This multi-threshold approach allows a more detailed and nuanced understanding of the models’ performance in predicting tail risks, which is essential for both regulatory compliance and strategic risk management.

Tables 6 and 7 show results of the unconditional and conditional coverage tests, respectively, which are qualitatively similar. First, for long positions, volatility models with normally distributed innovations fail both tests across all VaR confidence levels. This implies that the conditional normal distribution assumption is not supported by empirical evidence. When using a skewed student distribution, performance improves dramatically. For short positions, all the volatility models pass both tests with varying degrees of accuracy and the GARCH models with skewed student innovation performs better than their normally distributed counterparts at the high quantiles (i.e. 0.99 and 0.995).

If we focus on the long positions and models with non-normal distributed innovations, the DCC-T model outperforms all the univariate GARCH models since it passes both tests across all quantiles. Among the univariate GARCH models, the FHS performs relatively well at the lower quantiles, but fails at the high quantiles. The GJRGARCH-ST, APARCH-ST and FIGARCH-ST models are the opposite. This suggests that GARCH models with leverage or long memory improve VaR forecasts at high confidence levels, but are less accurate at lower confidence levels. It is worth noting that the intradaily models enhance VaR forecasts across all confidences levels when compared to the standard GARCH model. The HAR-ST model, in particular, does not reject the null hypothesis of correct unconditional and conditional coverage at the 5% significance level for all VaR quantiles.

Table 8 shows the results of the Berkowitz test for exception magnitudes which is generally consistent with the frequency tests. For example, volatility models with normally distributed innovations are rejected across all left quantiles, whereas some of them such as GARCH, GJRGARCH, and APARCH, are rejected across right quantiles. These volatility models perform better when combined with a skewed student distribution. Furthermore, the GJRGARCH and APARCH models improve standard GARCH’s accuracy at the 99% and 99.5% VaR confidence level for the long position, but remain inaccurate at lower quantiles. It is interesting to note that, despite passing the frequency tests, the DCC-T model fails the Berktowitz test at lower left quantiles. Several factors may contribute to the underperformance of the DCC-T model in this context. First, the DCC-T model’s reliance on the Student’s t-distribution assumption for the innovation distribution may not be sufficient to accurately forecast the extreme market movements, as these events often exhibit both skewness and heavy tails. Moreover, the complexity of the DCC-T model, with its additional parameters for estimating dynamic conditional correlations, poses potential estimation challenges. During periods of extreme market movements, the model may struggle to accurately update its correlation estimates, resulting in less reliable VaR forecasts at the lower quantiles where accurate prediction of extreme losses is critical. On the other hand, the intradaily models with skewed student distributions pass the tests across all left and right quantiles. Therefore, these results, when combined with the frequency tests, suggest that intradaily volatility models outperform GARCH models in terms of VaR forecast performance.

To further investigate the relative performance of GARCH versus intradaily models, we now focus on the loss functions which take the exception frequency and magnitude into account simultaneously. All univariate models with normally distributed innovations are excluded from the comparison, because they are less accurate in VaR forecasts than those with skewed student distributions, particularly for the long position. The HAR-ST model is used as a benchmark for comparison since it provides adequate VaR forecasts with the simplest specification among the intradaily models. Tables 9 and 10 report [67] test statistics with p-values. The null hypothesis is that the mean difference between each model and the HAR-ST model in tick loss or quadratic loss function values is equal to zero. Positive (negative) statistics indicate that the corresponding models have a bigger (smaller) loss, meaning that they are less (more) accurate than the HAR-ST model. It can be seen that the statistics for all GARCH models are generally positive across all quantiles. The null hypothesis of identical tick loss values is rejected in 64% and 80% of 56 cases (across 7 GARCH models, 4 confidence levels for both long and short positions), respectively. Similarly, at the 5% and 10% significance levels, the null hypothesis of equal Lopez loss value is rejected in 56% and 70% of the cases, respectively. The ARFIMA-ST and ASYHAR-ST models contain some negative tick loss values, but they are generally insignificant at the conventional significance levels. The results suggest that intradaily volatility models are more accurate at forecasting VaR than both univariate and multivariate GARCH models. There are no statistically significant differences in performance among intradaily models.

Finally, the mean and standard deviation of VaR estimates are calculated with results reported in Table 11. For the long position, intradaily models have marginally higher average VaR values but lower standard deviation than GARCH models, but the difference is less than 1% for the mean and within 2% for the standard deviation. For the short position, the intradaily models show lower mean and smaller standard deviations across all confidence levels, with a difference of about 1% for the mean and 3% for the standard deviation. Overall, the results show that intradaily models reduce the variability of VaR estimations while improving accuracy without increasing VaR levels significantly.

Robustness tests

There are two main limitations of the above analysis. First, the study is based on an equally weighted portfolio, however in reality portfolio managers may hold a variety of allocations based on their target returns, risk tolerance, and market view. As a result, it would be worthwhile to investigate if the outcomes are sensitive to the changes in portfolio risk profiles. Second, the out-of-sample period includes the 2008 financial crisis. [68] suggest that a Global Financial Crisis (GFC)-robust approach should not be overly sensitive to the period chosen for analysis so that risk managers do not have to continually adjust the rules for creating VaR forecasts when moving between tranquil and turbulent periods. Therefore, it would be interesting to examine whether the relative performance of each model remains stable throughout market conditions. In this part, we perform robustness tests to address the two constraints, respectively.

Portfolio allocations.

The pie chart of Fig 4 shows the covariance risk contribution of individual stock in the equally weighted portfolio ranges from 4.7% to 14.5%. We classify the entire stock population into two groups according to their risk contribution. The first category is a HighRisk portfolio, which is composed of an equal weighting of the top five risky stocks (i.e. BAC, JPM, AXP, AA, and GE). The second group is a LowRisk portfolio comprised of the remaining stocks (i.e. IBM, MSFT, XOM, DD, and KO) that are also equally weighted. The density graphs in the upper panel show that the HighRisk portfolio is more leptokurtic than the LowRisk portfolio. In order to fully assess the impact of different risk exposures on portfolio performance, we have systematically mixed the HighRisk and LowRisk portfolios in specific proportions. The mix ratios of 90%, 70%, 30%, and 10% for the HighRisk portfolio, balanced by 10%, 30%, 70%, and 90% for the LowRisk portfolio, were chosen to represent a wide range of risk scenarios, from predominantly high-risk to predominantly low-risk exposures. This selection was strategically designed to illuminate the effects of risk concentration and diversification within a portfolio. By incrementally adjusting the proportion of HighRisk and LowRisk stocks, we can observe the corresponding changes in the overall risk profile and performance metrics of the portfolio. These specific intervals have been chosen to ensure that the transitions between different levels of risk exposure are distinct and meaningful, allowing for a clear analysis of how incremental shifts in risk allocation affect portfolio performance. The boxplot displays the changes in portfolio distribution as a result of changes in allocation between HighRisk and LowRisk groups.

thumbnail
Fig 4. Portfolio distribution.

Notes: The pie chart shows the covariance risk contribution of each stock in an equally-weighted portfolio. The HighRisk portfolio is made up of the top 5 risky stocks, which are equally weighted. The LowRisk portfolio consists of the remaining 5 stocks, which are also equally weighted. Their return densities are shown in the upper panel. The bottom right graph (from top to bottom) shows the return distribution of the HighRisk portfolio mixed with 0%, 10%, 30%, 70%, 90% and 100% of LowRisk portfolio.

https://doi.org/10.1371/journal.pone.0303962.g004

Fig 5 compares the VaR performance of HAR-ST and GARCH-ST with varying allocation weightings to HighRisk and LowRisk portfolios. We examine representative interdaily and intradaily models rather than entire model set to reduce the amount of estimation effort for sensitivity analysis. The p-values of the Berkowitz, unconditional, and conditional coverage tests are presented. It is worth noting that both models are less accurate at the VaR 95% confidence level. The HAR-ST consistently outperforms the GARCH-ST across all confidence levels and portfolio risk profiles. The p-values of the tests fluctuate as the allocation changes. However, the results for unconditional and conditional coverage tests are fairly consistent, and there appears to be little evidence that the HAR-ST model’s performance is deteriorating for scenarios with very low or high portfolio risk. Fig 6 compares the performance of the multivariate EWMA and the DCC-T. The EWMA model fails the Berkowitz test regardless of the portfolio risk profile. Moreover, the unconditional and conditional coverage tests can only be passed for a very low risk profile at the VaR 95% confidence level. The DCC-T model generally performs poorly in the Berkowitz test and its unconditional and conditional coverage test performance deteriorates as the portfolio’s risk increases. Overall, the sensitivity analysis shows that the superior performance of HAR-ST model is robust against the changes in the portfolio’s risk profiles.

thumbnail
Fig 5. Portfolio allocation: Univariate models.

Notes: The graphs show the p-values of Berkowitz, unconditional and conditional coverage tests for the VaR models (HAR-ST and GARCH-ST) under four confidence levels, i.e. 95%, 97.5%, 99%, and 99.5%. The portfolio includes long positions in 10 DJIA stocks with allocations to the HighRisk portfolio ranging from 10%, 30%, 50%, 70%, and 90% with the remainder allocated to the LowRisk portfolio. The out-of-sample evaluation period is from January 27, 2005 to December 31, 2009 (1,242 observations).

https://doi.org/10.1371/journal.pone.0303962.g005

thumbnail
Fig 6. Portfolio allocation: Multivariate models.

Notes: The graphs show the p-values of Berkowitz, unconditional and conditional coverage tests for the VaR models (EWMA and DCC-T) under four confidence levels, i.e. 95%, 97.5%, 99%, and 99.5%. The portfolio includes long positions in 10 DJIA stocks with allocations to the HighRisk portfolio ranging from 10%, 30%, 50%, 70%, and 90% with the remainder allocated to the LowRisk portfolio. The out-of-sample evaluation period is from January 27, 2005 to December 31, 2009 (1,242 observations).

https://doi.org/10.1371/journal.pone.0303962.g006

Sub-periods.

We divide the entire out-of-sample evaluation period into two sub-periods. The pre-GFC starts from January 27, 2005 to August 29, 2008. The over-GFC period covers from September 02, 2008 to December 31, 2009. The left panel of Fig 7 displays the number of VaR backtesting exceptions before and during GFC. The expected number of exceptions is shown at the top of each graph as a reference. The right panel shows the mean absolute deviation of loss (MeanAD) throughout the full out-of-sample period and the maximum absolute deviation of loss (MaxAD) over the two sub-periods. The MeanAD (or MaxAD) measures the average (or maximum) absolute deviation of the violations from the VaR forecasts [16]. For each competing model, the upper bar is based on the results for VaR at the 99% confidence level while the lower bar is based on the 95% confidence level. The regulatory VaR backtesting is based on the comparison between one-day VaR at the 99% confidence level and the realized profit and loss [69].

thumbnail
Fig 7. Sub-periods analysis.

Notes: The graphs show the number of exceptions, mean absolute deviation (MeanAD), and maximum absolute deviation (MaxAD) over different sample periods. The full out-of-sample evaluation period is from January 27, 2005 to December 31, 2009. The Pre global financial crisis (GFC) is from January 27, 2005 to August 29, 2008. The over GFC period is from September 02, 2008 to December 31, 2009. Under each model, the upper bar is for VaR99 while the lower bar is for VaR95. The expected number of exceptions is shown at the top of each graph on the left.

https://doi.org/10.1371/journal.pone.0303962.g007

It can be seen that the univariate GARCH models produce more exceptions than expected for both confidence levels while the intradaily models improve the VaR performance. The results are generally consistent before and during GFC. For example, the FHS forecast is more accurate at 95% confidence but less so at 99% confidence. The GJRGARCH and APARCH models enhance the VaR forecasts at 99% confidence to some extent, but still generate more exceptions than expected. The DCC-T model produces the number of exceptions that are closest to the expected level at the 95% confidence before the GFC, but its MeanAD is substantially larger than the other models throughout the entire period, which is consistent with the Berkowitz test result. The intradaily models appear to be conservative at the 99% confidence level. The HAR-ST model, in particular, produces the smallest MaxAD before and during GFC. Therefore, the intradaily models would be favoured by financial institutions pursuing sufficient capital or regulators seeking financial system stability.

Discussion

This paper illuminates the advantages of leveraging intraday data for more precise portfolio variance and tail risk forecasting. Our findings are in line with recent studies that have highlighted the importance of capturing the nuances of market behaviour through high-frequency data. For instance, recent studies [70, 71] have demonstrated the effectiveness of high-frequency data in improving the accuracy of volatility forecasts for stock markets. Similarly, recent studies such as [32, 33, 72] have shown that incorporating high-frequency data can lead to more robust VaR estimates by better capturing the intraday dynamics of financial markets. However, our research extends the previous work in the following ways:

First, we critically address a gap in the existing literature by questioning the widespread belief in the superiority of multivariate GARCH models over their univariate counterparts. Previous studies, such as [5, 6], have shown that multivariate GARCH models, despite their complexity, do not significantly improve volatility forecasting. We show that univariate realized volatility models, particularly when augmented with high-frequency intraday data, not only outperform conventional multivariate models in terms of efficiency and predictive accuracy but also exhibit remarkable capability in high-dimensional settings. This challenges the notion of multivariate complexity as a proxy for predictive ability by demonstrating that intraday data mitigates the noise inherent in daily returns, which often obscures the true volatility signal [73]. The findings highlight the under-appreciated potential of univariate models as robust tools for both risk management and other applications dependent on volatility forecasting, such as derivative asset pricing.

Second, we extend the work of [20, 36] by investigating the distributions of realized volatilities and their long-memory properties. We unravel the persistence of realized volatility within equity portfolios and show that its structure can be captured well using univariate ARFIMA or HAR-RV models. It has been found that the efficacy of these models is broadly equivalent in predicting the realized volatility of portfolios, a finding consistent with [36] research within single indices. Given the striking similarities in realized volatility patterns within and across asset classes [35], the analysis could be expanded in a larger-dimensional setting or other asset classes as future research.

Third, our research represents a significant methodological breakthrough by extending the application of univariate realized volatility modelling from single assets or indices to the complex scenario of portfolio VaR forecasting. This novel analysis goes beyong previous research that has focused narrowly on volatility timing strategies within asset allocation [35, 74, 75]. By demonstrating that improved volatility risk forecasts lead to more accurate predictions of portfolio tail risks, our work identifies the crucial role of adjusting the innovation distribution to account for skewness and excess kurtosis. Significantly, we find that univariate models offer a superior balance between the accurate representation of the innovation distribution and computational efficiency compared to their multivariate counterparts.

Last but not least, our study addresses and resolves a critical issue identified by [16] regarding the tendency of both univariate and multivariate GARCH models to produce either excessive or insufficient violation counts. We show that the use of univariate models calibrated with intraday volatility data significantly refines VaR forecasting performance and overcomes the shortcomings of multivariate DCC-T models, as demonstrated by the work of [4]. While the DCC-T model may be suitable for diversified portfolios in general, our analysis reveals that it does not fully capture the nuanced characteristics of riskier asset profiles within our specific dataset. This limitation leads to less accurate predictions and potentially excessive loss magnitudes in outlier scenarios for the DCC-T model. In contrast, the HAR-RV model proves to be a robust tool that effectively maintains its predictive integrity across different portfolio risk profiles, providing a versatile and user-friendly approach to reliable portfolio VaR forecasting.

Conclusion

Our study advances the understanding of volatility forecasting by rigorously examining the efficacy of a univariate model based on high-frequency intraday data in improving portfolio VaR predictions, compared to traditional univariate and multivariate models using daily data. While previous research has applied high-frequency-based univariate models to index portfolios, our work provides a novel assessment of their robustness across diversified risk profiles and market conditions, particularly during financial crises.

We find that incorporating the skewness and excess kurtosis of return innovations into volatility models is crucial for accurate VaR forecasting. Our findings indicate that conditional realized volatility models based on high-frequency data, outperform both univariate and multivariate GARCH models. This superiority in forecasting accuracy and efficiency can be attributed to the high-frequency models’ ability to capture the fine-grained dynamics of market movements, which traditional GARCH models, relying on lower-frequency data, might miss. Moreover, we discover negligible performance differences between various conditional realized volatility models, indicating the efficacy of the standard HAR-RV model in capturing the long memory characteristic of realized volatility for portfolio VaR forecasting. The multivariate DCC model with Student t innovation, while effective in modeling time-varying correlations and outperforming the simpler EWMA technique, exhibits certain limitations. In particular, its sensitivity to changes in portfolio risk profiles and the magnitude of potential losses pose challenges. This may be due to its assumption of a symmetric innovation distribution, which does not always match the actual distribution of asset returns. The complexity of the model, involving the estimation of multiple parameters, also increases the risk of overfitting.

However, our approach is not without its limitations. The focus on the HAR model and its asymmetric extension was intentional, aiming to provide a focused examination of the impact of high-frequency data on volatility estimation. However, the potential of high-frequency intraday data to improve portfolio tail risk forecasting has not been fully realized and warrants further exploration. Future study could explore alternative models, such as Bayesian random fluctuation models to further validate the application of high-frequency data in risk management. Furthermore, future research could benefit from combining cutting-edge methodologies with high-frequency data. For instance, machine learning algorithms could be applied to analyze high-frequency data, while alternative data sources could be incorporated to provide a more comprehensive picture of market circumstances. The inclusion of comprehensive data indicators, distribution characteristics, and graphical analyses could further validate the application of these advanced methods in capturing market volatility and tail risks. Additionally, our empirical analysis, based on data from 2001 to 2009, captures a period of market stability and volatility, including the 2008 financial crisis. The evolving landscape of financial markets post-2009, with new regulatory frameworks and market dynamics, may influence the generalizability of our findings, suggesting the need for future research to assess the applicability of our results in the contemporary market environment. Moreover, the evaluation of Expected Shortfall (ES) forecasts, a complementary measure to VaR, would be a valuable addition to future work, providing a more comprehensive assessment of the predictive capabilities of models under different risk scenarios.

In conclusion, our research underscores the value of high-frequency intraday data in refining volatility predictions and portfolio VaR forecasts, while also highlighting areas for further investigation to enhance the practical application of risk management strategies in the financial industry.

Supporting information

S1 Checklist. Human participants research checklist.

https://doi.org/10.1371/journal.pone.0303962.s001

(DOCX)

Acknowledgments

The author thanks the editor and three referees for their detailed review and valuable comments that greatly improved the paper.

References

  1. 1. Sener E, Baronyan S, Mengütürk LA. Ranking the predictive performances of value-at-risk estimation methods. International Journal of Forecasting. 2012;28(4):849–73.
  2. 2. BCBS. Minimum capital requirements for market risk. Basel Committee on Banking Supervision; 2019.
  3. 3. Christoffersen P Value-at-Risk Models. In: Andersen TG, Davis RA, K JP, Mikosch TV, editors. Handbook of Financial Time Series. Berlin: Springer; 2009. p. 753–66.
  4. 4. Santos AAP, Nogales FJ, Ruiz E. Comparing Univariate and Multivariate Models to Forecast Portfolio Value-at-Risk. Journal of Financial Econometrics. 2013 10;11(2):400–41.
  5. 5. Diks C, Fang H. Comparing density forecasts in a risk management context. International Journal of Forecasting. 2020;36(2):531–51.
  6. 6. Brooks C, Persand G. Volatility forecasting for risk management. Journal of Forecasting. 2003;22(1):1–22.
  7. 7. Ferreira MA, Lopez JA. Evaluating Interest Rate Covariance Models Within a Value-at-Risk Framework. Journal of Financial Econometrics. 2005 01;3(1):126–168.
  8. 8. Lopez JA, Walter CA. Evaluating Covariance Matrix Forecasts in a Value-at-Risk Framework. Journal of Risk. 2001;3:69–98.
  9. 9. Rombouts JVK, Verbeek M. Evaluating portfolio Value-at-Risk using semi-parametric GARCH models. Quantitative Finance. 2009;9(6):737–45.
  10. 10. Giot P, Laurent S. Value-at-risk for long and short trading positions. Journal of Applied Econometrics. 2003;18(6):641–63.
  11. 11. Bauwens L, Laurent S. A New Class of Multivariate Skew Densities, With Application to Generalized Autoregressive Conditional Heteroscedasticity Models. Journal of Business & Economic Statistics. 2005;23(3):346–54.
  12. 12. Braione M, Scholtes NK. Forecasting Value-at-Risk under Different Distributional Assumptions. Econometrics. 2016;4(1):1–27.
  13. 13. Ku YHH, Wang JJ. Estimating portfolio value-at-risk via dynamic conditional correlation MGARCH model—an empirical study on foreign exchange rates. Applied Economics Letters. 2008;15(7):533–8.
  14. 14. Huang JJ, Lee KJ, Liang H, Lin WF. Estimating value at risk of portfolio by conditional copula-GARCH method. Insurance: Mathematics and Economics. 2009;45(3):315–24.
  15. 15. Luca GD, Rivieccio G, Corsaro S. Value-at-Risk dynamics: a copula-VAR approach. The European Journal of Finance. 2020;26(2-3):223–37.
  16. 16. McAleer M, da Veiga B. Single-index and portfolio models for forecasting value- at-risk thresholds. Journal of Forecasting. 2008;27(3):217–35.
  17. 17. Engle RF. Autoregressive Conditional Heteroscedasticity with Estimates of the Variance of United Kingdom Inflation. Econometrica. 1982;50(4):987–1007.
  18. 18. Bollerslev T. Generalized Autoregressive Conditional Heteroskedasticity. Journal of Econometrics. 1986;31:307–27.
  19. 19. Andersen TG, Bollerslev T, Diebold FX, Ebens H. The distribution of realized stock return volatility. Journal of Financial Economics. 2001;61(1):43–76.
  20. 20. Andersen TG, Bollerslev T, Diebold FX, Labys P. Modeling and Forecasting Realized Volatility. Econometrica. 2003;71(2):579–625.
  21. 21. Barndorff-Nielsen OE, Shephard N. Estimating Quadratic Variation Using Realized Variance. Journal of Applied Econometrics. 2002;17(5):457–77.
  22. 22. Shephard N, Sheppard K. Realising the Future: Forecasting with High- Frequency-based Volatility (HEAVY) Models. Journal of Applied Econometrics. 2010;25(2):197–231.
  23. 23. Hansen PR, Huang Z, Shek HH. Realized GARCH: A Joint Model for Returns and Realized Measures of volatility. Journal of Applied Econometrics. 2012;27(6):877–906.
  24. 24. Noureldin D, Shephard N, Sheppard K. Multivariate High-Frequency-based Volatility (HEAVY) models. Journal of Applied Econometrics. 2012;27(6):907–33.
  25. 25. Hansen PR, Lunde A, Voev V. Realized Beta GARCH: A Multivariate GARCH Model With Realized Measures OF Volatility. Journal of Applied Econometrics. 2014;29(5):774–99.
  26. 26. Gorgi P, Hansen PR, Janus P, Koopman SJ. Realized Wishart-GARCH: A Score-driven Multi-Asset Volatility Model. Journal of Financial Econometrics. 2018 04;17(1):1–32.
  27. 27. Giot P, Laurent S. Modelling daily Value-at-Risk using realized volatility and ARCH type models. Journal of Empirical Finance. 2004;11(3):379–98.
  28. 28. Beltratti A, Morana C. Statistical benefits of value-at-risk with long memory. Journal of Risk. 2005;7(4):21–45.
  29. 29. Brownlees CT, Gallo GM. Comparison of Volatility Measures: a Risk Management Perspective. Journal of Financial Econometrics. 2009 07;8(1):29–56.
  30. 30. Louzis DP, Xanthopoulos-Sisinis S, Refenes AP. The Role of High-Frequency Intra-daily Data, Daily Range and Implied Volatility in Multi-period Value-at-Risk Forecasting. Journal of Forecasting. 2013;32(6):561–76.
  31. 31. Meng X, Taylor JW. An approximate long-memory range-based approach for value at risk estimation. International Journal of Forecasting. 2018;34(3):377–88.
  32. 32. Gerlach R, Wang C. Semi-parametric dynamic asymmetric Laplace models for tail risk forecasting, incorporating realized measures. International Journal of Forecasting. 2020;36(2):489–506.
  33. 33. Lazar E, Xue X. Forecasting risk measures using intraday data in a generalized autoregressive score framework. International Journal of Forecasting. 2020;36(3):1057–72.
  34. 34. Andersen TG, Bollerslev T, Diebold FX, Labys P. The Distribution of Realized Exchange Rate Volatility. Journal of the American Statistical Association. 2001;96(453):42–55.
  35. 35. Bollerslev T, Hood B, Huss J, Pedersen LH. Risk Everywhere: Modeling and Managing Volatility. The Review of Financial Studies. 2018 05;31(7):2729–73.
  36. 36. Corsi F. A Simple Approximate Long-Memory Model of Realized Volatility. Journal of Financial Econometrics. 2009;7(2):174–96.
  37. 37. Louzis DP, Xanthopoulos-Sisinis S, Refenes AP. Stock index realized volatility forecasting in the presence of heterogeneous leverage effects and long-range dependence in the volatility of realized volatility. Applied Economics. 2012;44(27):3533–50.
  38. 38. Bollerslev T, Chou RY, Kroner KF. ARCH modeling in finance: A review of the theory and empirical evidence. Journal of Econometrics. 1992;52(1):5–59.
  39. 39. Bera AK, Higgins ML. ARCH Models: Properties, Estimation And Testing. Journal of Economic Surveys. 1993;7(4):305–66.
  40. 40. Palm FC. GARCH models of volatility. In: Statistical Methods in Finance. vol. 14 of Handbook of Statistics. Elsevier; 1996. p. 209–40.
  41. 41. Bauwens L, Laurent S, Rombouts JVK. Multivariate GARCH models: a survey. Journal of Applied Econometrics. 2006;21(1):79–109.
  42. 42. Silvennoinen A. Multivariate GARCH Models. In: Andersen TG, Davis RA, K JP, Mikosch TV, editors. Handbook of Financial Time Series. Berlin: Springer; 2009. p. 201–29.
  43. 43. Andersen TG, Bollerslev T. Answering the Skeptics: Yes, Standard Volatility Models do Provide Accurate Forecasts. International Economic Review. 1998;39(4):885–905.
  44. 44. Barndorff-Nielsen OE, Shephard N. Econometric Analysis of Realized Covariation: High Frequency Based Covariance, Regression, and Correlation in Financial Economics. Econometrica. 2004;72(3):885–925.
  45. 45. Müller UA, Dacorogna MM, Davé RD, Olsen RB, Pictet OV, von Weizsäker JE. Volatilities of different time resolutions-Analyzing the dynamics of market components. Journal of Empirical Finance. 1997;4(2):213–239.
  46. 46. Glosten LR, Jagannathan R, Runkle DE. On the Relation between the Expected Value and the Volatility of the Nominal Excess Return on Stocks. The Journal of Finance. 1993;48(5):1779–801.
  47. 47. Ding Z, Granger CWJ, Engle RF. A long memory property of stock market returns and a new model. Journal of Empirical Finance. 1993;1(1):83–106.
  48. 48. Baillie RT, Bollerslev T, Mikkelsen HO. Fractionally integrated general- ized autoregressive conditional heteroskedasticity. Journal of Econometrics. 1996;74(1):3–30.
  49. 49. Morgan J. RiskMetricesTM-Technical Document. 4th ed. New York; 1996.
  50. 50. Engle R. Dynamic Conditional Correlation: A Simple Class of Multivariate Generalized Autoregressive Conditional Heteroskedasticity Models. Journal of Business & Economic Statistics. 2002;20(3):339–50.
  51. 51. Fernández C, Steel MFJ. On Bayesian Modeling of Fat Tails and Skewness. Journal of the American Statistical Association. 1998;93(441):359–71.
  52. 52. Gerlach R, Lu Z, Huang H. Exponentially Smoothing the Skewed Laplace Distribution for Value-at-Risk Forecasting. Journal of Forecasting. 2013;32(6):534–50.
  53. 53. Kuang W. Dynamic VaR forecasts using conditional Pearson type IV distribution. Journal of Forecasting. 2021;40(3):500–11.
  54. 54. Barone-Adesi G, Giannopoulos K, Vosper L. VaR without correlations for portfolios of derivative securities. Journal of Futures Markets. 1999;19(5):583–602.
  55. 55. Corsi F, Mittnik S, Pigorsch C, Pigorsch U. The Volatility of Realized Volatility. Econometric Reviews. 2008;27(1-3):46–78.
  56. 56. Lambert P, Laurent S. Modelling financial time series using GARCH-type models with a skewed student distribution for the innovations. Working paper: University of Liege; 2001. 0125.
  57. 57. Cai J. A Markov model of switching-regime ARCH. Journal of Business & Economic Statistics. 1994 Jul 1;12(3):309–16.
  58. 58. Andersen T, Bollerslev T, Hadi A. ARCH and GARCH models. John Wiley & Sons. 2014.
  59. 59. Lambert P, Laurent S. Techniques for Verifying the Accuracy of Risk Measurement Models. Journal of Derivatives. 1995;3:73–84.
  60. 60. Christoffersen PF. Evaluating Interval Forecasts. International Economic Review. 1998;39:841–62.
  61. 61. Berkowitz J. Testing Density Forecasts, with Applications to Risk Management. Journal of Business & Economic Statistics. 2001;19(4):465–74.
  62. 62. Rosenblatt M. Remarks on a Multivariate Transformation. The Annals of Mathematical Statistics. 1952;23(3):470–2.
  63. 63. Giacomini R, Komunjer I. Evaluation and Combination of Conditional Quantile Forecasts. Journal of Business & Economic Statistics. 2005;23(4):416–31.
  64. 64. Lopez JA. Methods for Evaluating Value-at-Risk Estimates. FRBNY Economic Policy Review. 1998;4(3):119–24.
  65. 65. Harris RDF, Guermat C. Robust Conditional Variance Estimation and Value- at-Risk. Journal of Risk Research. 2002;4:25–42.
  66. 66. Giacomini R, Rossi B. Forecast comparisons in unstable environments. Journal of Applied Econometrics. 2010;25(4):595–620.
  67. 67. Diebold FX, Mariano RS. Comparing Predictive Accuracy. Journal of Business & Economic Statistics. 1995;20(1):134–44.
  68. 68. McAleer M, Jiménez-MartÃn JA, Pérez-Amaral T. International Evidence on GFC-Robust Forecasts for Risk Management under the Basel Accord. Journal of Forecasting. 2013;32(3):267–88.
  69. 69. BCBS. Supervisory framework for the use of backtesting in conjunction with the internal models approach to market risk capital requirements. Basel Committee on Banking Supervision; 1996.
  70. 70. Liu M, Lee CC, Choo WC. The role of high-frequency data in volatility forecasting: evidence from the China stock market. Applied Economics. 2021 May 9;53(22):2500–26.
  71. 71. Lyócsa Š, Molnár P, Výrost T. Stock market volatility forecasting: Do we need high-frequency data?. International Journal of Forecasting. 2021 Jul 1;37(3):1092–110.
  72. 72. Naimoli A, Gerlach R, Storti G. Improving the accuracy of tail risk forecasting models by combining several realized volatility estimators. Economic Modelling. 2022 Feb 1;107:105701.
  73. 73. Andersen TG, Bollerslev T, Diebold FX, Wu Gv. Realized Beta: Persistence and Predictability. In: Fomby TB, Terrell D, editors. Econometric Analysis of Financial and Economic Time Series. Bingley: Emerald Group Publishing Limited; 2006. p. 1–39.
  74. 74. Fleming J, Kirby C, Ostdiek B. The Economic Value of Volatility Timing Using “Realized” Volatility. Journal of Financial Economics. 2003;67(3):473–509.
  75. 75. Kuang W. Conditional covariance matrix forecast using the hybrid exponentially weighted moving average approach. Journal of Forecasting. 2021;40(8):1398–419.