Description of the Data
Investors can potentially improve the risk-adjusted performance of their portfolios by investing internationally and thereby take advantage of the associated return and diversification benefits. The potential gains provided by emerging markets have attracted significant investor attention which, in turn, has led to substantial capital inflows to these economies. Local currency-denominated sovereign bonds have been the fastest growing market in emerging market space in the past few years. More recently, the global quest for yield in a context of accommodative monetary policies in the advanced economies has created a positive external environment for emerging market bonds. In addition, the secondary market for emerging market local currency debt has also been supported by high interest and amortisation payments. Emerging market bonds are also benefiting from a track record of strong risk-adjusted returns and low correlations with other asset classes. Such characteristics are attractive from a portfolio optimisation perspective. These attributes have also seen considerable attention devoted to analysis of the various risk-return attributes of these markets. In particular, recent empirical literature has focused on the characterisation of the volatility profile of emerging market bond returns. Indeed, the accurate estimation of volatility plays a central role in many applications in finance, including optimal portfolio selection (e.g., diversification strategies), valuation of derivatives (e.g., option pricing) and risk management (e.g., value-at-risk calculation). These applications have motivated an extensive empirical literature on volatility modelling.
While recent empirical literature has focused on the characterisation of the volatility profile of a variety of asset classes; in particular, the long memory properties of these assets, surprisingly little attention has been devoted to the analysis of fixed income markets, especially in emerging markets. The empirical literature on long memory dynamics of fixed income volatility and its implications for portfolio and risk management appears to be limited. Most of the extant literature appears concentrated in the advanced economies. For example, several authors have examined the various aspects of long memory behaviour in interest rates and yield spreads. [1-3] The purpose of this chapter is to augment this line of analysis concerning the long memory attributes of fixed income volatility in emerging market local currency debt market in light of investor interest in the potential alpha generation of these markets, amid wider capital inflows into emerging market bonds as investors search for yield.
This study will focus on government bond markets from Hong Kong, Mexico and South Africa. According to the most recent survey from the Emerging Markets Trading Association (EMTA) these three local currency debt markets are among the most vibrant and actively traded in emerging markets.  As a result of their (comparatively high) liquidity, developed institutional frameworks and credible monetary policies, these markets are therefore of interest to investors. Indeed, local currency bond markets have emerged as an important asset class in many emerging markets; a point which becomes salient in the current low-yield environment, where investors targeting high returns and diversification benefits have channelled capital to emerging bond markets such as these.
As a result of regulatory initiatives and various reforms, emerging market local currency sovereign debt markets have has grown rapidly in size and sophistication. According to , emerging market debt trading volumes were at USD1.8 trillion in the third quarter of 2011. This represents a 3 percent increase from the USD1.7 trillion reported for the second quarter of 2011. Turnover in local market instruments was at USD1.3 trillion in the third quarter of 2011, (i.e., 76 percent of total reported volume). Mexican and Hong Kong securities were the first and second most frequently traded emerging market debt in the third quarter of 2011 at USD282 billion and USD176 billion, respectively compared to USD136 billion and USD201 billion a year earlier. The next most frequently traded local markets debt were those from Brazil (USD160 billion), South African (USD113 billion) and Turkey (USD69 billion). Analysis is on Hong Kong, Mexico and South Africa due to the availability of data.
Analysis is on Hong Kong, Mexico and South Africa due to the availability of data.
Domestic institutional investors are typically the largest investors in local currency bond markets. For example, pension funds tend to have long-term liabilities which are typically funded by investments in long-term investment grade securities that provide a prudent risk-return profile. As a result, the examination of long memory in volatility would appear to be of interest to institutional and other long-term investors. Furthermore, it has been shown that it is important to model the long memory volatility structure when pricing derivative contracts with long maturity.  In addition, in order to assess future returns from both active and passive investment strategies or the need for policy intervention, especially over long horizons, it is important to forecast volatility. These applications have motivated an extensive empirical literature on modelling long memory dynamics in asset return data. Analysis of the long-term volatility dynamics of emerging market local currency bonds appear limited. Therefore, this study attempts to help fill this gap by addressing a range of issues relating to the estimation and forecasting of fixed income return volatility especially over long horizons.
Against this background, this paper has three objectives. First, evidence of long memory in volatility within leading emerging fixed income markets is investigated. In particular, the existence of long memory behaviour in the volatility of returns from Hong Kong, Mexico and South Africa are examined which appears to have little or no previous research establishing the existence of its long memory properties. In order to estimate the long memory parameter
In total, the findings of this investigation will provide a range of volatility estimates and forecasts which could potentially inform portfolio management strategies and guide policymaking. In particular, while most empirical studies focus on the United States and other developed markets, recent research has begun to look at emerging markets, however, limited evidence exists with respect to these markets. This analysis contributes to the empirical literature by focusing on various aspects of long memory behaviour in local currency debt markets. The findings from this research complement those in previous studies and may provide an interesting comparison to existing studies.
The rest of the chapter is structured as follows. Section II presents a description of long memory in time series. Section III introduces the data. Section IV presents the empirical methodology and associated results. In particular, this starts with a presentation of the standard GARCH model which is often used to present initial evidence of long memory behaviour. Then the wavelet method and the estimator employed is introduced, along with the relevant findings. This includes a discussion on wavelet analysis, the discrete wavelet transform, the estimator employed and the relevant findings. Section V provides the forecast evaluation techniques used and the out-of-sample forecast results. Section VI considers the evaluation of value-at-risk in the context of the Basle adequacy criteria. Section VII concludes and identifies topics for further research.
2. Long memory in time series
Interest in long memory (or long range dependent) processes can be traced to the examination of data in the physical sciences. Formal models with long memory initially pertained to hydrological studies investigating how to regularise the flow of the Nile river in view of its nonperiodic (flooding) cycles.  This feature was described as the “Joseph effect” alluding to the biblical reference in which seven years of plenty where to be followed by seven years of famine.  In this sense, long memory process concern observations in the remote past that are highly correlated with observations in the distant future. The implications of long memory in financial markets was related to the use of Hurst’s ‘rescaled range’ statistic to detect long memory behaviour in asset return data.  It was observed that if security prices display long memory then the arrival of new market information cannot be arbitraged way, which in turn means that martingale models for security prices cannot be derived through arbitrage. As such, long memory processes can be characterised as having fractal dimensions, in the form of non-linear behaviour marked by distinct but nonperiodic cyclical patterns and long-term dependence between distant observations. 
A variety of measures have been used to detect long memory in time series. For example, in the time domain, long memory is associated with a hyperbolically decaying autocovariance function. Meanwhile, in the frequency domain, the presence of long memory is indicated by a spectral density function that approaches infinity near the zero frequency; in other words, such series display power at low frequencies.  Finally, a pattern of self-similarity in the aggregated sequences of a time series is an indicator of long memory (this refers to the property of a self-similar process, in which, different time aggregates display the same autocorrelation structure). These notions have led several authors to develop stochastic models that capture long memory behaviour, such as the fractionally-integrated I(
Since, non-zero values of the fractional differencing parameter imply dependence between distant observations, considerable attention has been directed to the analysis of fractional dynamics in financial time series data. Indeed, long memory behaviour has been reported in the returns of various asset classes.  Against this background, a rapidly expanding set of models has been developed to capture long memory dynamics is asset return data.
3. Data description
The data analysed in this study are obtained from the global bond index (GBI) series for emerging markets (EM) compiled by JP Morgan. In particular, the fixed income data used comprise of daily total returns for Hong Kong, Mexico, and South Africa from December 31, 2001 to April 9, 2012, representing 2571 observations.
More formally, the change in the local bond index can be expressed as:
where is the local currency return and is the currency return. and and the coupon return is the return derived from the interest payment made on the fixed income product. Therefore, when an investor buys a local market index, equation (1) suggests that fixed income returns can be decomposed into its predictable coupon return, FX changes, and changes in local yields. Furthermore, in order to compute return volatility, this analysis focuses on squared daily returns, as a proxy for the volatility of the selected emerging markets. In addition, the volatility series is standardised prior to further analysis.
4. Empirical methodology and results
4.1. Preliminary observations
Table 1 presents the time series properties of the data using some basic methods. The results of the Augmented Dickey-Fuller (ADF) unit root test offer evidence in favour of stationary fixed income returns. While this test may be deficient in terms of its ability to capture an order of integration that may not be an integer, the finding of stationary bond returns is consistent with those of many previous studies.  However, based on the standard normality and Lagrange Multiplier ARCH tests, fixed income return data exhibit non-normality and ARCH effects. [16-17] These non-white noise characteristics of the data motivate estimation of GARCH(1,1) model using the assumption of the Student
|Mexico||South Africa||Hong Kong|
ARCH (5) test
ARCH (10) test
ADF unit root test:
Constant & Trend
4.2. GARCH(1,1) model
The GARCH (1,1) specification comprises a return (or mean) and a variance equation. In particular, the returns generating process can be described by:
where denotes the returns process, which may include autoregressive and moving average components, and
The GARCH (1,1) model estimates are reported in Table 2. The results confirm the previous findings on the importance of GARCH effects by showing that the GARCH and ARCH terms are all statistically significant. The parameters of the conditional variance equations are all positive and statistically significant. Furthermore, they satisfy the positivity constraint for the GARCH(1,1) model. Furthermore evidence of persistence in variance as measured by the GARCH model is reflected in the magnitude and significance of the ARCH and GARCH terms (indeed, as this sum approaches unity the greater the degree of persistence). Therefore, in order to have an indication of long memory in fixed income return volatility the level of volatility persistence (i.e., α+β) is assessed.
The results indicate that volatility in these markets is very persistent, with the level of volatility persistence being 0.9775 for Mexico, 0.9782 for South Africa and 0.9912 for Hong Kong, which underscores the highly persistent nature of shocks to volatility, which also in turn is suggestive of a long memory component to volatility behaviour in these fixed income markets. The models’ appropriateness has also been checked by applying the Box-Pierce Q statistic test to standardised and squared standardised residuals. Basic diagnostics indicate that the GARCH models are well-specified.
4.3. Wavelet analysis
To estimate the long memory parameter
|Mexico||South Africa||Hong Kong|
Sign bias test
Negative size bias test
Positive size bias test
A wavelet is defined as a wave-like function with an amplitude that oscillates around zero and has a finite or quickly decreasing time support. These functions are well suited to locally approximating variables in time or space as they have the ability to be manipulated by being either ‘stretched’ or ‘squeezed’ so as to mimic the series under investigation.  The power of wavelet analysis is that it makes it possible to decompose a time series into its high- and low-frequency components, which are localised in time. Wavelets, also allow the selection of an appropriate trade-off between resolution in the time and frequency domains, while traditional Fourier analysis stresses resolution in the frequency domain at the expense of the time domain.  Wavelets therefore provide a convenient and efficient method to analyse complex signals.  Wavelet theory is applicable to several subjects. They are especially useful where a signal (e.g. long memory) lasts for a finite time or shows markedly different behaviour in different time periods. These methods, have emerged as a useful tool in the empirical finance literature where long-run and short-run relationships can be distinguished. 
4.4. The discrete wavelet transform
A discrete wavelet transform (DWT) is any wavelet transform for which the wavelets are discreetly sampled, as is often the case in econometric analysis. A wavelet transform is a scaling function used to transform a signal into a father and motherwavelet, where the former, are representations of a signal’s trend component (i.e., the approximation coefficients) and the latter represent the deviations from the trend component (i.e., the detail coefficients). The discrete wavelet series approximation to a continuous signal
In the empirical literature, Haar and Daubechies wavelets represent typical wavelets and have been used in the characterisation of the time series properties of asset return data. The Haar wavelet is the simplest wavelet and provides a basis for studying more complex wavelets. The Haar transform assumes a discrete signal and decomposes the signal into two sub-signals of half its length reflecting the trend process and fluctuations from the trend process.
The Haar transform assumes a discrete signal and decomposes the signal into two sub-signals of half its length reflecting the trend process and fluctuations from the trend process.
The data used in this study are discreetly sampled, accordingly, the discrete wavelet transform is used, which permits the generation of the approximation coefficients, , which capture the trend of a time series and the detail coefficients, , reflecting the deviations from the trend at each scale. Because the original function can be represented in terms of a wavelet expansion, data operations can be performed using the corresponding wavelet coefficients. This leads to a continuum of time-scale representations of the signal, all with different resolutions. Hence, multi-resolution analysis, which allows the computation of the coefficients corresponding to the wavelet transform of the observed time series.
The analysis of fractionally integrated processes through the use of wavelets is based on the result that the detail coefficients of a zero mean long memory process are asymptotically normally distributed with variance, where is a finite constant,
where is the variance of the detail coefficients associated with the value of the scaling parameter
To estimate, this study uses multi-resolution analysis via the Haar wavelet to generate the respective detail coefficients of each volatility series at each dimension of scaling parameter. From here, the variance of the detail coefficients at each scale are computed and then the regression specified in equation (5) is performed, where the slope coefficient provides an estimate of. To check for the robustness of the results, and, therefore, avoid spurious conclusions of long memory dynamics the Daubechies 4 (D4) wavelet is also examined.
The regression results of this analysis are presented in Table 3 and 4. The slope coefficient of the regression given in equation (3) provides an estimate,
|Volatility series||Identifier||Parameter Estimate||Standard Error|
In econometric analysis, it is important to perform diagnostic checks in order to assess the validity of the initial estimates of. Therefore, to avoid spurious evidence of long memory (due to the choice of wavelet employed) in the volatility process of the time series, equation (4) is re-estimated using the Daubechies 4 (D4) wavelet. These results are presented in Table 4. The results are broadly similar in magnitude to those obtained using the Haar wavelet. The noticeable exception relates to the case of South Africa where the long memory parameter falls from 0.2679 (when the Haar wavelet is used) to 0.1784 (when the D4 wavelet is used). This notwithstanding, the results are all statistically significant. In sum, the results of this analysis suggest that bond return volatility in emerging markets is characterised by stochastic processes which have a long memory component.
|Volatility series||Identifier||Parameter Estimate||Standard Error|
The analysis indicates robust evidence of long memory behaviour in the return volatility of emerging market debt. Further, wavelet methods provide a robust fit for the data as evidence by the readings presented in the final columns of both Table 3 and 4. If fixed income data exhibit long memory, then it displays significant autocorrelation between distant observations. This, in turn, implies that the series realisations may have a predictable component; and, perhaps, past trends in the data can be used to predict future volatility. Therefore, attention now turns to an exploration of the forecast performance of models with long memory relative to the standard volatility models.
5. Volatility forecasting
The evidence accumulated so far suggests that fixed income return volatility in emerging markets follows a long memory process. This, in turn, implies the existence of fractional dynamics in the data which may be exploited to potentially construct improved volatility forecasts, especially over longer forecasting horizons. In order to evaluate the forecasting performance of long memory models (especially over long(er) horizons) versus short memory models (i.e., the GARCH model), the respective data sets are simply split in half and then each model is estimated for all series covering the first part of the sample and then these estimates are used to forecast volatility over the sample period covered by the second half of the data. In this manner, out-of-sample forecast accuracy is evaluated. In addition to calculating the daily forecasts, this study also calculates monthly forecasts using the well-known property that volatility forecasts are additive, such that the sum of five daily volatility forecasts produces the weekly forecasts. And, the summation of weekly forecasts produces monthly forecasts.
In addition to the GARCH and long memory model the RiskMetrics model is also considered for comparative purposes. The RiskMetrics model was popularised by the investment bank JP Morgan and is widely used by financial institutions to model and forecast volatility, especially in the context of the Basle Committee adequacy criteria. This model is essentially an exponentially weighted moving average (EWMA). Under the EWMA, the fitted variance from the model, which provides the multi-step ahead volatility forecast, is a weighted function of the immediately preceding volatility forecast and actual volatility is given below:
where is the smoothing parameter, such that when= 0 the model reduces to a random walk process and when =1 the model is equivalent to the prior period forecast of volatility. The value of is determined empirically by the value that minimizes the ‘in-sample’ sum of squared prediction errors. In this study is set to 0.94 following standard market practice, which is also consistent with previous research which indicates that this value produces accurate forecasts. 
5.1. Standard forecast evaluation
Two standard symmetric measures are used to evaluate forecast accuracy, namely, the mean absolute error (MAE) and the root mean square error (RMSE). They are defined below:
where is the number of forecast data points and is the proxy for volatility. Both the MAE and RMSE assume the underlying loss function to be symmetric. Furthermore, under these evaluation criteria the model which minimises the loss function is preferred.
Table 5 reports out-of-sample performance of the estimated models based on the MAE and RMSE forecast error statistics. At the daily level, the results are not unexpected. That is, the GARCH model dominates forecast accuracy for South Africa on the basis of both the MAE and RMSE. For Mexico, the GARCH model dominates forecast performance on the basis of the MAE while the RiskMetrics models delivers the most accuracy when the RMSE is used as a criterion. For Hong Kong the GARCH process is preferred on the basis of the MAE, which surprisingly, the long memory model delivers the best performance when the RMSE is used as a reference. However, in some cases the forecast accuracy of all the models are close; for instance at the daily level the forecast MAE statistics for the GARCH, RiskMetrics and FIGARCH models are virtually indistinguishable. More generally, the findings of GARCH superiority at the daily level are consistent with a wide empirical literature attesting to the forecast superiority of the GACH model at forecasting volatility over daily frequencies or short horizons.
|Forecast Error Statistic||MAE||RMSE||MAE||RMSE||MAE||RMSE|
At the monthly level (i.e., at a longer horizon) the GARCH model also delivers the most accurate results. This finding is surprising. Long memory implies that widely separated observations are associated with each other which in turn suggests that volatility realizations are connected over long lags. The results shows that at even comparatively longer horizons the GARCH model still delivers the most accurate volatility forecasts. Indeed, Table 6 shows that the forecast MAE statistics for Mexico and South Africa are 3.13e-03 and 3.92e-03, respectively, which are smaller than those from long memory models. The same results holds true for the forecast RMSE error statistics. For Hong Kong fixed returns This result appears to suggest that long memory models while theoretically appealing are not particularly helpful in deriving accurate volatility forecast especially over long horizons.
|Forecast Error Statistic||MAE||RMSE||MAE||RMSE||MAE||RMSE|
6. Value-At-Risk evaluation
VaR is a widely used measure to capture the exposure of a portfolio to market risk. The VaR of a position describes the expected maximum loss over a target horizon within a given confidence interval due to an adverse movement in the relevant fixed income yield (or price). VaR is now widely used as an internal risk management tool by financial institutions and as a regulatory measure of risk exposure.  In addition, the VaR method is the cornerstone of the 1996 market risk amendment to the Basle Accord (Bank of International Settlements, (BIS), 1996). The Basle Accord prescribes the VaR method in order that financial institutions can meet the capital requirements to cover the market risk they incur in the process of their daily business operations. Under this framework, operational evaluation takes the form of backtesting volatility forecasts and exception reporting.
In particular, the Basle Accord stipulates that for the purpose of calculating regulatory market risk capital it is required that VaR estimates be calculated at the 99 percent probability level using daily data over a minimum sample period of at least one business year (equivalent to 250 trading days) and that these estimates be updated at least every quarter (i.e., 60 trading days). Against this background, the well-known delta-normal specification is employed:
where is the appropriate standard normal deviate, is the volatility forecast, the number three represents the minimum regulatory Basle multiplicative factor and is the initial portfolio value. While Basle Accord prescribes a 99 percent probability the 97.5 percent and 95 percent confidence level is also examined for greater comprehensive and consistency with previous studies. The validity of such VaR calculations are assessed or ‘backtested’ by comparing actual daily trading (net) losses with the estimated VaR and noting the number of ‘exceptions’, in the sense of days when the VaR estimate was insufficient to cover actual trading losses. Regulatory scrutiny is therefore triggered where such exceptions occur frequently, and in practice this leads to a range of penalties for the financial institution concerned. 
In line with the rolling window approach to VaR evaluation mandated by the Basle Committee rules initial volatility forecasts and VaR measures are constructed over intervals of 60 trading days, with the initial estimation sample then rolled forward and the models updated every 60 observations before the next set of volatility forecasts are produced. The first 3-years of data (representing 752 observations) are used for initial model parameter estimation, leaving 1819 observations for volatility forecasting and the construction and evaluation of VaR measures. Specifically, this provides 30 sub-samples of 60 trading days length over which VaR is assessed. This assessment is conducted through appraisal of the out-of-sample VaR failure rates associated with VaR measures constructed using the forecast values of those volatility measures. The assessment of VaR performance is conducted through appraisal of the ‘out-of-sample’ VaR failure rates associated with VaR measures constructed using the forecast values derived from the GARCH, RiskMetrics and long memory model. The focus on the ‘out-of-sample’ failure rates is motivated by the requirements of risk managers, who obtain VaR estimates in real time and must use parameters obtained from an already observed sample in order to evaluate the risks associated with current and future random movements in risk factors. As a result, credible test of VaR construction methods under alternative volatility forecasting models is their performance outside the sample used to estimate the underlying parameters.
Table 7 reports the out-of-sample VaR failure rates. The results are very diverse and highlight that in many of the markets considered the forecasting model that minimises the percentage number of daily VaR exceedances is sensitive to the specification of the probability level. When the Basle Committee rules are applied (i.e., the 99 percent probability level) the results indicate that the GARCH and RiskMetrics that provide the exceedance-minimising method for the fixed income markets considered. At the 99 percent probability level the long memory model is the generally the weakest performer. However, in the case of Hong Kong and South Africa the long memory model is second best model in terms of delivering accurate VaR measures. In addition, it is important to note that in many cases is level of accuracy between the various models is close as reflected by the closeness of the VaR failure rates. At the 97.5 and 95 percent probability levels model performance is more varied with all models demonstrating varied degrees of accuracy. As a generalization, these results are mixed but the evidence suggests that at the Basle prudential level the simpler models help in providing improved VaR estimates that minimise occasions when the minimum capital requirement identified by the VaR methodology would have fallen short of actual trading losses.
|Hong Kong||Mexico||South Africa|
Recent empirical evidence concerning the nature of volatility dynamics in fixed income markets suggests the existence of a long memory component. Since volatility in fixed income returns is an important aspect of portfolio management it is essential to accurately characterise the time series properties of fixed income volatility especially in the context of emerging markets where local currency-denominated sovereign bonds have been the fastest growing market in recent years. Accordingly, the objective of this analysis was to examine the existence of long memory behaviour in the volatility structure of total return indices for the local currency bond markets of Hong Kong, Mexico and South Africa. Against this background, the long memory parameter is estimated using methods based on wavelets, which have gained prominence in recent years. Furthermore, this study has compared and evaluated the performance of a long memory model versus a standard volatility models (the ubiquitous GARCH and RiskMetrics processes) in order to evaluate their power in delivering accurate volatility forecasts over long(er) horizons in an out-of-sample setting. This endeavour is motivated by recognition of the importance of accurate volatility forecasts in a wide range of applications, including tactical and strategic decision making and the limited empirical evidence available to date for emerging fixed income markets. Then, the performance of the standard GARCH, RiskMetrics and FIGARCH models are evaluated in the context of value-at-risk (VaR) estimation given the Basle regulatory framework.
The main findings of this research are threefold. First, evidence of long memory is conclusively demonstrated in emerging market local currency sovereign debt markets. In addition, to counteract the possibility of finding spurious evidence of long memory a variety of wavelet forms are considered. The findings from these tests are complementary and therefore suggest that the finding of long memory is not spurious. Second, the presence of a long memory structure in the volatility of these fixed income markets suggests volatility observations in the recent past and the remote past are associated with each other. Since the series realisations are not independent over time then past volatility may potentially be exploited to predict future volatility, especially over long horizons. Accordingly, the out-of-sample forecasting performance of the long memory model and the standard GARCH and RiskMetrics models are compared. While, none of the estimated models consistently outperforms the others, a key generalisation can be made. In particular, on the basis of the forecast MAE and RMSE statistics it is shown that the information content of long memory models does not consistently generate improved volatility forecasts, especially over long horizons, relative to the standard GARCH model. Indeed, the GARCH model generally provides the most accurate forecasts at the monthly horizon. With respect to VaR estimation, the results show that both the standard GARCH and RiskMetrics models generally deliver more accurate VaR measures relative to the long memory process.
These findings have three important implications. First, the exploitation of long memory models based on wavelet analysis may not have great relevance in the context of emerging market debt in terms of delivering superior forecast performance. Second, the existence of a long memory structure in volatility is not an essential condition for the derivation of accurate volatility forecasts at any time horizon, especially over a long horizon. Indeed, this research suggests that long memory models appear to be of limited practical forecast value, especially over long horizons, for Hong Kong, Mexico and South Africa. Put differently, the computational complexity of long memory modelling is not commensurate with the benefits (in terms of forecast power). Third, the results of the VaR estimation may provide guidance on more effective prudential standards for operational risk measurement and, as result, may help ensure adequate capitalisation and reduce the probability of financial distress. The results highlight the importance of using out-of-sample forecasting techniques and the stipulated probability level for the identification of methods that minimise the occurrence of VaR exceptions. Standard models – RiskMetrics and GARCH – that are already widely used by market participants are generally shown to outperform the more computationally intensive wavelet-derived FIGARCH model in estimating VaR across the probability levels considered.
In sum, this research has evaluated the long memory properties of return volatility in fixed income markets. This paper also complements the literature on long memory models and the forecast performance of these models that has attracted interest in other asset classes. In addition, the results of this study may potentially be used to inform portfolio and risk analysis. In particular, it is shown that in the context of VaR estimation existing models based on the GARCH and/or RiskMetrics process are more accurate (and simpler) than their long memory counterpart. Some caveats to these results exist, however. First, squared returns provide a noisy proxy for the ‘true’ volatility. In the case of this analysis, data constraints limited alternative options. However, future research may find that the application of realised variance may produce more accurate forecasts. Second, future research may also consider exploring the relevance of other long memory models, for example, models with asymmetric effects given that market volatility is often reported as being ‘directional’, i.e., volatility is higher in a down – than an upmarket.
- Analysis is on Hong Kong, Mexico and South Africa due to the availability of data.
- The Haar transform assumes a discrete signal and decomposes the signal into two sub-signals of half its length reflecting the trend process and fluctuations from the trend process.