Generalized Autoregressive Conditional Heteroscedasticity (GARCH) model and its variations have been widely adopted in the study of financial volatilities, while the extension of GARCH-type models to high-dimensional data is always difficult because of over-parameterization and computational complexity. In this article, we propose a multi-variate GARCH-type model that can simplify the parameterization by utilizing the network structure that can be appropriately specified for certain types of high-dimensional data. The asymmetry in the dynamics of volatilities is also considered as our model adopts a threshold structure. To enable our model to handle data with extremely high dimension, we investigate the near-epoch dependence (NED) of our model, and the asymptotic properties of our quasi-maximum-likelihood-estimator (QMLE) are derived from the limit theorems for NED random fields. Simulations are conducted to test our theoretical results. At last we fit our model to log-returns of four groups of stocks and the results indicate that bad news is not necessarily more influential on volatility if the network effects are considered.
1 Introduction
To pursue maximum return or to circumvent potential risk, investors constantly revise their portfolio according to any related information. Understanding how the volatility of financial assets responds to new information is crucial in risk management and a widely studied area in econometrics and statistics. In the literature, statistical models that describe the formation of financial risks have been developed and conducted in practice. The Autoregressive Conditional Heteroscedasticity (ARCH) model was proposed by Engle (1982) for estimating the variance of United Kingdom's inflation. In an ARCH() model, the volatilities of returns are affected by up to lags of past observations. Bollerslev (1986) then proposed a generalized ARCH model (GARCH), to accommodate longer memory of past observations. It has become one of the most popular models in econometrics ever since, and numerous variations of GARCH model have been developed for modeling volatility with complicated structures. See Teräsvirta (2009) for a survey of different GARCH-type models.
When we study risks of multiple assets simultaneously, the conditional variances that represent individual risks are of interest as well as conditional covariances that represent risk-sharing relationships. On the other hand, risk of a particular individual could be affected by covariates of itself, and of those who closely related to it. This leads to the need of extending the GARCH-type models into the multi-variate case. For an dimensional time series , a canonical expression of multi-variate GARCH would be
(1)
where random vector satisfies and and there could be various specifications of the function as it represents the structure of the conditional covariance matrix . For more details on this subject, an excellent survey paper by Bauwens et al. (2006) on the family of multi-variate GARCH models is recommended.
However, in terms of parameter estimation, there are major challenges that could make multi-variate GARCH (MGARCH) models inapplicable in empirical analysis when it comes to dealing with high-dimensional data. For example, the number of parameters rises at the speed of in vectorized GARCH model (VEC-GARCH) proposed by Bollerslev et al. (1988). Over parameterized specification causes high computational complexity, and makes it problematic to derive conditions for positive definiteness of the conditional covariance matrix . Plenty of efforts have been made in the literature, diagonal VEC-GARCH (DVEC-GARCH) model by Bollerslev et al. (1988) and Baba-Engle-Kraft-Kroner GARCH (BEKK-GARCH) model by Engle and Kroner (1995) are proposed with the aim of simplifying the conditions of positive definite by imposing structural restrictions on the conditional covariance matrix. The number of parameters could also be significantly reduced to in the Constant Conditional Correlation GARCH (CCC-GARCH) by Bollerslev (1990) and Dynamic Conditional Correlation GARCH (DCC-GARCH) by Engle (2002) and Tse and Tsui (2001). On the other hand, as an alternative way to overcome the over-parameterization problem, the idea of factor variables was imposed by Engle et al. (1990) on the multi-variate ARCH model as a dimension reduction technique. This idea is later introduced to MGARCH model by Bollerslev and Engle (1993), as well as succeeding work of Pan et al. (2010), Hu and Tsay (2014) and Li et al. (2016). At certain application scenarios when there is network structure behind the data we are interested in, multiple variables are connected and a multi-variate GARCH-type model could be fitted. These variations of MGARCH models solved the over-parameterization problem to some extent, but the number of parameters still expands along with the number of dimension nevertheless. With this flaw, MGARCH models could only be imposed on data with a small number of dimensions, such as stock indices of multiple markets or exchange rates of two currencies (see Karolyi, 1995 and Tse and Tsui, 2001).
Despite aforementioned difficulties due to dimensionality, for some specific types of multi-variate data where the connections between different components are actually observable, it is still possible to significantly simplify the model setup in the following aspects:
Instead of considering both volatilities and co-volatilities, we focus on studying the dynamics of volatilities only,
And instead of parameterizing every cross-individual effect, appropriate network structure can be embedded into the model.
In many cases such network structure can provide sufficient information about how the influence of pulses travel through edges between individual nodes. For instance, Nitzan and Libai (2011) found that customers connected with a defecting neighbor are 80% more likely to cancel their cellular service, and Goel and Goldstein (2014) concluded that the accuracy of individual behavior prediction can be significantly improved based on network data compared with conventional marketing practices.
Zhou et al. (2020) proposed a network GARCH model (see (2) for detailed specification of this model) that significantly reduces the parameterization complexity – the number of parameters remains fixed in their model no matter how large is the dimension . However, they did not fully utilize such an advantage as their discussion on parameter estimation is limited to the case when is fixed. Such setting narrows the variety of scenarios where their model could be applied since the size of network is often extremely large. In a study on social network that consists of 2982 users, Zhu et al. (2017) proposed a network AR model and the corresponding least squares estimator is proved to remain valid when the sample size and the dimension . Compared with their AR-type model, the unobserved volatility processes raise difficulties in extending such properties to GARCH-type models. We manage to address this problem by considering our network model as a spatial process on a two-dimensional lattice and adopting the asymptotic theorems for random fields proposed by Jenish and Prucha (2012) in the estimation of parameters. Since their limit theorems require NED, we will show such properties under certain restrictions on parameters and network structure. The idea of using limit theorems of spatial processes in the inference of high-dimensional time series has been considered by Xu et al. (2022) in the instrumental variable quantile regression estimation of their dynamic network quantile regression model. In this article, we will first introduce this idea into the estimation of high-dimensional GARCH-type models.
Aside from data with a diverging number of dimensions, we also aim to enable our model to handle data with asymmetry that was observed in empirical work such that positive and negative pulses affect volatilities differently in magnitude as well as in direction. While most GARCH-type models have an implicit assumption that the volatilities respond equally to the magnitude of positive and negative returns, Glosten et al. (1993) proposed a Glosten-Jagannathan-Runkle GARCH (GJR-GARCH) model with a threshold structure, allowing the volatility to act asymmetrically in magnitude responding to positive and negative pulses. The threshold GARCH (TGARCH) by Zakoïan (1994) also accommodates the asymmetry, but in magnitude of influence on conditional standard deviation. The exponential GARCH (EGARCH) of Nelson (1991) takes log transformation on the conditional variances, lifting the limitation of non-negative coefficients in conventional GARCH-type models and making it possible for their model to explain asymmetry in the direction of how volatility change corresponding to positive and negative news.
To study the asymmetrical dynamics in the volatilities of high-dimensional financial data, we propose a threshold network GARCH (TNGARCH) in Section 2. Stationarity conditions of this model are derived in Section 3.1 with fixed . In Section 3.2 we prove the -NED of proposed model under certain restrictions. The asymptotic properties of QMLE are investigated in Section 4, in the case when and at a lower rate. Then we propose a Wald statistic in Section 5.1, to test the existence of threshold effect, and in Section 5.2 we introduce a test for high-dimensional white noise proposed by Li et al. (2019). In Section 6, our methodology is tested on simulated data that are generated based on four different kinds of network structure. We observed an asymmetry that is different from existing literature, in how much the volatility responds to good news and bad news at individual level by applying our model to high-dimensional time series of log returns in Section 7. At last in Section 8, conclusions and potential directions for future research are summarized.
2 Model Setup
Consider an undirected and weightless network with nodes. Define the adjacency matrix , where if there is a connection between node and node , otherwise . Besides, self-connection is not allowed for any node by letting .
The connection can be defined differently with respect to practical scenarios, such as two social network accounts in a mutual followship, or two stocks who share at least one of top shareholders. As an interpretation of the network structure, is symmetric since , hence for any node , the out-degree is equal to the in-degree and we use to denote both for convenience.
For any node in this network, let be the observation at time , and be the unobservable conditional heteroscedasticity of , i.e. where denotes the -algebra consisting of all available information up to . A network GARCH(1, 1) specification of the conditional variance incorporates the network effect:
(2)
Model (2) indicates that the volatility of each stock , is influenced by not only its own previous change of price measured by , but also the average (with weight ) of for all node that are related to node . To ensure the positiveness of conditional variance, it is constrained that while .
To model the asymmetry in volatility, our TNGARCH model contains the threshold structure comparing with model (2). A TNGARCH(1,1) model is specified as follows:
(3)
where is the indicator function. To assure the positiveness of , the coefficients and are assumed to have the same constraints as in (2). is a white noise process satisfying the following assumption:
Assumption 1. is i.i.d. across and , with non-degenerate distribution, mean 0 and variance 1.
This assumption allows us to investigate, in the next section, the conditions for our model to have a unique strictly stationary solution, which serves as a precondition for further discussion on parameter estimation and statistical inference.
3 Stationarity and Near-Epoch Dependence
To derive the conditions under which model (3) is strictly stationary, we rewrite the conditional variance process in vector form
(4)
with notations as follows:
In Section 3.1, the stationarity of (4) when is a fixed number will be discussed. However, to estimate the parameters when , limit theorems based on stationarity and ergodicity of fixed-dimensional time series are not sufficient. Therefore, in Section 3.2 we will discuss the near-epoch dependence for a random field that supports the adoption of limit theorems for spatial processes in the subsequent sections.
3.1 Stationarity with N being Fixed
Since , is equivalent to . Hence
In this case, the random matrices are i.i.d. and model (4) is a generalized autoregressive equation by definition 1.4 in Bougerol and Picard (1992). It is easy to verify that . Therefore, the top Lyapunov exponent associated to is well-defined as follows:
(5)
where is an operator norm of matrices, corresponding to any norm on through
According to theorem 3.2 in Bougerol and Picard (1992), the series
(6)
is the unique strictly stationary and ergodic solution of model (4) if and only if the Lyapunov exponent . Under this condition, process is also strictly stationary and ergodic where since we could easily construct a continuous function according to (3) such that . Besides, since , the almost sure convergence of (6) guarantees that for any . Thus, with being an Euclidean norm.
By the subadditive ergodic theorem in Kingman (1973),
almost surely. In this case, could be approximated through computer simulation technique given a specific distribution of . For the purpose of reducing computation complexity, we derive a sufficient condition that is simple and much easier to verify.
Theorem 1.Under Assumption 1, model (4) has a unique strictly stationary and ergodic solution in the form (6) if
(7)
3.2 Near-Epoch Dependence for Random Fields
Let be a lattice on space , and measures the distance between any two locations . Assume we have observations from model (3) , then these observations could be regarded as triangular array of random fields with being a series of finite rectangular lattices . Then the growth of sample size is ensured by unbounded expansion of as . Such expansion is represented as , where is the cardinality of . The discussions on the asymptotic behaviors of random fields concern only the expansion of sample region, therefore the theoretical results derived in this section will apply as long as .
Let denote the -norm, i.e. for an arbitrary random variable . The definition of NED random fields is given as follows (see definition 1 in Jenish and Prucha, 2012):
Definition 1.A triangular array of random fields is said to be -NED () on if , and
where , is some non-negative sequence with , and is an array of finite positive constants.
Remark.If for some , then is said to be -NED on of size-; If for some , then is said to be -NED on geometrically; If , then is said to be uniformly -NED on random field . Note that geometric NED means NED of size- for all .
We need following assumptions before discussing the NED property of . Assumption 2 is needed to prove that ; Assumption 3 put restriction on the sparsity of the network: the power of connections between two nodes decays with their distance in case (a), or two nodes are only connected if they are sufficiently close in case (b). Similar restrictions on the network structure could also be seen in assumption 3 by Xu and Lee (2015) and assumption 3.2 by Xu et al. (2022).
Assumption 2.There exists , such that
Assumption 3.The row-normalized adjacency matrix satisfies one of following conditions:
(a).
for some ;
(b).
if for some constant , and otherwise.
Theorem 2.If condition (7) holds, under Assumptions 1, 2 and 3(a), is uniformly -NED on of size-, where the NED size is the constant in Assumption 3(a). Moreover, if Assumption 3(b) holds instead of 3(a), is uniformly and geometrically -NED on .
Remark.Note that , then Assumption 2 facilitates the -NED of 's given the -NED of 's. Besides, since , it is easy to verify that is a Lipschitz transformation of using mean value theorem. Then proposition 2 in Jenish and Prucha (2012) allows 's to inherit the NED properties from 's, therefore we could also verify that 's are also -NED.
4 Parameter Estimation
From model (3) we have observations with respect to true parameters . Based on the infinite past of observations, the quasi log-likelihood function is
Since the evaluation of the exact value of (8) is infeasible in practice, it is convenient to approximate (8) with
(9)
where is also generated from model (3) but with initial value . And the QMLE of is given by
To prove the asymptotic properties of , we need following assumptions aside from those required by Theorem 2:
Assumption 4. is a compact subset of such that all satisfy (7) and Assumption 2, and the true parameter is an interior point of .
Assumption 5. for some .
Assumption 6. for some , and following bounds exists:
where denotes the th component of parameter vector .
Assumption 7.The NED-size in Theorem 2 satisfies with being the one in Assumption 6.
Assumptions 4 is also required by Zhou et al. (2020) to prove the asymptotic properties in the case when being fixed. With both and , additional assumptions as above are required to adopt the limit theorems of random fields. Specifically, Assumption 5 is required for to satisfy the bound condition of law of large numbers (LLN) for random fields (assumption 2(a) in Jenish and Prucha, 2012); Assumption 6 facilitates the heredity of NED property from to the more complicated forms of first-order and second-order derivatives of ; Assumption 7 is a constraint on the decaying rate of NED coefficients, which is required by the central limit theorem (CLT) for random fields (assumption 4(c) in Jenish and Prucha, 2012). Of course, as we have remarked after Definition 1 that geometric NED means NED of size- for all , therefore Assumption 7 would be trivial under geometric NED.
Theorem 3.Under Assumptions required by Theorem 2, Assumption 4 and Assumption 5, the quasi-maximum likelihood estimator is consistent, i.e.
as ; If Assumptions 6 and 7 also hold, and the smallest eigenvalue
of
satisfies that
(10)
then is asymptotically normal as and :
where is the identity matrix.
Remark.Condition (10) can be implied if the smallest eigenvalues of satisfy that
As we will show in the proof of Proposition 5.1, and above could be approximated by
(11)
and
(12)
respectively. The later could be calculated recursively as where
5 Tests on Threshold Effect and Residuals
5.1 A Wald Test for the Threshold Effect
Given a null hypothesis
(13)
where is an matrix with rank and is an -dimensional vector, we could define a Wald test statistic as follows:
By the asymptotic normality of , could also be proved to follow a canonical asymptotic distribution as in Proposition 5.1.
Proposition 5.1.Under the same assumptions required by Theorem 3, as and , the Wald test statistic defined in (14) asymptotically follows a
distribution with degree of freedom , i.e.
5.2 A White Noise Test on the Residuals
There has been a large literature investigating high-dimensional time series models, including Xu and Lee (2015), Zhu et al. (2017) and Xu et al. (2022) among others, but none of them has used diagnostic tools to check the model adequacy. In this section, we will introduce a high-dimensional white noise test developed by Li et al. (2019) that can be applied to the diagnostic of high-dimensional models including ours.
Assume we have residuals , where . We want to test whether are high-dimensional white noises, i.e. there exists a matrix such that
(15)
where . The test statistic is the sum of squared singular values of first lagged sample autocovariance matrices:
(16)
where with when .
If is unknown, the sample covariance matrix of is . According to (2.8) in Li et al. (2019), we reject (15) if
where , and is the upper- quantile of standard normal distribution.
Note that being white-noise means that the residuals are uncorrelated over . However, it doesn't indicate that the residuals are uncorrelated over both and . The later indicates a stronger adequacy of high-dimensional model. We could assume that in the null hypothesis, and by (2.5) in Li et al. (2019), we reject if
6 Simulation Study
6.1 Network Simulation
The symmetric matrix in model 4 represents an undirected network structure, the pattern of which varies over different application scenarios. In this simulation study, we tend to use four different mechanisms of simulating corresponding network. The network structure in Example 1 adapts to Assumption 3, which is required by geometric NED as we have shown in Theorem 2. Simulation mechanisms introduced in Examples 2–4
are for testing the robustness of our estimation, against network structures that may violate Assumption 3.
Example 1.For each node , it is connected to node only if is inside 's -neighborhood. That is, in the adjacency matrix, if and otherwise. Figure 1(a) is a visualization of such a network with and .
Visualized network structures with N = 100. (a) Example 1 (D = 10) (b) Example 2; (c) Example 3; (d) Example 4 (K = 10)
Example 2. (Network structure with random distribution)For each node , we generate from uniform distribution , and then draw samples randomly from to form a set ( denotes the integer part of ). could be generated by letting if and otherwise. In a network simulated with such mechanism, as it is indicated in Figure 1(b), there is no significantly influential node (i.e. node with extremely large in-degree).
Example 3. (Network structure with power-law distribution)According to Clauset et al. (2009), for each node in such a network, is generated the same way as in Example 2. Instead of uniformly selecting samples from , these samples are collected w.r.t. probability where is generated from a discrete power-law distribution with scaling parameter . As shown in Figure 1(c), a few nodes have much larger in-degrees while most of them have less than 2. Compared to Example 2, network structure with power-law distribution exhibits larger gaps between the influences of different nodes. This type of network is suitable for modeling social media such as Twitter and Instagram, where celebrities have huge influence while the ordinary majority has little.
Example 4. (Network structure with stochastic blocks)As it was proposed in Nowicki and Snijders (2001), in a network with stochastic block structure, all nodes are divided into blocks and nodes from the same block are more likely to be connected comparing to those from different blocks. To simulate such structure, these nodes are randomly divided into groups by assigning labels to every nodes with equal probability. For any two nodes and from the same group, let while for those two from different groups, . Hence, it is very unlikely for nodes to be connected across groups. Our simulated network successfully mimics this characteristic as Figure 1(d) shows clear boundaries between groups. Block network also has its advantage in practical perspective. For instance, the price of one stock is highly relevant to those in the same industry sector.
In the next section, the simulation study is carried out on datasets that are generated according to the process (3) in conjunction with three types of adjacency matrices in Examples 1–4.
6.2 Simulation Results
Setting the true parameters as , we generate data according to process (3) with different sample sizes and number of dimensions . In our setting, increases from 50 to 4000, while also increases at relatively slower rates of and respectively, as it is showed in the following table:
50
100
200
500
800
1000
1500
2000
2500
3000
4000
7
10
14
22
28
31
38
44
50
54
63
12
21
37
80
119
144
205
263
319
374
482
For each combination of , datasets will be simulated independently, according to (3). Based on the th () dataset, the estimation of will be carried out and the estimation result is denoted as . For , the following two measurements are used to evaluate the performance of simulation results:
root-mean-square error: ,
coverage probability: .
is the 95% confidence interval defined as
where the estimated SE could be calculated as the square root of th diagonal element of and is the 0.975th quantile of standard normal distribution. To eliminate the effect of starting points, a different initial guess of is used for each .
As it is demonstrated in line graphs (c) and (d) in Figures 2-5, the consistency of the estimator is obvious since RMSE drops toward zero when and increases. Additionally, is proved to provide reliable estimation of true SE since the coverage probability (CP) converges to its theoretical value of 95% in graphs (a) and (b) in Figures 2-5. In conclusion, the asymptotic properties of our estimator in Theorem 3 are well supported by our simulation results, even for network structures in Examples 2–4 that may violate Assumption 3.
Remark.It is worth noticing that the CPs show a lower efficiency of convergence in general when , comparing with the case when . Such phenomenon raises an assumption that the performance of the estimator is highly related to the ratio of and . We repeat the simulation for 61 different combinations of and the scatter graph Figure 6 indicates that such assumption could be true, and the convergence of only requires , which includes what we have in Theorem 3, where and at a lower rate.
Coverage probabilities varies with the ratio of and
7 Empirical Data Analysis
In addition to simulation studies, we want to test our model using real data from Chinese Shanghai Stock Exchange (SSE) and Shenzhen Stock Exchange (SZSE). The dataset consists of daily log returns of 286 stocks, which are observed in two consecutive years of 2019 and 2020 ( except for closing days). These stocks come from four industry sectors as follows:
75 stocks from automotive industry sector;
73 stocks from financial industry sector;
68 stocks from information industry sector;
70 stocks from pharmaceutical industry sector.
And our model is tested within each sector, in which the number of stocks is approximately . Hence the estimates and inferences could be trusted according to the simulation study.
As an initial impression of data from each category the time plots of daily average log returns are presented in Figure 7. We also have the shareholder information of each stock, based on which two stocks are considered as connected when they share at least one common shareholder among their top ten shareholders. By this principle, four adjacency matrices are constructed and visualized as Figure 8 for four different industry sectors. Although it is quite intuitive to tell from Figure 8 the sparsity of these four networks, we tend to use the network density (ND) as a quantified measurement, which is defined by the ratio of the number of existing edges to the number of potential connections:
Average log returns of stocks from different industry sectors. (a) Automotive industry; (b) Financial industry; (c) Information industry; (d) Pharmaceutical industry
Visualization of networks for stocks from different industry sectors. (a) Automotive industry (ND = ); (b) Financial industry (ND = ); (c) Information industry (ND = ); (d) Pharmaceutical industry (ND = )
The results of parameter estimation is summarized in Table I. It is worth noting that the estimated network effect for automotive industry sector is not statistically significant while the other estimates for other coefficients or estimates from other sectors are all significant at 5% level. As indicated in Figure 8(a), this could be caused by the sparsity of the network structure as the data from automotive industry has the lowest network density comparing to others. Positive estimates of indicate positive correlation between the return of a stock and the returns of its neighbors. Comparing with other parameters, the estimates of are much larger for all four categories. Strong memory of volatility has been observed in many econometric studies on daily data, and such persistence would be stronger with data sampled at higher frequency according to Nelson (1991).
Table I.
Estimation results based on daily log-returns (2019&2020) of stocks from four industries.
Parameter
Estimation
SE
p-Value
Parameter
Estimation
SE
p-Value
Automotive industry
Financial industry
0.000099
5.83e07
0.05
0.000043
3.12e06
0.05
0.199408
1.08e02
0.05
0.247765
1.41e02
0.05
0.136423
1.01e02
0.05
0.202237
1.47e02
0.05
0.004591
4.71e03
0.16465
0.010469
5.35e03
0.05
0.727756
1.17e02
0.05
0.737272
1.09e02
0.05
Information industry
Pharmaceutical industry
0.000105
6.39e06
0.05
0.000063
4.15e06
0.05
0.172737
9.34e03
0.05
0.180950
1.05e02
0.05
0.122312
8.86e03
0.05
0.131722
1.06e02
0.05
0.009475
4.03e03
0.05
0.012929
4.06e03
0.05
0.745699
1.11e02
0.05
0.753305
1.11e02
0.05
We now conduct a Wald test on the existence of threshold effect based on the estimated parameters. By letting - and in (13), we can make a null hypothesis as follows:
As it is indicated in Table II, we could reject the null hypothesis with strong confidence and conclude that there exists extremely significant threshold effect within each industry sector.
Table II.
p-Values of Wald test on
Automotive industry
Financial industry
Information industry
Pharmaceutical industry
1.09e10
2.16e07
3.8e06
3.17e06
Using the diagnostic tool introduced in Section 5.2, we could check the model adequacy by inspecting the correlations between residual vectors . We will test null hypothesis with being unknown and respectively, the results are summarized in Table III. In all sectors, we cannot reject the hypothesis that the residual vectors are high-dimensional white noises with and over . However, the stronger hypothesis is rejected, as there exist correlations between residuals with different . We might be able to eliminate such deficiency in the adequacy of our model by heterogeneous parameterization with coefficients as . , , and , or by considering a dynamic network structure. However, the purpose of the introduction of network structure is to reduce the number of parameters of high-dimensional time series. Besides, deriving limit theorems for models with heterogeneous parameters or dynamic network could be theoretically challenging.
Table III.
Results of high-dimensional white noise test on with and
Automotive industry
Financial industry
Information industry
Pharmaceutical industry
is unknown
Not rejected
Not rejected
Not rejected
Not rejected
Rejected
Rejected
Rejected
Rejected
On the other hand, our results on asymmetric effect of positive and negative news are quite different compared to what was derived from univariate data in the literature. For instance, in a study by Engle and Ng (1993) on daily returns of Japanese stock index TOPIX, it was found that negative news would have larger impact on future volatility. Such phenomenon is reasonable in stock market since investors would lose confidence to a certain asset when it performs badly, hence they would adjust their portfolio and add more uncertainty to the future. However, it is not necessarily the case if we take into consideration the whole picture instead of looking at one individual and ignoring possible impact of its neighbors in the same system. In our estimation results, are uniformly larger than , indicating a larger impact of good news on volatility. A more precise conclusion would be that the volatility of one individual is more sensitive to its own good news, which actually does not contradict the conclusion of Engle and Ng (1993), since in the univariate case, how much proportion of the ‘bad news’ effect is actually contributed by bad performance in systematic perspective remains unknown. Our results show that good news has larger ‘local influence’ as it is indicated by , while there is a possibility that bad news, despite of having less ‘local influence’, spreads faster and has larger ‘global influence’ on the neighbors through network connection. Such potential leads to a future extension of our model that the threshold effect could be further applied on the coefficient , allowing good news and bad news to have asymmetric network effect.
8 Conclusion
In this article, we propose a TNGARCH model by taking consideration of network effect, as well as the threshold effect of shocks on volatilities. Our model can be applied to describe asymmetric properties for volatilities of high dimensional time series without increasing the parameterization complexity. Strict stationarity when is fixed, as well as near-epoch dependence when are discussed. Then the parameters are estimated by quasi-maximum likelihood estimation, the consistency and the asymptotic normality of the proposed estimator are proved as well. The results of simulation study support the theoretical properties of QMLE. At last, our model is fitted to real stock data containing 710 stocks of four industries from SSE and SZSE. Empirical results reveal that although volatility is more sensitive to bad news in univariate case, with network structure being considered, there is a possibility that majority of the revision of individual volatility is due to the impact of bad news of its neighbors, hence the ‘local influence’ of bad news is not necessarily larger than that of good news in such case.
There is room for extension of our methodology, which could lead to interesting topics for future research. In Theorem 3 we have derived asymptotic properties when and at a lower rate, our estimation method enables us to make reliable inference on parameters even when the data has hundreds of dimensions according to the simulation study. The limitation is also obvious, for example, as shown in Figures 2-5, to get a decent approximation of standard errors, we need to collect 4000 samples even though the number of dimensions is about 482, which could be even higher in real-world situations: user data collected from a social network often consists of millions of individual accounts whereas it may be impossible to collect sufficient number of samples over time even at daily frequency. Therefore our model would be applicable in a much larger scale if its statistical properties could be derived when increases at the same rate, or even higher rate compared to . As far as we know, there is no published work in the literature to solve this problem theoretically for GARCH-type models. Another limitation of our model is that the way we consider the network effect is simplified in two aspects: The network structure is deterministic rather than stochastic over time, embedding a random network in our model would make more sense, it would nevertheless raise the complexity of the model, and may cause problems in the estimation of parameters (see Chandrasekhar and Lewis, 2011); Moreover, there is only one type of individual-to-individual relation considered since the network structure is constructed solely based on common shareholders. Zhu et al. (2023) constructed their factor-augmented network using several types of relations, including individual-to-individual relations and factor-to-individual relations. Bringing more information into consideration would possibly improve the adequacy of the model and we will leave it for future research.
Acknowledgments
Many thanks to the Editor, the Co-Editor and the Referees for their insightful comments. The referees' suggestions lead to significant improvement of our article. The second author's work was partially supported by the National Natural Science Foundation of China (Grant No.12171161).
Our dataset consists of daily log returns of 286 stocks, which are observed in two consecutive years of 2019 and 2020 (except for holidays) from Chinese Shanghai Stock Exchange (SSE) and Shenzhen Stock Exchange (SZSE). The data are available at request. The data that support the findings of this study are available at request.
Appendix S1. The proofs for Theorem 1, Theorem 2, Theorem 3 and Proposition 5.1 are included in a separated supplementary material.
Please note: The publisher is not responsible for the content or functionality of any supporting information supplied by the authors. Any queries (other than missing content) should be directed to the corresponding author for the article.
References
Bauwens L, Laurent S, Rombouts JV. 2006. Multivariate GARCH models: a survey. Journal of Applied Econometrics21(1): 79–109.
Bollerslev T. 1990. Modelling the coherence in short-run nominal exchange rates: a multivariate generalized arch model. The Review of Economics and Statistics72(3): 498–505.
Engle R. 2002. Dynamic conditional correlation: a simple class of multivariate generalized autoregressive conditional heteroskedasticity models. Journal of Business & Economic Statistics20(3): 339–350.
Engle RF, Ng VK, Rothschild M. 1990. Asset pricing with a factor-ARCH covariance structure: empirical estimates for treasury bills. Journal of Econometrics45(1): 213–237.
Glosten LR, Jagannathan R, Runkle DE. 1993. On the relation between the expected value and the volatility of the nominal excess return on stocks. The Journal of Finance48(5): 1779–1801.
Karolyi GA. 1995. A multivariate GARCH model of international transmissions of stock returns and volatility: the case of the United States and Canada. Journal of Business & Economic Statistics13(1): 11–25.
Nowicki K, Snijders TAB. 2001. Estimation and prediction for stochastic block structures. Journal of the American Statistical Association96(455): 1077–1087.
Pan J, Polonik W, Yao Q. 2010. Estimating factor models for multivariate volatilities: an innovation expansion method. In Proceedings of COMPSTAT'2010, Physica-Verlag, Heidelberg; 305–314.
Please check your email for instructions on resetting your password.
If you do not receive an email within 10 minutes, your email address may not be registered,
and you may need to create a new Wiley Online Library account.
Request Username
Can't sign in? Forgot your username?
Enter your email address below and we will send you your username
If the address matches an existing account you will receive an email with instructions to retrieve your username
The full text of this article hosted at iucr.org is unavailable due to technical difficulties.