Department of Statistics and Finance, International Institute of Finance, School of Management, University of Science and Technology of China, Hefei 230026, China
Fanglin Bao is a graduate student under the tutelage of Prof. Bo Zhang at the University of Science and Technology of China. Her research mainly focuses on high-dimensional statistics
Bo Zhang is an Associate Professor at the University of Science and Technology of China (USTC). He obtained a Bachelor’s degree in Statistics from USTC in 2011, a Master’s degree in Probability Theory and Mathematical Statistics from USTC in 2013, and a Ph.D. degree in Mathematics from Nanyang Technological University in Singapore in 2017. His research mainly focuses on high-dimensional random matrices, high-dimensional time series, and complex networks
We consider the fluctuation of eigenvalues in factor models and propose a new method for testing the model. Based on the characteristics of eigenvalues, variables of unknown distribution are transformed into statistics of known distribution through randomization. The test statistic checks for breaks in the structure of factor models, including changes in factor loadings and increases in the number of factors. We give the results of simulation experiments and test the factor structure of the stock return data of China’s and U.S. stock markets from January 1, 2017, to December 31, 2019. Our method performs well in both simulations and real data.
Graphical Abstract
The randomization algorithm is used to construct the test statistics of the model structure.
Abstract
We consider the fluctuation of eigenvalues in factor models and propose a new method for testing the model. Based on the characteristics of eigenvalues, variables of unknown distribution are transformed into statistics of known distribution through randomization. The test statistic checks for breaks in the structure of factor models, including changes in factor loadings and increases in the number of factors. We give the results of simulation experiments and test the factor structure of the stock return data of China’s and U.S. stock markets from January 1, 2017, to December 31, 2019. Our method performs well in both simulations and real data.
Public Summary
Study the fluctuation characteristics of eigenvalues in factor models.
Construct the test statistic using a randomization algorithm.
Check for breaks in the structure of factor models, involving changes in factor loadings and increases in the number of factors.
Currently, people can easily obtain higher-dimensional and larger amounts of data. The factor model, a well-established technique for dimension reduction, is one of the most commonly used methods in high-dimensional research. For the theory related to factor models, see papers such as Refs. [1, 2]. With the wide application of factor models, related theories have also been developed, such as articles related to determining the number of factors in the model[3, 4]. These methods often involve the use of covariance matrices. Let N and T denote the cross-sectional and time series dimensions, respectively. When the dimension N is fixed while T tends to infinity, the eigenvalues of the sample covariance matrix are good estimates of the eigenvalues of the population covariance matrix. However, when both N and T tend to infinity, the estimation at this time is biased. For the study of the eigenvalues of the sample covariance matrix, there is a substantial body of work dealing with the limiting behavior of the eigenvalues. It forms an important subset of what is often called random matrix theory. Related research includes Refs. [5, 6].
In this paper, we consider spiked covariance matrices in the high-dimensional setting, which arise naturally from factor models. A special type of spiked covariance matrix whose nonspiked eigenvalues are equal to 1 is proposed by Ref. [7]. Subsequent studies based on the spiked matrix have been extended in various directions, including the asymptotic distribution of spiked eigenvalues (e.g., Refs. [8–12]). Related studies extend the scope to more generalized matrices, for example, by removing the restriction that all nonspiked eigenvalues equal 1. Based on the eigenvalues of the covariance matrix, Ref. [13] proposes a randomization method and constructs a test statistic to determine the number of factors in a factor model. Applications of this randomization method can be traced back to Ref. [14], and Ref. [15] also applies the method in the financial field. This randomization method is also the main method used in our article.
In a typical factor model, the whole sample is often used to construct a model suitable for the whole period. Time invariance of factor loadings is a standard assumption in the analysis of large factor models. However, such an assumption is often too ideal to suit practice. In reality, the factor structure may constantly change due to the influence of various factors. This study focuses on two types of changes: the instability of factor loadings and the fact that new factors emerge after the change point. Breaks in the factor structure have been explored in many papers. Refs. [16, 17] focus on structural instability in the loadings. Some articles consider the breaks in factor models before and after some significant events. Ref. [18] assesses the evidence for a change in the number of factors during the Great Recession. However, their approach requires prior knowledge of the breaking point and cannot distinguish the two change types we just mentioned. Studies also aim to find the exact time point when a break occurs, which we call a change point, with Ref. [19] being one example.
The problem is that when there is no structural change in the process, some detection methods may give a false result. To avoid such false test results, we need to test whether there are structural changes before conducting in-depth research, and when there are no structural changes, additional studies are unnecessary. In this article, we consider the fluctuation of the eigenvalues in the factor model and use the randomization method to construct a statistic to test whether there is a structural change in the model within the time range. Employing a sliding window, our technique is more sensitive to detecting changes in factor loadings than simply dividing the sample into two. The test statistic performs well in the absence of change points. The presence of a change point is consistently detected under both types of structural changes (loading instability and new factors).
Our research always assumes that √T/N=o(1). For any matrix A, ‖A‖ is the square root of the largest eigenvalue of A⊤A. The rest of the paper is organized as follows. In Section 2, we introduce the model and basic results, and Section 3 presents the theory on which the method is mainly based. Section 4 explains how we construct statistics using the randomization method. Then, in Section 5, simulation experiments are carried out to test the effect of the statistic. In addition, a real data example is reported in Section 6. Finally, we give our conclusions in Section 7. All technical proofs are presented in Appendix.
where {\boldsymbol{X}} and {\boldsymbol{E}} are two N \times T matrices, {\boldsymbol{A}} is an N \times r matrix, and {\boldsymbol{F}} is an r \times T matrix. Define T^{-1}\displaystyle\sum\limits_{t=1}^Tf_{kt}=\bar{f}_k.
Assumption 1. Both N and T go to infinity. {\boldsymbol{A}}^\top{\boldsymbol{A}}={\boldsymbol{I}}_r . Moreover, \dot{\lambda}_1 \geqslant \cdots \geqslant \dot{\lambda}_r\asymp N(1+NT^{-1}) such that
Let {\boldsymbol{M}} be a T \times T matrix that has elements M_{ij}={\boldsymbol{1}}\{i=j\}-T^{-1} . Define {\boldsymbol{\varLambda}}={\rm diag}(\dot{\lambda}_1,\cdots,\dot{\lambda}_r), then Eq. (3) implies that
Remark 1. Eq. (3) holds when f_{t}=\left(f_{1 t}, \cdots, f_{r t}\right) satisfies the following conditions: f_{t} is strictly stationary and \psi -mixing with \max _{1 \leqslant i \leqslant r} E\left(\lambda_{i}^{-2} f_{i t}^{4}\right) < C<\infty, \operatorname{Cov}\left(f_{t}\right)={\boldsymbol{\varLambda}} and \displaystyle \sum \nolimits _{t \geqslant 1} t \psi(t)^{1 / 2} < \infty. If the population covariance matrix is not diagonal, it is easy to adjust it by eigendecomposition and a different {\boldsymbol{A}} . However, Eq. (3) may still hold in some other cases even though f_{t} is a nonrandom time trend. Thus, we impose Assumption 1 on the samples, not the populations, mainly to cover a more extensive range of situations.
Assumption 2. {\boldsymbol{E}}={\boldsymbol{\varSigma^{1/2} Y \varOmega}} . Here, {\boldsymbol{Y}} has i.i.d. elements Y_{ij} with mean 0, variance 1, and a finite fourth moment. {\boldsymbol{\varSigma}} is an N \times N nonnegative definition matrix. {\boldsymbol{\varOmega}} is a \tilde{T} \times T matrix with \tilde{T} \asymp T . There exists a constant that does not depend on N and T such that \|{\boldsymbol{\varSigma}}\| \leqslant C and \|{\boldsymbol{\varOmega}}\|\leqslant C.
It also implies that \|T^{-1/2}{\boldsymbol{Y}}\|=O_p(1+N^{1/2}T^{-1/2}) and \|T^{-1/2}{\boldsymbol{E}}\|=O_p(1+N^{1/2}T^{-1/2}) . Let {\boldsymbol{S}}=T^{-1}{\boldsymbol{XMX^\top}} be the sample covariance matrix of {\boldsymbol{X}} . Let \lambda_1 \geqslant \cdots \geqslant \lambda_N \geqslant 0 be the eigenvalues of {\boldsymbol{S}} .
Proposition 1 is a simple result, but it provides spectrum separation. Let {\boldsymbol{U}}=({\boldsymbol{A}}\ {\boldsymbol{B}}) be an N \times N orthogonal matrix. Denote {\boldsymbol{(F+A^\top E)M}} as {\boldsymbol{\tilde{F}}} and {\boldsymbol{B^\top EM}} as {\boldsymbol{\tilde{E}}} .
for any 1\leqslant k \leqslant r. Here, {\boldsymbol{P_F}}={\boldsymbol{MF^\top}}({\boldsymbol{FMF}}^\top)^{-1}{\boldsymbol{FM}} .
Theorem 1 shows the fluctuation of \lambda_k from \dot{\lambda}_k for 1\leqslant k \leqslant r. Note that when {\boldsymbol{F}} and {\boldsymbol{E}} are independent, {\boldsymbol{A^\top EP_F}} degenerates to a separable model with rank r on both sides. Thus, \|{\boldsymbol{A^\top EP_F}}\|=O_p(1+N^{1/2}T^{-1/2}) holds. It can also extend to other cases where there is only weak dependence between {\boldsymbol{F}} and {\boldsymbol{E}} .
Remark 2. We note that \dot{\lambda}_k only depends on the population of {\boldsymbol{F}} . Eqs. (7) and (8) imply that the fluctuation of \lambda_k is small even if we replace {\boldsymbol{E}} by a different population satisfying Assumption 2. Thus, \lambda_k becomes large only if the population of {\boldsymbol{F}} becomes large.
Assumption 3. N/T \rightarrow c< \infty as N and T go to infinity, where c is away from 0. Let \mu_1 \geqslant \cdots \geqslant \mu_{N-r} \geqslant 0 be the eigenvalues of T^{-1}{\boldsymbol{B^\top EE^\top B}} . There exists a constant C such that
\tag{9} \sup\limits_{1 \leqslant k \leqslant K}|\mu_k-C|=O_p(T^{-2/3}),
where K is any finite integer.
Eq. (9) is a common result for nonspiked eigenvalues in random matrix theory, and it covers many cases. For example, the asymptotic distribution for the first K eigenvalues of sample covariance matrices with the general population has been stated under some moment constraints in Ref. [20]. Ref. [21] also established the Tracy–Widom asymptotic for the K largest eigenvalues of a general class of random Gram matrices. In fact, C is the right endpoint of the support for LSD of T^{-1}{\boldsymbol{B^\top EE^\top B}} . For more details, one can also refer to Refs. [6, 22, 23].
Theorem 2. Under Assumptions 1–3,
\tag{10} \sup\limits_{1 \leqslant k \leqslant K}|\lambda_{r+k}-C|=O_p(T^{-2/3}),
where K is any finite integer.
Theorem 2 demonstrates that excluding the first r largest eigenvalues, the remaining first K eigenvalues will be located within a small border around the right endpoint C . Although Assumption 3 covers the nonspiked cases in random matrix theory, there are other cases for {\boldsymbol{\tilde{E}}} with so-called “spiked eigenvalues”. Spiked eigenvalues often have a convergence rate of T^{-1/2} , which makes them differ from nonspiked eigenvalues. Addressing spiked eigenvalues in random matrix theory is sometimes easier than nonspiked eigenvalues, but the situation is different in our problems. The main reason is that the important terms in our proof of Theorem 2, such as {\boldsymbol{I-P_{\tilde{F}}}} and {\boldsymbol{I}}+T^{-2}{\boldsymbol{\tilde{E} \tilde{F}^\top}}{\boldsymbol{F_1}}(T^{-1}{\boldsymbol{\tilde{F}\tilde{F}^\top-\lambda{\boldsymbol{I}}}})^{-1}{\boldsymbol{\tilde{F}\tilde{E}^\top}} , may change the spiked structure.
We will consider a special case for {\boldsymbol{\tilde{E}}} with spiked eigenvalues.
Assumption 4. Let {\boldsymbol{F}} and {\boldsymbol{E}} be independent. N \asymp T . Moreover, {\boldsymbol{E}} can be rewritten as
where {\boldsymbol{Y}} has the i.i.d. elements Y_{ij} with mean 0, variance 1, and a finite fourth moment. {\boldsymbol{\varSigma}} is an N \times N matrix that is not negative definite, with eigenvalues \infty > \sigma_1 \geqslant \cdots \geqslant \sigma_N \geqslant 0.
Theorem 3. Under Assumptions 1 and 4, the spectrum of {\boldsymbol{B^\top\varSigma B}} and the distributions of Y_{ij} satisfy the conditions of Theorem 3.6 in Ref. [22]. Then, there exist constants c_1 \geqslant \cdots \geqslant c_K such that
\tag{12} \lambda_{r+k}-c_k=O_p(N^{-1})+O_p(T^{-1/2}), \ 1 \leqslant k \leqslant K.
Here, c_1 \geqslant \cdots \geqslant c_K depend only on {\boldsymbol{\varSigma}} .
Assumption 4 gives a situation with spiked eigenvalues in {\boldsymbol{\tilde{E}}} . In this case, (12) establishes a similar result to Theorem 2. The eigenvalues \lambda_{r+1},\dots,\lambda_{r+K} will still fluctuate in a small range near the limits.
3.
Fluctuation of eigenvalues
Now, we consider x_{it} with 1 \leqslant t \leqslant T^{*}. Define the N \times T matrix {\boldsymbol{X}}^{(s)} with elements x_{it} (1 \leqslant i \leqslant N,\ s \leqslant t \leqslant s+T-1) and {\boldsymbol{X}}^{(s)}={\boldsymbol{A}}^{(s)} {\boldsymbol{F}}^{(s)} +{\boldsymbol{E}}^{(s)} . For each s , we can calculate the r+1 largest eigenvalues of the sample covariance matrix for {\boldsymbol{X}}^{(s)} . We use \lambda_k^{(s)} to represent the k th eigenvalue of {\boldsymbol{X}}^{(s)} .
We do not impose any conditions on the correlations between {\boldsymbol{X}}^{(s)} for varying s . In contrast, we want to study whether differences exist between the {\boldsymbol{X}}^{(s)} structures based on the eigenvalues of their sample covariance matrices. We define {\boldsymbol{S}}^{(s)} and \lambda_1^{(s)} \geqslant \cdots \geqslant \lambda_N^{(s)} \geqslant 0 by replacing {\boldsymbol{X}} by {\boldsymbol{X}}^{(s)} in {\boldsymbol{S}} . Similarly, we can define {\boldsymbol{\varLambda}}^{(s)} by replacing {\boldsymbol{F}} by {\boldsymbol{F}}^{(s)} in {\boldsymbol{S}} . Denote that
Remark 3. If N is not large enough, we may consider a faster rate of \lambda_k^{(s)} . However, {\boldsymbol{A}}^{(s)} and {\boldsymbol{E}}^{(s)} impact \lambda_k^{(s)} in this case. Eq. (17) can have a faster rate provided that there are no differences (in distributions) between ({\boldsymbol{A}}^{(s)}, {\boldsymbol{E}}^{(s)}) .
We also consider an alternative hypothesis and result: Assume s_1 and s_2 exist s.t. {\boldsymbol{\varLambda}}^{(s_1)} \neq {\boldsymbol{\varLambda}}^{(s_2)} . Then, there may exist 1 \leqslant k \leqslant r such that the left part of Eq. (17) has a larger order than T^{-1/2}+N^{-1/2} .
Theorem 4 states that for the first r eigenvalues, the left part of Eq. (15) should be small and not fluctuate greatly when the structure has not changed. It is a natural conclusion drawn from Theorem 1.
Theorem 5. Suppose that Assumptions 1–3 hold for all {\boldsymbol{X}}^{(s)} . Then, according to Eq. (10),
\tag{18} \sup\limits_{1 \leqslant k \leqslant K}|\lambda_{r+k}^{(s)}-C^{(s)}|=O_p(T^{-2/3}),
where K is any finite integer. If C^{(s_1)}=C^{(s_2)} for \forall s_1 , s_2 with \ 1 \leqslant s_1 , s_2 \leqslant T^*-T+1, then
\tag{19} \dfrac{|\lambda_{r+k}^{(s)}-\bar{\lambda}_{r+k}|}{\bar{\lambda}_{r+k}}=O_p(T^{-2/3}), \ 1 \leqslant k \leqslant K.
Theorem 6. Suppose that the conditions of Theorem 3 hold for all {\boldsymbol{X}}^{(s)} . Then,
\tag{20} \lambda_{r+k}^{(s)}-c_k^{(s)}=O_p(N^{-1})+O_p(T^{-1/2}), \ 1 \leqslant k \leqslant K.
For \forall s_1 , s_2 with 1 \leqslant s_1, s_2 \leqslant T^*-T+1, if {\boldsymbol{\varSigma}}^{(s_1)} and {\boldsymbol{\varSigma}}^{(s_2)} have the same spectrum, then
\tag{21} \dfrac{|\lambda_{r+k}^{(s)}-\bar{\lambda}_{r+k}|}{\bar{\lambda}_{r+k}}=O_p(N^{-1}+T^{-1/2}), \ 1 \leqslant k \leqslant K.
Remark 4. With an alternative hypothesis, we have the following: If C^{(s_1)} \neq C^{(s_2)} or {\boldsymbol{\varSigma}}^{(s_1)} and {\boldsymbol{\varSigma}}^{(s_2)} have different spectra, then there may exist k such that the left part of (21) has a larger order than T^{-1/2}+N^{-1} .
Following Theorems 2 and 3, it is easy to obtain Theorems 5 and 6. We know that the left part of (21) should not fluctuate greatly for both the nonspiked and spiked cases when there is no break in the factor structure.
4.
Randomization algorithm
In Section 3, we consider only the results of orders. We can use the randomization algorithm in Ref. [13] to construct a new test statistic.
Based on Theorems 1–6, we construct the new statistic using
\tag{22} \varphi(s,k) = \dfrac{|\lambda_k^{(s)}-\bar{\lambda}_k|}{\bar{\lambda}_k} ,\;\; \phi=\max\limits_{1 \leqslant s \leqslant T^*-T+1, 1 \leqslant k \leqslant r_{\max}}{\varphi(s,k)}.
Here, r_{\max} is a value given in advance. It controls the number of possible factors in all sliding windows and is predetermined in the algorithm by a simple initial estimate: For {\boldsymbol{X}}^{(s)} in each sliding window, compute r_s= \mathop{\arg\max}\limits_{1\leqslant k\leqslant 19}\lambda_k^{(s)}/\lambda_{k+1}^{(s)} to estimate the number of factors for that window. When corresponding estimates are made for all windows, r_{\max} is chosen to be slightly larger than the maximum of \left\{r_s\right\}_{s=1}^{T^*-T+1} so that it can always include the r + 1 largest eigenvalues:
\tag{23} r_{\max}=\max\limits_{1 \leqslant s \leqslant T^*-T+1}{r_s}+1.
To detect whether there is a change point in 1\leqslant t \leqslant T^*, the null hypothesis and the alternative hypothesis can be set as follows:
Here, \phi is defined as Eq. (22). Then, the new statistic is constructed by
\tag{25} \varPhi=\exp{\{1/\phi\}}.
Since \phi has an order of T^{-1/2} under the null hypothesis, then \varPhi should diverge to infinity at a rate of \exp{\{T^{1/2}\}} . The randomization algorithm follows as in Ref. [24]:
Step 1. For \varPhi , generate an {\rm i.i.d.} sample \left\{\xi_{j}\right\}_{j=1}^{R} with common distribution G (the standard normal distribution).
Step 2. For any u drawn from a distribution F(u) with support U\subset R/\{0\} , define
\tag{26} \zeta_{j}(u)=I\left\{\xi_{j} \times \varPhi \leqslant u \right\}.
The considerations in Ref. [24] point toward choosing u=\pm \sqrt{2} . A relevant explanation of how the test statistics work under the null and alternative hypotheses is as follows. Under the null hypothesis, \varPhi will approach infinity as mentioned before; therefore, the \rm i.i.d. sequence \left\{\zeta_{j}(u)\right\}_{j=1}^{R} follows a Bernoulli distribution with E\left[\zeta_{j}(u)\right]= G(0) . Then in Step 3, \vartheta(u) should be N(0,1) because the central limit theorem should hold when R is large enough. Consequently, \varTheta in Step 4 should follow a chi-square distribution with one degree of freedom. Under the alternative hypothesis, \varPhi remains finite, so for any u \neq 0 , E\left[\zeta_{j}(u)\right] \neq G(0) . Naturally, a sum of {\rm i.i.d.} random variables with nonzero means is obtained in Step 3. Hence, \varTheta also diverges to infinity.
5.
Simulation
To verify the effect of the constructed statistic, repeated experiments were carried out under each group of conditions, where r represents the number of factors under the null hypothesis. Under {H}_0, we simulate data x_{it} with 1\leqslant i \leqslant N, 1\leqslant t \leqslant T^* according to the model:
where T^* means the whole process of data monitoring, and T is the chosen window width. We fix N=100 and T^*=800 for the whole process and set the change point \tau=500 . Additionally, we consider r \in \{1, 2, 3, 4\} factors and sliding window width T \in \{150,200,250,300\} .
Common factors are allowed to have some time dependence:
with e_t\stackrel{\text{i.i.d.}}{\sim}N(0,I_r) . We simulate each element of a_i as {\rm i.i.d.} \ N(0,1). The idiosyncratic components u=D \varepsilon G with each element of \varepsilon \;\; {\rm i.i.d.}\ N(0,1). D and G are two Toeplitz matrices of size N \times N and T^* \times T^* , respectively. The entries in the k th diagonal place are given by 0.3^{k - 1} for D and 0.5^{k - 1} for G . R=N is always chosen in the randomization step because the construction of the final statistic is based on the central limit theorem. The distribution G in Step 1 is set to be N(0,1) , while F in Step 2 is chosen to have nonzero and equal mass at \pm \sqrt{2} .
When the factor loadings have changed after the change point, the factor model can be expressed in the following form:
where each element of a and \widetilde{a}\stackrel{\text{i.i.d.}}{\sim}N(0,1) .
When using the traditional way to look at two samples separately, the eigenvalues reflect only the information of each one; they cannot reflect the change in loading. However, putting two samples together and calculating through the sliding window method can ensure the eigenvalues reflect that change. One of the examples is that when the common factors stay constant in (29), but there is such a relationship between the factor loadings {\boldsymbol{A_1=PA_2}} , {\boldsymbol{P}} is an orthogonal matrix. If we simply study the eigenvalues of the two covariance matrices, we find that they have almost the same eigenvalues, but at this time, the two loading matrices are obviously not equal. At this point, our sliding window method can identify such changes very well.
Consider the case where r_1 new factors appear after the change point, and the new factors satisfy: g_t= 0.7I_{r_1} g_{t-1}+v_t,\ 1\leqslant t\leqslant T^*, with v_t\stackrel{\text{i.i.d.}}{\sim}N(0,I_{r_1}) . The factor model at this time can be written as follows:
Elements of b_i are generated as {\rm i.i.d.}\ N (0,1).
To evaluate the performance of our statistic, we repeat the simulation 300 times under varying conditions and present the outcomes in Tables 1–4. Table 1 reports the fraction of false rejections under {H}_0 (i.e., when there is no break in the factor model) at the 1% and 5% significance level. In simulations, the sample size T^*=800 , and the window width T \in \{150,200,250,300\} .
Table
1.
False rejection at 1% and 5% significance ( T^*=800 ).
Under the null hypothesis, it can be seen from the data in Table 1 that the constructed test statistic can reach a good confidence level when a large window width T is chosen. Moreover, the performance of the test statistic is not affected by increases in the number of factors in the original model.
In Table 2, we show the fraction of detections for three types of structural changes when a break takes place and when testing at 1% and 5% significance levels. Under the alternative hypothesis, considering the changes in the factor loadings and the number of factors, the statistic can reject the null hypothesis with a probability of 1, which indicates that it can test the changes in the structure very well. At the same time, the results in Table 2 show that it can have strong power under the alternative assumptions of different factor numbers in the original model and newly added factor numbers.
Considering a smaller sample size T^*=200 with change point \tau=100 , Table 3 shows the false rejection rate under H_0 of selecting the window width T \in \{100,125,150\} . According to the results in the table, even if T^* is reduced to 200, our method can still be used when window width T = 125 . This suggests that a larger window width can bring a better size of the test. However, it is not appropriate to select a window width T that is too large for the whole sample size T^* . Table 4 shows the fraction of detections under the alternative hypothesis when testing at 1% and 5% significance levels. Alghouth Table 3 shows that it has a better size with window width T=150 , the power in Table 4 is too low for the test.
6.
Real data
Taking the daily returns of the S&P 500 and CSI 300 constituent stocks in China’s and U.S. stock markets, respectively, from January 1, 2017, to December 31, 2019, as the research object, we examine the factor structure of the stock markets of the two countries. Stocks that have been suspended for more than 60 consecutive days are eliminated, and the remaining missing stock return data are filled with 0. Aggregating the data according to the same opening date for both markets, we obtain the daily returns of 749 stocks for 708 trading days.
We use the randomization algorithm in Section 4 to test the factor structure of the above data. The window width is selected as T = 200 , and the final test statistic value is \varTheta = 58.94 , which exceeds the critical value c_{\alpha}= 6.635 when the significance level is \alpha = 0.01 . This suggests that there are changes in the factor structure during this period. To show the result more clearly, Fig. 1 presents the value of \lambda_1^{(s)} in each sliding window {\boldsymbol{X}}^{(s)} . In the whole process, \lambda_1^{(s)} changes greatly with the movement of the window, which also shows that the factor structure has changed.
Figure
1.
The largest eigenvalue \lambda_1^{(s)} in each sliding window (for 1 \leqslant s\leqslant 509).
The selected data include the period of the U.S.-China trade war, during which many important trade war events occurred, including multiple mutual taxation and meeting negotiations between China and the United States. The trade relationship between these two countries has experienced huge volatility, and stock markets have also fluctuated accordingly. Hence, the test result is consistent with the actual situation.
7.
Conclusions
We study the fluctuation of eigenvalues in factor models and construct a test statistic of the model structure. One only need to detect the change points after the test shows the existence of change points. It reduces unnecessary work. We also apply our method to simulations and the real data from China’s and U.S. stock markets. Our method performs well in both simulations and real data.
At the same time, the sliding window method also avoids the insensitivity of the method of simply comparing two samples for eigenvalue research to determine the change in factor loadings. Our proposed test statistic performs well under both types of structural changes (the instability of factor loadings and the fact that new factors emerge after the change point) and is not affected by the number of factors in the original model. Accordingly, this novel statistic has good universality for application in future studies.
Acknowledgements
We are grateful to reviewers for their careful reading and helpful remarks. This work was supported by the National Natural Science Foundation of China (12001517, 72091212), the USTC Research Funds of the Double First-Class Initiative (YD2040002005), and the Fundamental Research Funds for the Central Universities (WK2040000026, WK2040000027).
Conflict of interest
The authors declare that they have no conflict of interest.
Study the fluctuation characteristics of eigenvalues in factor models.
Construct the test statistic using a randomization algorithm.
Check for breaks in the structure of factor models, involving changes in factor loadings and increases in the number of factors.
Bai J. Inferential theory for factor models of large dimensions. Econometrica,2003, 71 (1): 135–171. DOI: 10.1111/1468-0262.00392
[2]
Bai J, Li K. Statistical analysis of factor models of high dimension. The Annals of Statistics,2012, 40 (1): 436–465. DOI: 10.1214/11-AOS966
[3]
Bai J, Ng S. Determining the number of factors in approximate factor models. Econometrica,2002, 70 (1): 191–221. DOI: 10.1111/1468-0262.00273
[4]
Onatski A. Determining the number of factors from empirical distribution of eigenvalues. The Review of Economics and Statistics,2010, 92 (4): 1004–1016. DOI: 10.1162/REST_a_00043
[5]
El Karoui N. Spectrum estimation for large dimensional covariance matrices using random matrix theory. The Annals of Statistics,2008, 36 (6): 2757–2790. DOI: 10.1214/07-AOS581
[6]
Bao Z, Pan G, Zhou W. Universality for the largest eigenvalue of sample covariance matrices with general population. The Annals of Statistics,2015, 43 (1): 382–421. DOI: 10.1214/14-AOS1281
[7]
Johnstone I M. On the distribution of the largest eigenvalue in principal components analysis. The Annals of Statistics,2001, 29 (2): 295–327. DOI: 10.1214/aos/1009210544
[8]
Baik J, Silverstein J W. Eigenvalues of large sample covariance matrices of spiked population models. Journal of Multivariate Analysis,2006, 97 (6): 1382–1408. DOI: 10.1016/j.jmva.2005.08.003
[9]
Paul D. Asymptotics of sample eigenstructure for a large dimensional spiked covariance model. Statistica Sinica,2007, 17 (4): 1617–1642.
[10]
Bai Z, Yao J. On sample eigenvalues in a generalized spiked population model. Journal of Multivariate Analysis,2012, 106: 167–177. DOI: 10.1016/j.jmva.2011.10.009
[11]
Wang W, Fan J. Asymptotics of empirical eigenstructure for high dimensional spiked covariance. The Annals of Statistics,2017, 45 (3): 1342–1374. DOI: 10.1214/16-AOS1487
[12]
Cai T T, Han X, Pan G. Limiting laws for divergent spiked eigenvalues and largest nonspiked eigenvalue of sample covariance matrices. The Annals of Statistics,2020, 48 (3): 1255–1280. DOI: 10.1214/18-AOS1798
[13]
Trapani L. A randomized sequential procedure to determine the number of factors. Journal of the American Statistical Association,2018, 113 (523): 1341–1349. DOI: 10.1080/01621459.2017.1328359
[14]
Pearson E S. On questions raised by the combination of tests based on discontinuous distributions. Biometrika,1950, 37: 383–398. DOI: 10.1093/biomet/37.3-4.383
[15]
Corradi V, Swanson N R. The effect of data transformation on common cycle, cointegration, and unit root tests: Monte Carlo results and a simple test. Journal of Econometrics,2006, 132 (1): 195–229. DOI: 10.1016/j.jeconom.2005.01.028
[16]
Chen L, Dolado J J, Gonzalo J. Detecting big structural breaks in large factor models. Journal of Econometrics,2014, 180 (1): 30–48. DOI: 10.1016/j.jeconom.2014.01.006
[17]
Cheng X, Liao Z, Schorfheide F. Shrinkage estimation of high-dimensional factor models with structural instabilities. The Review of Economic Studies,2016, 83 (4): 1511–1543. DOI: 10.1093/restud/rdw005
[18]
Stock J H, Watson M W. Disentangling the channels of the 2007–2009 recession. Cambridge, MA: National Bureau of Economic Research, 2012: 18094.
[19]
Breitung J, Eickmeier S. Testing for structural breaks in dynamic factor models. Journal of Econometrics,2011, 163 (1): 71–84. DOI: 10.1214/17-AAP1341
[20]
Ding X, Yang F. A necessary and sufficient condition for edge universality at the largest singular values of covariance matrices. The Annals of Applied Probability,2018, 28 (3): 1679–1738. DOI: 10.1214/17-AAP1341
[21]
Ding X, Yang F. Tracy–Widom distribution for heterogeneous Gram matrices with applications in signal detection. IEEE Transactions on Information Theory,2022, 68 (10): 6682–6715. DOI: 10.1109/TIT.2022.3176784
[22]
Ding X, Yang F. Spiked separable covariance matrices and principal components. The Annals of Statistics,2021, 49 (2): 1113–1138. DOI: 10.1214/20-AOS1995
[23]
Knowles A, Yin J. Anisotropic local laws for random matrices. Probability Theory and Related Fields,2017, 169 (1): 257–352. DOI: 10.1007/s00440-016-0730-4
[24]
Barigozzi M, Trapani L. Sequential testing for structural stability in approximate factor models. Stochastic Processes and Their Applications,2020, 130 (8): 5149–5187. DOI: 10.1016/j.spa.2020.03.003
Figure
1.
The largest eigenvalue \lambda_1^{(s)} in each sliding window (for 1 \leqslant s\leqslant 509).
References
[1]
Bai J. Inferential theory for factor models of large dimensions. Econometrica,2003, 71 (1): 135–171. DOI: 10.1111/1468-0262.00392
[2]
Bai J, Li K. Statistical analysis of factor models of high dimension. The Annals of Statistics,2012, 40 (1): 436–465. DOI: 10.1214/11-AOS966
[3]
Bai J, Ng S. Determining the number of factors in approximate factor models. Econometrica,2002, 70 (1): 191–221. DOI: 10.1111/1468-0262.00273
[4]
Onatski A. Determining the number of factors from empirical distribution of eigenvalues. The Review of Economics and Statistics,2010, 92 (4): 1004–1016. DOI: 10.1162/REST_a_00043
[5]
El Karoui N. Spectrum estimation for large dimensional covariance matrices using random matrix theory. The Annals of Statistics,2008, 36 (6): 2757–2790. DOI: 10.1214/07-AOS581
[6]
Bao Z, Pan G, Zhou W. Universality for the largest eigenvalue of sample covariance matrices with general population. The Annals of Statistics,2015, 43 (1): 382–421. DOI: 10.1214/14-AOS1281
[7]
Johnstone I M. On the distribution of the largest eigenvalue in principal components analysis. The Annals of Statistics,2001, 29 (2): 295–327. DOI: 10.1214/aos/1009210544
[8]
Baik J, Silverstein J W. Eigenvalues of large sample covariance matrices of spiked population models. Journal of Multivariate Analysis,2006, 97 (6): 1382–1408. DOI: 10.1016/j.jmva.2005.08.003
[9]
Paul D. Asymptotics of sample eigenstructure for a large dimensional spiked covariance model. Statistica Sinica,2007, 17 (4): 1617–1642.
[10]
Bai Z, Yao J. On sample eigenvalues in a generalized spiked population model. Journal of Multivariate Analysis,2012, 106: 167–177. DOI: 10.1016/j.jmva.2011.10.009
[11]
Wang W, Fan J. Asymptotics of empirical eigenstructure for high dimensional spiked covariance. The Annals of Statistics,2017, 45 (3): 1342–1374. DOI: 10.1214/16-AOS1487
[12]
Cai T T, Han X, Pan G. Limiting laws for divergent spiked eigenvalues and largest nonspiked eigenvalue of sample covariance matrices. The Annals of Statistics,2020, 48 (3): 1255–1280. DOI: 10.1214/18-AOS1798
[13]
Trapani L. A randomized sequential procedure to determine the number of factors. Journal of the American Statistical Association,2018, 113 (523): 1341–1349. DOI: 10.1080/01621459.2017.1328359
[14]
Pearson E S. On questions raised by the combination of tests based on discontinuous distributions. Biometrika,1950, 37: 383–398. DOI: 10.1093/biomet/37.3-4.383
[15]
Corradi V, Swanson N R. The effect of data transformation on common cycle, cointegration, and unit root tests: Monte Carlo results and a simple test. Journal of Econometrics,2006, 132 (1): 195–229. DOI: 10.1016/j.jeconom.2005.01.028
[16]
Chen L, Dolado J J, Gonzalo J. Detecting big structural breaks in large factor models. Journal of Econometrics,2014, 180 (1): 30–48. DOI: 10.1016/j.jeconom.2014.01.006
[17]
Cheng X, Liao Z, Schorfheide F. Shrinkage estimation of high-dimensional factor models with structural instabilities. The Review of Economic Studies,2016, 83 (4): 1511–1543. DOI: 10.1093/restud/rdw005
[18]
Stock J H, Watson M W. Disentangling the channels of the 2007–2009 recession. Cambridge, MA: National Bureau of Economic Research, 2012: 18094.
[19]
Breitung J, Eickmeier S. Testing for structural breaks in dynamic factor models. Journal of Econometrics,2011, 163 (1): 71–84. DOI: 10.1214/17-AAP1341
[20]
Ding X, Yang F. A necessary and sufficient condition for edge universality at the largest singular values of covariance matrices. The Annals of Applied Probability,2018, 28 (3): 1679–1738. DOI: 10.1214/17-AAP1341
[21]
Ding X, Yang F. Tracy–Widom distribution for heterogeneous Gram matrices with applications in signal detection. IEEE Transactions on Information Theory,2022, 68 (10): 6682–6715. DOI: 10.1109/TIT.2022.3176784
[22]
Ding X, Yang F. Spiked separable covariance matrices and principal components. The Annals of Statistics,2021, 49 (2): 1113–1138. DOI: 10.1214/20-AOS1995
[23]
Knowles A, Yin J. Anisotropic local laws for random matrices. Probability Theory and Related Fields,2017, 169 (1): 257–352. DOI: 10.1007/s00440-016-0730-4
[24]
Barigozzi M, Trapani L. Sequential testing for structural stability in approximate factor models. Stochastic Processes and Their Applications,2020, 130 (8): 5149–5187. DOI: 10.1016/j.spa.2020.03.003