Suppose that we model our data as
If we split our data into two groups, then we have
and
The null hypothesis of the Chow test asserts that a 1 = a 2 {\displaystyle a_{1}=a_{2}} , b 1 = b 2 {\displaystyle b_{1}=b_{2}} , and c 1 = c 2 {\displaystyle c_{1}=c_{2}} , and there is the assumption that the model errors ε {\displaystyle \varepsilon } are independent and identically distributed from a normal distribution with unknown variance.
Let S C {\displaystyle S_{C}} be the sum of squared residuals from the combined data, S 1 {\displaystyle S_{1}} be the sum of squared residuals from the first group, and S 2 {\displaystyle S_{2}} be the sum of squared residuals from the second group. N 1 {\displaystyle N_{1}} and N 2 {\displaystyle N_{2}} are the number of observations in each group and k {\displaystyle k} is the total number of parameters (in this case 3, i.e. 2 independent variables coefficients + intercept). Then the Chow test statistic is
The test statistic follows the F-distribution with k {\displaystyle k} and N 1 + N 2 − 2 k {\displaystyle N_{1}+N_{2}-2k} degrees of freedom.
The same result can be achieved via dummy variables.
Consider the two data sets which are being compared. Firstly there is the 'primary' data set i={1,..., n 1 {\displaystyle n_{1}} } and the 'secondary' data set i={ n 1 {\displaystyle n_{1}} +1,...,n}. Then there is the union of these two sets: i={1,...,n}. If there is no structural change between the primary and secondary data sets a regression can be run over the union without the issue of biased estimators arising.
Consider the regression:
y t = β 0 + β 1 x 1 t + β 2 x 2 t + . . . + β k x k t + γ 0 D t + ∑ i = 1 k γ i x i t D t + ε t . {\displaystyle y_{t}=\beta _{0}+\beta _{1}x_{1t}+\beta _{2}x_{2t}+...+\beta _{k}x_{kt}+\gamma _{0}D_{t}+\sum _{i=1}^{k}\gamma _{i}x_{it}D_{t}+\varepsilon _{t}.\,}
Which is run over i={1,...,n}.
D is a dummy variable taking a value of 1 for i={ n 1 {\displaystyle n_{1}} +1,...,n} and 0 otherwise.
If both data sets can be explained fully by ( β 0 , β 1 , . . . , β k ) {\displaystyle (\beta _{0},\beta _{1},...,\beta _{k})} then there is no use in the dummy variable as the data set is explained fully by the restricted equation. That is, under the assumption of no structural change we have a null and alternative hypothesis of:
H 0 : γ 0 = 0 , γ 1 = 0 , . . . , γ k = 0 {\displaystyle H_{0}:\gamma _{0}=0,\gamma _{1}=0,...,\gamma _{k}=0}
H 1 : otherwise {\displaystyle H_{1}:{\text{otherwise}}}
The null hypothesis of joint insignificance of D can be run as an F-test with n − 2 ( k + 1 ) {\displaystyle n-2(k+1)} degrees of freedom (DoF). That is: F = ( R S S R − R S S U ) / ( k + 1 ) R S S U / D o F {\displaystyle F={\frac {(RSS^{R}-RSS^{U})/(k+1)}{RSS^{U}/DoF}}} .
Remarks