回归分析
2022-05-10 21:22:36 0 举报AI智能生成
线性基本模型 线性回归与方差分析---王松桂老师版本
统计
检验流程图
假设检验
模版推荐
作者其他创作
大纲/内容
二、随机向量
2.1均值向量与协方差阵
定理2.1.1
Y=AX+b,则<span class="equation-text" contenteditable="false" data-index="0" data-equation="E(Y)=AE(X)+E(b)"><span></span><span></span></span>
推论:<span class="equation-text" contenteditable="false" data-index="0" data-equation="trCov(X)=\sum_{i=1}^n Var(X_i)"><span></span><span></span></span>
定理2.1.2
<b>协方差阵半正定,而且对称!</b>
证半正定思路:令<span class="equation-text" contenteditable="false" data-index="0" data-equation="Y=c{'}X"><span></span><span></span></span>,<span class="equation-text" contenteditable="false" data-index="1" data-equation="Var(Y)=c^{'}Cov(X)c>0\Longrightarrow Cov(X)>0"><span></span><span></span></span>
定理2.1.3
<font color="#ff0000">Y=AX,则</font><span class="equation-text" data-index="0" data-equation="Cov(Y)=ACov(X)A^{'}"><font color="#ff0000"></font></span><font color="#000000"><b>(2.3节的定理说的都是这个事儿</b>)</font>
定理2.1.4
<span class="equation-text" contenteditable="false" data-index="0" data-equation="Cov(AX,BY)=ACov(X,Y)B^{'}"><span></span><span></span></span>
2.2随机向量的二次型
定义
<span class="equation-text" contenteditable="false" data-index="0" data-equation="X^{'}AX=\sum_{i=1}^n \sum_{j=1}^n a_{ij}X_{i}Y_{j}"><span></span><span></span></span>
<font color="#000000">定理2.2.1</font>
<b><font color="#ff0000">设<span class="equation-text" data-index="0" data-equation="E(X)=\mu,Cov(X)=\sum,则E(X^{'}AX)=\mu^{'}A\mu+tr(A\sum)" contenteditable="false"><span></span><span></span></span></font></b>
证明思路:①<span class="equation-text" contenteditable="false" data-index="0" data-equation="X^{'}AX=(X-\mu+\mu)^{'}A(X-\mu+\mu)=(X-\mu)^{'}A(X-\mu)+\mu^{'}A(X-\mu)+(X-\mu)^{'}A\mu+\mu^{'}A\mu"><span></span><span></span></span><br>②由定理2.1.1,得<span class="equation-text" contenteditable="false" data-index="1" data-equation="E[\mu^{'}A(X-\mu)]=E(\mu^{'}AX)-\mu^{'}A\mu=0"><span></span><span></span></span>,同理,第三项也为0<br>③<span class="equation-text" contenteditable="false" data-index="2" data-equation="E[(X-\mu)^{'}A(X-\mu)]=E[trA(X-\mu)(X-\mu)^{'}]"><span></span><span></span></span>(二次型的迹就是他本身)<br> <span class="equation-text" contenteditable="false" data-index="3" data-equation="=trE[A(X-\mu)(X-\mu)^{'}]"><span></span><span></span></span>(迹是求和运算,利用了期望的线性)<br> <span class="equation-text" contenteditable="false" data-index="4" data-equation="=trAE[(X-\mu)(X-\mu)^{'}]"><span></span><span></span></span>(trAB=trBA)<br> <span class="equation-text" contenteditable="false" data-index="5" data-equation="=tr(A\sum)"><span></span><span></span></span>
推论:<span class="equation-text" contenteditable="false" data-index="0" data-equation="若\mu=0,\sum=I,则E(X^{'}AX)=tr(A)"><span></span><span></span></span>
2.3正态随机变量
三个等价定义
(从密度函数上)<span class="equation-text" data-index="0" data-equation="f(x)=(2\pi)^{\frac{-n}{2}}(det\sum)^{-1\over2}e^{{-1\over2}(x-\mu)^{'}(x-\mu)}" contenteditable="false"><span></span><span></span></span>+<font color="#ff0000"><span class="equation-text" data-index="1" data-equation="\sum>0" contenteditable="false"><span></span><span></span></span>(非退化)</font>
(从正态性质上)若存在列满秩矩阵<span class="equation-text" contenteditable="false" data-index="0" data-equation="A_{n\times r}"><span></span><span></span></span>,<span class="equation-text" contenteditable="false" data-index="1" data-equation="X=Au+\mu,其中\mu=(\mu_1,\mu_2....\mu_r)^{'}\sim N(0,1)"><span></span><span></span></span>,则<span class="equation-text" contenteditable="false" data-index="2" data-equation="X\sim N_n(\mu,\sum=AA^{'})"><span></span><span></span></span>
(从特征函数上)<span class="equation-text" contenteditable="false" data-index="0" data-equation="X\sim N_n(\mu,\sum)\Longleftrightarrow\phi_x(t)=exp\left\{it^{'}\mu-\frac{t^{'}\sum t}{2}\right\}"><span></span><span></span></span>
定理2.3.1
对正态向量而言,相互独立与不相关是等价的(<font color="#ff0000">不相关一般推不出来独立</font>)
定理2.3.2
<span class="equation-text" contenteditable="false" data-index="0" data-equation="X_n\sim N(\mu,\sum),Y=AX+b,则Y\sim N(A\mu+b,A\sum A^{'})"><span></span><span></span></span>
推论1:设<span class="equation-text" data-index="0" data-equation="X\sim N (\mu,\sum),则Y=\sum^{-1\over2}X\sim N(\sum^{-1\over2}\mu,I)" contenteditable="false"><span></span><span></span></span>(<b>A特殊化了</b>)
推论2:设<span class="equation-text" data-index="0" data-equation="X\sim N_n(\mu,\sigma^2I),Q_n为正交阵,QX\sim N_n(Qu,\sigma^2I)" contenteditable="false"><span></span><span></span></span>(<b><span class="equation-text" data-index="1" data-equation="\sum" contenteditable="false"><span></span><span></span></span>特殊化了</b>)
定理2.3.3
<span class="equation-text" data-index="0" data-equation="X \sim N_n(\mu,\sum),X=(X_1,X_2)^{'},\mu=(\mu_1,\mu_2)^{'},\sum=\begin{vmatrix}\sum_{11} & \sum_{12} \\\sum_{21} & \sum_{22}\end{vmatrix}" contenteditable="false"><span></span><span><font color="#ff0000"></font></span></span>,<font color="#ff0000">则<span class="equation-text" data-index="1" data-equation="X_1=N_m(\mu_1,\sum_{11})" contenteditable="false"><span></span><span></span></span></font>
联合正态<span class="equation-text" contenteditable="false" data-index="0" data-equation="\Longrightarrow"><span></span><span></span></span>边际正态,反之不成立
定理2.3.4
<span class="equation-text" contenteditable="false" data-index="0" data-equation="X\sim N_n(\mu,\sum),r(A_{m\times n})=m(<n),则Y=AX\sim N_m(A\mu,A\sum A^{'})"><span></span><span></span></span>
2.4<span class="equation-text" contenteditable="false" data-index="0" data-equation="\chi^2"><span></span><span></span></span>分布
定理2.4.1
<span class="equation-text" contenteditable="false" data-index="0" data-equation="X\sim N_n(0,\sum),\sum正定,则X^{'}\sum^{-1}X\sim \chi_n^2"><span></span><span></span></span>
定理2.4.2
<span class="equation-text" contenteditable="false" data-index="0" data-equation="X\sim \chi^2_n,则E(X)=n.Var(X)=2n"><span></span><span></span></span>
定理2.4.3
可加性
<font color="#ff0000">定理2.4.4</font>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="X\sim N_n(0,I_n)"><span></span><span></span></span>,A为对称阵,r(A)=r,当A幂等时,二次型<span class="equation-text" contenteditable="false" data-index="1" data-equation="X^{'}AX\sim \chi^2_r"><span></span><span></span></span>
证明思路:①A幂等,r(A)=r<span class="equation-text" data-index="0" data-equation="\Longrightarrow" contenteditable="false"><span></span><span></span></span>A的特征值中有r个为1,n-r个为0<br>②A对称阵<span class="equation-text" contenteditable="false" data-index="1" data-equation="\Longrightarrow \exist正交阵Q,A=Q^{'}\begin{pmatrix}I_r & 0 \\0 & 0\end{pmatrix}Q"><span></span><span></span></span>(正交谱分解)<br>③令Y=QX,则<span class="equation-text" contenteditable="false" data-index="2" data-equation="Y\sim N(0,I_n)"><span></span><span></span></span>,且Y的各个分量独立<br><span class="equation-text" contenteditable="false" data-index="3" data-equation="X{'}AX=X{'}Q{'}\begin{pmatrix}I_r & 0 \\0 & 0\end{pmatrix}QX=Y_1^2+……+Y_r^2\sim \chi^2_r"><span></span><span></span></span>
<font color="#ff0000"> 定理2.4.5<br>(判别独立的另一方法)</font><br>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="X\sim N_n(0,I_n)"><span></span><span></span></span>,A为n阶对称阵,若BA=0,则BX与<span class="equation-text" contenteditable="false" data-index="1" data-equation="X^{'}AX"><span></span><span></span></span>相互独立
<span style="color: rgb(255, 0, 0); font-size: inherit;">定理2.4.6</span><br>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="X\sim N_n(0,I),"><span></span><span></span></span>A和B皆为对称阵,且AB=0,则<span class="equation-text" contenteditable="false" data-index="1" data-equation="X^{'}AX与X^{'}BX相互独立"><span></span><span></span></span>
2.5 关于正交
回顾
<span class="equation-text" contenteditable="false" data-index="0" data-equation="X_i\sim N(\mu,\sigma^2),iid,"><span></span><span></span></span>则<span class="equation-text" contenteditable="false" data-index="1" data-equation="\bar{X}"><span></span><span></span></span>与<span class="equation-text" contenteditable="false" data-index="2" data-equation="S^2"><span></span><span></span></span>独立
引理
x与y正交,则g(x)与h(y)正交(非随机变换)
定理2.5.1
<span class="equation-text" data-index="0" data-equation="X\sim N_n(\mu,I),A^{'}=A,CA=0,则CX\bot X^{'}AX" contenteditable="false"><span></span><span></span></span>(<b>正态情形,也可以说独立</b>)
推论:<span class="equation-text" data-index="0" data-equation="X\sim N_n(\mu,\sum),\sum>0,A^{'}=A,C\sum A=0,则CX\bot X^{'}AX" contenteditable="false"><span></span><span></span></span>(<b>协方差阵一般化了</b>)
定理2.5.2
<span class="equation-text" data-index="0" data-equation="X\sim N_n(\mu,I),A^{'}=A,B^{'}=B,AB=0,则X^{'}BX\bot X^{'}AX" contenteditable="false"><span></span><span></span></span>(<b>正态情形,也可以说独立</b>)
推论:<span class="equation-text" data-index="0" data-equation="X\sim N_n(\mu,\sum),\sum>0,A^{'}=A,B^{'}=B,A\sum B=0,则X^{'}BX\bot X^{'}AX" contenteditable="false"><span></span><span></span></span>(<b>协方差阵一般化了</b>)
<font color="#ff0000"><b>辨析正交 独立不相关</b></font>
正态情形下<span class="equation-text" contenteditable="false" data-index="0" data-equation="\Longrightarrow"><span></span><span></span></span>独立
X、Y有一期望为0<span class="equation-text" contenteditable="false" data-index="0" data-equation="\Longrightarrow"><span></span><span></span></span>正交
独立<span class="equation-text" data-index="0" data-equation="\Longrightarrow" contenteditable="false"><span></span><span></span></span>不相关,正交<span class="equation-text" data-index="1" data-equation="\Longrightarrow" contenteditable="false"><span></span><span></span></span>不相关
三、回归参数的估计
3.1最小二乘估计
<br><span class="equation-text" contenteditable="false" data-index="0" data-equation="y=X\beta+e"><span></span><span></span></span><br><span class="equation-text" contenteditable="false" data-index="1" data-equation="\begin{cases}E(e_i)=0 \\Var(e_i)=\sigma^2 \\Cov(e_i,e_j)=0\end{cases}"><span></span><span></span></span>
2.<span class="equation-text" contenteditable="false" data-index="0" data-equation="\hat{\beta}=(X^{'}X)^{-1}X^{'}Y"><span></span><span></span></span>
证明:<span class="equation-text" contenteditable="false" data-index="0" data-equation="Q(\beta)=||y-X\beta||^2=(y-X\beta)^{'}(y-X\beta)=y^{'}y-2y^{'}X\beta+\beta^{'}X^{'}X\beta"><span></span><span></span></span><br><span class="equation-text" contenteditable="false" data-index="1" data-equation="\hat{\beta}=arcminQ(\beta)"><span></span><span></span></span>,<span class="equation-text" contenteditable="false" data-index="2" data-equation="Q(\beta)取最小时,\hat{\beta}=\beta=(X^{'}X)^{-1}X^{'}Y"><span></span><span></span></span>
3.中心化和标椎化
3.2 最小二乘估计性质
1.<span class="equation-text" contenteditable="false" data-index="0" data-equation="\hat{\beta}=(X^{'}X)^{-1}X^{'}Y"><span></span><span></span></span>的性质
定理3.2.1:<span class="equation-text" contenteditable="false" data-index="0" data-equation="\begin{cases}E (\hat{\beta})=\beta\\Cov(\hat{\beta})=\sigma^2(X^{'}X)^{-1} \end{cases}"><span></span><span></span></span>
求<span class="equation-text" contenteditable="false" data-index="0" data-equation="Cov(\beta)时用到了定理2.1.3"><span></span><span></span></span>+<span class="equation-text" contenteditable="false" data-index="1" data-equation="(X^{'}X)^{-1}对称"><span></span><span></span></span>
推论3.2.1:<span class="equation-text" contenteditable="false" data-index="0" data-equation="\begin{cases}E (c^{'}\hat{\beta})=c^{'}\beta\\Cov(c^{'}\hat{\beta})=\sigma^2c^{'}(X^{'}X)^{-1}c \end{cases}"><span></span><span></span></span>
定理3.2.2(Gauss-Markov):在所有线性、无偏估计中,最小二乘最优(方差最小)
证明思路:①假设<span class="equation-text" data-index="0" data-equation="a^{'}y是c^{'}\beta" contenteditable="false"><span></span><span></span></span>的任一线性无偏估计<br>②利用无偏(<span class="equation-text" contenteditable="false" data-index="1" data-equation="E(a^{'}y)=c^{'}\beta"><span></span><span></span></span>)得到c和a的关系<br>③求<span class="equation-text" contenteditable="false" data-index="2" data-equation="Cov(a^{'}y)-Cov(c^{'}\beta)"><span></span><span></span></span>
2.<span class="equation-text" contenteditable="false" data-index="0" data-equation="\hat{e}=y-X\hat{\beta}"><span></span><span></span></span>的性质
<span class="equation-text" contenteditable="false" data-index="0" data-equation="RSS=\hat{e^{'}}\hat{e}(=\sum_{i=1}^n\hat{e_i^2})"><span></span><span></span></span>
注:RSS和SSE是一样的意思
定理3.2.3:<span class="equation-text" contenteditable="false" data-index="0" data-equation="\begin{cases}RSS=y{'}(I-X(X{'}X)^{-1}X{'})y\\\hat{\sigma^2}={RSS\over{n-p} }是\sigma^2的无偏估计\end{cases}"><span></span><span></span></span>
证RSS时用到了<font color="#ff0000"><span class="equation-text" data-index="0" data-equation="I-P_x=(I-X(X{'}X)^{-1}X{'})" contenteditable="false"><span></span><span></span></span>是对称幂等</font>的性质
证<span class="equation-text" data-index="0" data-equation="\sigma^2" contenteditable="false"><span></span><span></span></span>时用到了"<b>定理2.2.1</b>"+"trBA=trAB"
3.<span class="equation-text" contenteditable="false" data-index="0" data-equation="y=X\beta+e的性质"><span></span><span></span></span>
定理3.2.4:<span class="equation-text" contenteditable="false" data-index="0" data-equation="\begin{cases}1.\hat{\beta}_{ls}=\hat{\beta}_{MLE}=(X{'}X)^{-1}X{'}Y\sim N(\beta,\sigma^2(X{'}X)^{-1})\\\hat{\sigma^2}_{MLE}={RSS\over n}={(n-p)\over n }\hat{\sigma^2}_{ls}\\2.{RSS\over \sigma^2}\sim \chi_{n-p}^2\\3.\hat{\beta}与RSS相互独立\end{cases}"><span></span><span></span></span>
证明<span class="equation-text" data-index="0" data-equation="\hat{\beta}_{ls}=\hat{\beta}_{MLE}" contenteditable="false"><span></span><span></span></span><b>思路</b>:①先令<span class="equation-text" data-index="1" data-equation="\mu=Ey=X\beta,e\sim N(0,\sigma^2In)" contenteditable="false"><span></span><span></span></span><br>②写出对数似然函数<span class="equation-text" data-index="2" data-equation="l(\mu,\sigma^2)" contenteditable="false"><span></span><span></span></span>(y还是服从正态分布的)<br>③发现<span class="equation-text" data-index="3" data-equation="l(\mu,\sigma^2)" contenteditable="false"><span></span><span></span></span>取最大时,<span class="equation-text" data-index="4" data-equation="||y-\mu||_2^2" contenteditable="false"><span></span><span></span></span>取最小,即<span class="equation-text" data-index="5" data-equation="\mu_{ls}=\mu_{MLE}" contenteditable="false"><span></span><span></span></span>,所以<span class="equation-text" data-index="6" data-equation="\hat{\beta}_{ls}=\hat{\beta}_{MLE}" contenteditable="false"><span></span><span></span></span>
证明<span class="equation-text" data-index="0" data-equation="\hat{\sigma^2}_{MLE}={RSS\over n}" contenteditable="false"><span></span><span></span></span><b>思路</b>:①写出对数似然函数<span class="equation-text" data-index="1" data-equation="l(\mu,\sigma^2)" contenteditable="false"><span></span><span></span></span><br>②把<span class="equation-text" data-index="2" data-equation="\sigma^2" contenteditable="false"><span></span><span></span></span>当做整体,求偏导,令其为零,即可得MLE。<br><span class="equation-text" data-index="3" data-equation="\hat{\sigma^2}_{MLE}={||y-\hat{\mu}||_2^2\over n}={||y-X\hat{\beta}||_2^2\over n}={||\hat{e}||_2^2\over n}={RSS\over n}" contenteditable="false"><span></span><span></span></span>
证明<span class="equation-text" data-index="0" data-equation="{RSS\over \sigma^2}\sim \chi_{n-p}^2" contenteditable="false"><span></span><span></span></span><b>思路</b>:①由<b>定理3.2.3</b>得,<span class="equation-text" data-index="1" data-equation="RSS=y{'}(I-P_x)y=(X\beta+e){'}(I-P_x)(X\beta+e)=e{'}(I-P_x)e" contenteditable="false"><span></span><span></span></span>(因为<span class="equation-text" data-index="2" data-equation="I-P_x与X正交,乘积为零" contenteditable="false"><span></span><span></span></span>)<br>②<span class="equation-text" data-index="3" data-equation="I-P_x" contenteditable="false"><span></span><span></span></span>是幂等阵,<span class="equation-text" data-index="4" data-equation="rank(I-P_x)=n-p," contenteditable="false"><span></span><span></span></span>由<b>定理2.4.4</b>,得<span class="equation-text" data-index="5" data-equation="{RSS\over \sigma^2}\sim \chi_{n-p}^2" contenteditable="false"><span></span><span></span></span>
证明<span class="equation-text" data-index="0" data-equation="\hat{\beta}与RSS相互独立" contenteditable="false"><span></span><span></span></span><b>思路</b>:①<span class="equation-text" data-index="1" data-equation="\hat{\beta}=(X^{'}X)^{-1}X^{'}(X\beta+e)=\beta+(X^{'}X)^{-1}X^{'}e,RSS=e^{'}(I-P_x)e" contenteditable="false"><span></span><span></span></span><br>②因为<span class="equation-text" data-index="2" data-equation="(X^{'}X)^{-1}X^{'}(I-P_x)=0" contenteditable="false"><span></span><span></span></span>,由<b>定理2.4.5</b>得,<span class="equation-text" data-index="3" data-equation="(X^{'}X)^{-1}X^{'}e与RSS相互独立" contenteditable="false"><span></span><span></span></span><br>③因此<span class="equation-text" data-index="4" data-equation="\hat{\beta}与RSS" contenteditable="false"><span></span><span></span></span>相互独立
3.3约束最小二乘估计
<br><span class="equation-text" data-index="0" data-equation="y=X\beta+e" contenteditable="false"><span></span><span></span></span><br><span class="equation-text" data-index="1" data-equation="\begin{cases}E(e_i)=0 \\Var(e_i)=\sigma^2 \\Cov(e_i,e_j)=0\end{cases}+确定的约束H\beta=d(其中,r(H)=r(H\ d)\Rightarrow解存在)" contenteditable="false"><span></span><span></span></span>
2.<span class="equation-text" data-index="0" data-equation="\hat{\beta}_H=\hat{\beta}-(X^{'}X)^{-1}H^{'}[H(X^{'}X)^{-1}H^{'}]^{-1}(H\hat{\beta}-d)" contenteditable="false"><span></span><span></span></span>
证明:(利用Lagurange乘子法)<br>①令<span class="equation-text" contenteditable="false" data-index="0" data-equation="F(\beta,\lambda)\ =Q(\beta)+2\lambda(H^{'}\beta-d)=Q(\beta)+2\begin{matrix} \sum_{i=1}^k \lambda_i(h_i^{'}\beta-d_i) \end{matrix}"><span></span><span></span></span><br>②对<span class="equation-text" contenteditable="false" data-index="1" data-equation="\beta"><span></span><span></span></span>求偏导,得<span class="equation-text" contenteditable="false" data-index="2" data-equation="\hat{\beta}_H=\hat{\beta}-(X^{'}X)^{-1}H^{'}\hat{\lambda}_H"><span></span><span></span></span><br>③对<span class="equation-text" contenteditable="false" data-index="3" data-equation="\lambda"><span></span><span></span></span>求偏导,得<span class="equation-text" contenteditable="false" data-index="4" data-equation="H\beta=d,带入\hat{\beta}_H,得\hat{\lambda}_H"><span></span><span></span></span><br>④将<span class="equation-text" data-index="5" data-equation="\hat{\lambda}_H" contenteditable="false"><span></span><span></span></span> 带入②,得<span class="equation-text" data-index="6" data-equation="\hat{\beta}_H=\hat{\beta}-(X^{'}X)^{-1}H^{'}[H(X^{'}X)^{-1}H^{'}]^{-1}(H\hat{\beta}-d)" contenteditable="false"><span></span><span></span></span>
3.验证<span class="equation-text" contenteditable="false" data-index="0" data-equation="\hat{\beta}_H"><span></span><span></span></span>确实是该约束条件下的最小二乘估计
(a)证明<span class="equation-text" contenteditable="false" data-index="0" data-equation="H\hat{\beta}_H=d"><span></span><span></span></span>
(b)证明<span class="equation-text" data-index="0" data-equation="||y-X\beta||_{H\beta=d}^2\geq||y-X\hat{\beta}_H||^2" contenteditable="false"><span></span><span></span></span><br><br><span class="equation-text" data-index="1" data-equation="||y-X\beta||^2=||y-X\beta+X\beta-X\hat{\beta}_H+X\hat{\beta}_H-X\beta||^2" contenteditable="false"><span></span><span></span></span><span class="equation-text" data-index="2" data-equation="=∣∣y−X\hat{β}∣∣^2+∣∣X(\hat{β}−\hat{β}_H)∣∣^2+∣∣X(\hat{β}_H−β)∣∣^2" contenteditable="false"><span></span><span></span></span><br>所以<span class="equation-text" data-index="3" data-equation="∣∣y−Xβ∣∣^2\geq∣∣y−X\hat{β}∣∣^2+∣∣X(\hat{β}−\hat{β}_H)∣∣^2" contenteditable="false"><span></span><span></span></span><b><font color="#ff0000">①</font></b>当且仅当<span class="equation-text" data-index="4" data-equation="\hat{\beta}_H=\beta" contenteditable="false"><span></span><span></span></span>时等号成立<br>当<span class="equation-text" data-index="5" data-equation="\hat{\beta}_H=\beta" contenteditable="false"><span></span><span></span></span>时,代入①式,得<span class="equation-text" data-index="6" data-equation="∣∣y−X\hat{β}_H∣∣^2=∣∣y−X\hat{β}∣∣^2+∣∣X(\hat{β}−\hat{β}_H)∣∣^2" contenteditable="false"><span></span><span></span></span>(得到<span class="equation-text" data-index="7" data-equation="∣∣X(\hat{β}−\hat{β}_H)∣∣^2" contenteditable="false"><span></span><span></span></span>表达式)<br>结合①式,得<span class="equation-text" contenteditable="false" data-index="8" data-equation="∣∣y−Xβ∣∣^2\geq∣∣y−X\hat{β}∣∣^2+∣∣y-X\hat{β}_H∣∣^2-∣∣y−X\hat{β}∣∣^2=∣∣y-X\hat{β}_H∣∣^2"><span></span><span></span></span>
3.4 回归诊断<span class="equation-text" contenteditable="false" data-index="0" data-equation="(\hat{e}=y-X\hat{\beta})"><span></span><span></span></span>
1.残差诊断<span class="equation-text" data-index="0" data-equation="\hat{e}=y-X\hat{\beta}=(I-P_x)y" contenteditable="false"><span></span><span></span></span>
定理3.4.1 ①<span class="equation-text" contenteditable="false" data-index="0" data-equation="E(\hat{e})=0,cov(\hat{e})=\sigma^2(I-P_x)"><span></span><span></span></span><br> ②<span class="equation-text" contenteditable="false" data-index="1" data-equation="若e\sim N(0,\sigma^2I_n),则\hat{e}\sim N(0,\sigma^2(I_n-P_x)"><span></span><span></span></span><br> ③<span class="equation-text" contenteditable="false" data-index="2" data-equation="cov(\hat{y},\hat{e})=0"><span></span><span></span></span><br> ④<span class="equation-text" contenteditable="false" data-index="3" data-equation="1^{'}\hat{e}=0"><span></span><span></span></span>(说明是中心化的向量)
2.残差图
(1)以拟合值<span class="equation-text" contenteditable="false" data-index="0" data-equation="\hat{y}_i"><span></span><span></span></span>为横轴,以<span class="equation-text" contenteditable="false" data-index="1" data-equation="r_i"><span></span><span></span></span>为纵轴,其中<span class="equation-text" data-index="2" data-equation="r_i=\frac{\hat{e}_i}{\hat{\sigma}\sqrt{1-P_{ii}}}\sim N(0,\sigma^2I)" contenteditable="false"><span></span><span></span></span>,这里<span class="equation-text" contenteditable="false" data-index="3" data-equation="\hat{\sigma}=\frac{SSE}{n-p}"><span></span><span></span></span>
(2)合理情形:点在<span class="equation-text" data-index="0" data-equation="(\mu-2\sigma,u+2\sigma)" contenteditable="false"><span></span><span></span></span> 范围大致分布均匀,不呈现任何趋势
(3)不合理的两种情况
(i).残差图呈现对勾型或抛物线型,说明因变量Y对自变量 X的依赖不仅仅是线性关系,引入二次项<span class="equation-text" contenteditable="false" data-index="0" data-equation="Z_1=X_1^2,Z_2=X_2^2,Z_3=X_1X_2"><span></span><span></span></span>转化为线性
(ii).只有<span class="equation-text" data-index="0" data-equation="\hat{y}_i" contenteditable="false"><span></span><span></span></span>偏大或偏小或中等时<span class="equation-text" data-index="1" data-equation="r_i" contenteditable="false"><span></span><span></span></span>偏大,或呈线性趋势,说明误差方差不相等,则<br>①对因变量做变换,使变换过后具有近似相等的方差②应用加权最小二乘估计③对因变量做Box-Cox变换<br>
3.强影响点<br>(<span class="equation-text" data-index="0" data-equation="\hat{\beta}-\hat{\beta}_{(i)}" contenteditable="false"><span></span><span></span></span>可以反映影响力大小)<br>
<span class="equation-text" data-index="0" data-equation="y_{(i)}、X_{(i)}、e_{(i)}" contenteditable="false"><span></span><span></span></span>分别表示从Y、X、e中剔除第i行所得到的向量或矩阵,<span class="equation-text" contenteditable="false" data-index="1" data-equation="\hat{\beta}_{(i)}=(X_{(i)}^{'}X_{(i)})^{-1}X_{(i)}^{'}y_{(i)}"><span></span><span></span></span>
影响力大小判断
量化<span class="equation-text" data-index="0" data-equation="\hat{\beta}-\hat{\beta}_{(i)}" contenteditable="false"><span></span><span></span></span>的Cook统计量(原始公式):<span class="equation-text" data-index="1" data-equation="D_i=\frac{(\hat{\beta}-\hat{\beta}_{(i)})^{'}X^{'}X(\hat{\beta}-\hat{\beta}_{(i)})}{p\hat{\sigma}^2}" contenteditable="false"><span></span><span></span></span>,其中<span class="equation-text" data-index="2" data-equation="\hat{\sigma}=\frac{SSE}{n-p}" contenteditable="false"><span></span><span></span></span>,但要计算n+1个回归<font color="#ff0000">计算量太大</font>
<b>定理3.4.2(简便公式)</b> <span class="equation-text" data-index="0" data-equation="D_i=\frac{1}{p}(\frac{P_{ii}}{1-P_{ii}})r_i^2,1\leq i\leq n,其中P_{ii}=X_{i}^{'}(X_{i}^{'}X_{i})^{-1}X_i=e_i^{'}(X(X^{'}X)^{-1}X^{'})e_i=e_i^{'}P_xe_i" contenteditable="false"><span></span><span></span></span>
证明思路:用到了<br><span class="equation-text" contenteditable="false" data-index="0" data-equation="(X_{(i)}^{'}X_{(i)})^{-1}=(X^{'}X-X_iX_i^{'})^{-1}"><span></span><span></span></span>和<br><span class="equation-text" contenteditable="false" data-index="1" data-equation="(A-uv)^{-1}=A^{-1}+\frac{A^{-1}u\ v^{'}A^{-1}}{1-u^{'}A^{-1}v}"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="D_i \propto \frac{P_{ii}}{1-P_{ii}}"><span></span><span></span></span>,只需计算<span class="equation-text" contenteditable="false" data-index="1" data-equation="r_i和P_x的对角元P_{ii}"><span></span><span></span></span>即可
难以判断强影响数据的<font color="#ff0000">临界值</font>,需引入F统计量
<span class="equation-text" data-index="0" data-equation="随机事件\frac{(\hat{\beta}-\beta)^{'}X^{'}X(\hat{\beta}-\beta)}{p\hat{\sigma}^2}\leq F_{p,n-p}(\alpha)" contenteditable="false"><span></span><span></span></span>(<span class="equation-text" data-index="1" data-equation="\hat{\beta}_{(i)}" contenteditable="false"><span></span><span></span></span>落在以 <span class="equation-text" contenteditable="false" data-index="2" data-equation="\hat{\beta}"><span></span><span></span></span>为中心的置信椭球里)发生的概率为<span class="equation-text" data-index="3" data-equation="1-\alpha" contenteditable="false"><span></span><span></span></span>,<span class="equation-text" contenteditable="false" data-index="4" data-equation="\alpha"><span></span><span></span></span>越大,越不容易落在椭球里,影响力越大
分子:<span class="equation-text" contenteditable="false" data-index="0" data-equation="因为\hat{\beta}\sim N(\beta,\sigma^2(X^{'}X)^{-1}),由定理2.4.1,得\frac{(\hat{\beta}-\beta)^{'}X^{'}X(\hat{\beta}-\beta)}{\sigma^2}\sim \chi^2_p"><span></span><span></span></span>
分母:<span class="equation-text" contenteditable="false" data-index="0" data-equation="\frac{SSE}{\sigma^2}=\frac{(n-p)\hat{\sigma}^2}{\sigma^2}\sim\chi^2_{n-p}"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="\frac{(\hat{\beta}-\beta)^{'}X^{'}X(\hat{\beta}-\beta)}{p\hat{\sigma}^2}\sim F_{p,n-p}"><span></span><span></span></span>
3.5 Box-Cox变换<br>(对因变量y做变换,使满足线性+误差的G-M假设)<br>
Step1.给定 <span class="equation-text" contenteditable="false" data-index="0" data-equation="\lambda"><span></span><span></span></span>范围,<span class="equation-text" contenteditable="false" data-index="1" data-equation="\lambda \in \left\{\lambda_1,\lambda_2.......\lambda_n\right\}"><span></span><span></span></span>
Step2.计算<span class="equation-text" contenteditable="false" data-index="0" data-equation="z_i^{(\lambda)}=\begin{cases}\frac{y_i^{\lambda}}{J^{\frac{1}{n}}}=\frac{y_i^{\lambda}}{(\prod_{i=1}^n y_i)^{\frac{\lambda-1}{n}}} \ \ \ \ \ \ \ \ \ \lambda\neq 0 \\(ln y_i)((\prod_{i=1}^n y_i)^{\frac{1}{n}} \ \ \ \ \ \ \ \ \ \ \ \lambda=0 \\\end{cases}"><span></span><span></span></span>
Step3.计算<span class="equation-text" data-index="0" data-equation="SSE(\lambda,z^{(\lambda)})=z^{(\lambda){'}}(I-P_x)z^{(\lambda)}" contenteditable="false"><span></span><span></span></span>,其中<span class="equation-text" data-index="1" data-equation="z^{(\lambda)}=(z_1^{(\lambda)},z_2^{(\lambda)}...z_n^{(\lambda)})" contenteditable="false"><span></span><span></span></span>,取<span class="equation-text" data-index="2" data-equation="\hat{\lambda}" contenteditable="false"><span></span><span></span></span>为SSE最小值的<span class="equation-text" contenteditable="false" data-index="3" data-equation="\lambda"><span></span><span></span></span>值
Step4.计算<span class="equation-text" contenteditable="false" data-index="0" data-equation="\hat{\beta}(\lambda)=(X^{'}X)^{-1}X^{'}y^{(\hat{\lambda})}"><span></span><span></span></span>,<span class="equation-text" contenteditable="false" data-index="1" data-equation="\hat{\sigma}(\lambda)=\frac{SSE(\hat{\lambda)}}{n}"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="Y^{(\lambda)}=\begin{cases}\frac{Y^{\lambda}-1}{\lambda} ,\lambda\neq0\\lnY ,\lambda=0\\\end{cases}"><span></span><span></span></span>
3.6广义最小二乘估计<br>(<span class="equation-text" contenteditable="false" data-index="0" data-equation="cov(e)\neq\sigma^2I=\sigma^2\sum,其中\sum正定)"><span></span><span></span></span>
令<span class="equation-text" contenteditable="false" data-index="0" data-equation="\tilde{e}=\sum^{-\frac{1}{2}}e,\tilde{y}=\sum^{-\frac{1}{2}}y,\tilde{X}=\sum^{-\frac{1}{2}}X,则\tilde{y}=\tilde{X}\beta+\tilde{e}, E(\tilde{e})=0,cov(\tilde{e})=\sigma^2I_n"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="\tilde{\beta}(\beta^*)=(\tilde{X}^{'}\tilde{X})^{-1}\tilde{X}^{'}\tilde{y}=(X^{'}\sum^{-1}X)^{-1}X^{'}\sum^{-1}y"><span></span><span></span></span>
定理3.6.1 <br>(a)<span class="equation-text" data-index="0" data-equation="E(\beta^*)=\beta" contenteditable="false"><span></span><span></span></span><br>(b)<span class="equation-text" data-index="1" data-equation="cov(\beta^*)=\sigma^2(X^{'}\sum^{-1}X)^{-1}" contenteditable="false"><span></span><span></span></span><br>(c)<span class="equation-text" contenteditable="false" data-index="2" data-equation="c^{'}\beta^*是c^{'}\beta"><span></span><span></span></span>唯一的最小方差无偏估计
<b>注意:</b>①证明定理时注意,只有y是随机的!<br> ②<span class="equation-text" data-index="0" data-equation="\beta^*" contenteditable="false"><span></span><span></span></span> 和 <span class="equation-text" data-index="1" data-equation="\hat{\beta}" contenteditable="false"><span></span><span></span></span>都是<span class="equation-text" contenteditable="false" data-index="2" data-equation="\beta"><span></span><span></span></span> 的无偏估计,对于一般的线性回归模型,广义最小二乘优于简单最小二乘
<span class="equation-text" contenteditable="false" data-index="0" data-equation="\sigma^{2*}=\frac{(y-X\beta^*)^{'}\sum^{-1}(y-X\beta^*)}{n-p}=\frac{e^{*'}\sum^{-1}e^*}{n-p}"><span></span><span></span></span>
证明同<span class="equation-text" data-index="0" data-equation="\hat{\sigma}^2" contenteditable="false"><span></span><span></span></span>,多了个<span class="equation-text" contenteditable="false" data-index="1" data-equation="\sum^{-1}"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="e^*=y-X\beta^*=\sum^{\frac{1}{2}}(I-P_{\sum^{-\frac{1}{2}}X})\sum^{-\frac{1}{2}}y"><span></span><span></span></span>
<span class="equation-text" data-index="0" data-equation="E(e^*)=0,cov(e^*)=\sigma^2" contenteditable="false"><span></span><span></span></span><span class="equation-text" contenteditable="false" data-index="1" data-equation="\sum^{\frac{1}{2}}(I-P_{\sum^{-\frac{1}{2}}X})\sum^{\frac{1}{2}}"><span></span><span></span></span>
注意:①<span class="equation-text" data-index="0" data-equation="\sum" contenteditable="false"><span></span><span></span></span>的次数什么时候为正什么时候为负<br> <font color="#ff0000">②<span class="equation-text" data-index="1" data-equation="e^*和\tilde{e}" contenteditable="false"><span></span><span></span></span>不一样!只有<span class="equation-text" data-index="2" data-equation="\beta^*和\tilde{\beta}" contenteditable="false"><span></span><span></span></span>可当做一样的,都是估计出来的</font>
事实上,<span class="equation-text" data-index="0" data-equation="\sum" contenteditable="false"><span></span><span></span></span>通常无法确定,可采用<b>两步估计法</b>:先假设<span class="equation-text" data-index="1" data-equation="\sum=I" contenteditable="false"><span></span><span></span></span>,求出简单最小二乘估计,然后通过残差分析,可以估计出<span class="equation-text" data-index="2" data-equation="\sum" contenteditable="false"><span></span><span></span></span>
3.7 复共线性<br>(自变量之间存在近似线性关系)
定理3.7.1 <span class="equation-text" contenteditable="false" data-index="0" data-equation="MSE(\hat{\theta})=trCov(\hat{\theta})+||E\hat{\theta}-\theta||^2"><span></span><span></span></span>
证明思路:①<span class="equation-text" data-index="0" data-equation="MSE(\hat{\theta})=E(\hat{\theta}-\theta)^{'}(\hat{\theta}-\theta)=E[(\hat{\theta}-E\hat{\theta}+E\hat{\theta}-\theta)^{'}(\hat{\theta}-E\hat{\theta}+E\hat{\theta}-\theta)" contenteditable="false"><span></span><span></span></span><br> (拆开后)<span class="equation-text" data-index="1" data-equation="=E[(\hat{\theta}-E\hat{\theta})^{'}(\hat{\theta}-E\hat{\theta})]+E[( E\hat{\theta}-\theta)^{'}( E\hat{\theta}-\theta)]+2E[(\hat{\theta}-E\hat{\theta})^{'}( E\hat{\theta}-\theta)]" contenteditable="false"><span></span><span></span></span><br> ②第一项(<span class="equation-text" data-index="2" data-equation="\triangle_1" contenteditable="false"><span></span><span></span></span>)=<span class="equation-text" data-index="3" data-equation="trCov(\hat{\theta})" contenteditable="false"><span></span><span></span></span>,第二项(<span class="equation-text" contenteditable="false" data-index="4" data-equation="\triangle_2"><span></span><span></span></span>)=<span class="equation-text" data-index="5" data-equation="(E\hat{\theta}-\theta)^{'}(E\hat{\theta}-\theta)=||E\hat{\theta}-\theta||^2" contenteditable="false"><span></span><span></span></span>,第三项(交叉项)为0
考虑线性回归模型:<span class="equation-text" data-index="0" data-equation="y=\alpha_01+X\beta+e,Ee=0,Cov(e)=\sigma^2I" contenteditable="false"><span></span><span></span></span><br>(X已中心化,<span class="equation-text" contenteditable="false" data-index="1" data-equation="r(X_{n\times p})=p,"><span></span><span></span></span>满秩)<br>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="MSE(\hat{\beta})=\sigma^2\sum_{i=1}^{p}\frac{1}{\lambda_i}"><span></span><span></span></span>
<span class="equation-text" data-index="0" data-equation="\triangle_1=\sigma^2tr(X^{'}X)^{-1},X^{'}X正定,对称,对其作正交谱分解,(X^{'}X)^{-1}=\phi\begin{bmatrix}\frac{1}{\lambda_i}& \cdots & 0\\\vdots & \ddots & \vdots \\0& \cdots & \frac{1}{\lambda_p}\end{bmatrix}\phi^{'}" contenteditable="false"><span></span><span></span></span><br>所以<span class="equation-text" contenteditable="false" data-index="1" data-equation="\triangle_1=\sigma^2\sum_{i=1}^{p}\frac{1}{\lambda_i}"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="\triangle_2=0(\hat{\beta}无偏)"><span></span><span></span></span>
若<span class="equation-text" contenteditable="false" data-index="0" data-equation="X^{'}X"><span></span><span></span></span>至少有一个特征根非常小,MSE就会很大!
①与Gauss-Markov定理并不矛盾,最小二乘在线性无偏中方差最小,而这个最小的方差值本身却很大
②可以从MSE(<span class="equation-text" contenteditable="false" data-index="0" data-equation="\hat{\beta}"><span></span><span></span></span>)表达式直观地看出
③另一解释:<span class="equation-text" data-index="0" data-equation="MSE(\hat{\beta})=E(\hat{\beta}-\beta)^{'}(\hat{\beta}-\beta)=E||\hat{\beta}||^2-\beta^{'}\beta=E||\hat{\beta}||^2-||\beta||^2" contenteditable="false"><span></span><span></span></span><br>当<span class="equation-text" data-index="1" data-equation="X^{'}X" contenteditable="false"><span></span><span></span></span>至少有一个特征根非常小,最小二乘估计<span class="equation-text" contenteditable="false" data-index="2" data-equation="\hat{\beta}"><span></span><span></span></span>的长度要比真正长度长的多
<font color="#ff0000">④意味着什么</font>:设<span class="equation-text" data-index="0" data-equation="X^{'}X" contenteditable="false"><span></span><span></span></span>的非常小的特征值为<span class="equation-text" data-index="1" data-equation="\lambda\approx0" contenteditable="false"><span></span><span></span></span>,对应特征向量为<span class="equation-text" data-index="2" data-equation="\phi" contenteditable="false"><span></span><span></span></span>,其中<span class="equation-text" data-index="3" data-equation="\phi^{'}\phi=1" contenteditable="false"><span></span><span></span></span><br>则<span class="equation-text" data-index="4" data-equation="X^{'}X\phi=\lambda\phi\approx0" contenteditable="false"><span></span><span></span></span><br>用<span class="equation-text" data-index="5" data-equation="\phi^{'}" contenteditable="false"><span></span><span></span></span>左乘,<span class="equation-text" data-index="6" data-equation="\phi^{'}X^{'}X\phi=||X\phi||^2=\lambda\approx0" contenteditable="false"><span></span><span></span></span><br>所以<span class="equation-text" data-index="7" data-equation="X\phi\approx0" contenteditable="false"><span></span><span></span></span>,说明X的列向量之间具有线性关系。
<span class="equation-text" contenteditable="false" data-index="0" data-equation="k=\frac{\lambda_1}{\lambda_{p-1}}"><span></span><span></span></span>(<span class="equation-text" data-index="1" data-equation="\lambda_i是X^{'}X" contenteditable="false"><span></span><span></span></span>的特征根,已排序)
<span class="equation-text" contenteditable="false" data-index="0" data-equation="k\approx1时,"><span></span><span></span></span>表明各特征值差别不大,<span class="equation-text" data-index="1" data-equation="X^{'}X" contenteditable="false"><span></span><span></span></span>随机性不强,复共线性不强
<span class="equation-text" contenteditable="false" data-index="0" data-equation="\begin{cases}k<100 \ \ \ \ \ \ \ \ 弱共线性 \\100\leq k <1000 \ \ \ \ \ \ \ \ \ 中共线性\\ k \geq1000 \ \ \ \ \ \ \ \ \ 强共线性\end{cases}"><span></span><span></span></span>
3.8岭估计<br>(有偏估计)
<span class="equation-text" contenteditable="false" data-index="0" data-equation="\hat{\beta}(k)=(X^{'}X+kI)^{-1}X^{'}y"><span></span><span></span></span>
<b>推导</b>:由于复共线性的存在,最小二乘偏差较大,①用拉格朗日乘法,引入k,令<span class="equation-text" data-index="0" data-equation="Q_k(\beta)=||y-X\beta||^2+k||\beta||^2" contenteditable="false"><span></span><span></span></span><br>②求<span class="equation-text" data-index="1" data-equation="\beta,使Q_k(\beta)最小" contenteditable="false"><span></span><span></span></span>,先展开<span class="equation-text" data-index="2" data-equation="Q_k(\beta)" contenteditable="false"><span></span><span></span></span>,对<span class="equation-text" data-index="3" data-equation="\beta" contenteditable="false"><span></span><span></span></span>求偏导,得<span class="equation-text" data-index="4" data-equation="\hat{\beta}(k)=(X^{'}X+kI)^{-1}X^{'}y" contenteditable="false"><span></span><span></span></span><br>(下面任务就是确定k,然后就能确定<span class="equation-text" data-index="5" data-equation="\hat{\beta}(k)" contenteditable="false"><span></span><span></span></span>)
<b>注意</b>:①<span class="equation-text" data-index="0" data-equation="X^{'}X+k I" contenteditable="false"><span></span><span></span></span>仍对称,正定,特征值为<span class="equation-text" data-index="1" data-equation="\lambda_i+k" contenteditable="false"><span></span><span></span></span><br>②k=0时,<span class="equation-text" data-index="2" data-equation="\hat{\beta}(0)=\hat{\beta}_{ls}" contenteditable="false"><span></span><span></span></span><br>③<span class="equation-text" data-index="3" data-equation="E(\hat{\beta}_k)=(X^{'}X+kI)^{-1}X^{'}X\beta\neq\beta" contenteditable="false"><span></span><span></span></span>(有偏估计)
典则模型
典则<b>模型</b>:<span class="equation-text" data-index="0" data-equation="y=\alpha_01+Z\alpha+e,Ee=0,Cov(e)=\sigma^2I_n" contenteditable="false"><span></span><span></span></span><br>(其中<span class="equation-text" data-index="1" data-equation="\phi为正交阵,X^{'}X=\phi \lambda \phi^{'},Z=X\phi,\alpha=\phi^{'}\beta" contenteditable="false"><span></span><span></span></span>)
<span class="equation-text" data-index="0" data-equation="\begin{cases}\hat{\alpha}_0 =\bar{y}\\\hat{\alpha}=(Z^{'}Z)^{-1}Z^{'}Y=\lambda^{-1}Z^{'}y \ \ \end{cases}" contenteditable="false"><span></span><span></span></span><b>估计</b>
<span class="equation-text" data-index="0" data-equation="E(\hat{\alpha})=" contenteditable="false"><span></span><span></span></span><span class="equation-text" contenteditable="false" data-index="1" data-equation="\phi^{'}E(\hat{\beta})=\phi^{'}\beta=\alpha"><span></span><span></span></span><br><span class="equation-text" contenteditable="false" data-index="2" data-equation="cov(\hat{\alpha})=\sigma^2(Z^{'}Z)^{-1}=\sigma^2\lambda^{-1}"><span></span><span></span></span>
<span class="equation-text" data-index="0" data-equation="\hat{\alpha}与\hat{\beta}" contenteditable="false"><span></span><span></span></span>的<b>关系</b>
<span class="equation-text" data-index="0" data-equation="带入Z=X\phi ,\hat{\alpha}=\phi^{'}\hat{\beta}\Longrightarrow MSE(\hat{\alpha})= MSE(\hat{\beta})" contenteditable="false"><span></span><span></span></span><br><span class="equation-text" data-index="1" data-equation="\hat{\alpha}(k)=(Z^{'}Z+kI)^{-1}Z^{'}y=\phi^{'}\hat{\beta}(k)\Longrightarrow MSE(\hat{\alpha}(k))= MSE(\hat{\beta}(k))" contenteditable="false"><span></span><span></span></span>
性质
定理3.8.1 <span class="equation-text" data-index="0" data-equation="\exist k>0,s.t.MSE(\hat{\beta}(k))<MSE(\hat{\beta})" contenteditable="false"><span></span><span></span></span>
证明思路:只需证明<span class="equation-text" data-index="0" data-equation="MSE(\hat{\alpha}(k))<MSE(\hat{\alpha})" contenteditable="false"><span></span><span></span></span><br><span class="equation-text" data-index="1" data-equation="利用Z^{'}Z=\lambda和Z^{'}1=0" contenteditable="false"><span></span><span></span></span>(已中心化),求出<span class="equation-text" data-index="2" data-equation="MSE(\hat{\alpha}(k))" contenteditable="false"><span></span><span></span></span>表达式,对k求偏导,发现在0附近单调递减,<br>所以<span class="equation-text" contenteditable="false" data-index="3" data-equation="MSE(\hat{\alpha}(k))<MSE(\hat{\alpha}(0))=MSE(\hat{\alpha})"><span></span><span></span></span>
找k,让<span class="equation-text" contenteditable="false" data-index="0" data-equation="MSE(\hat{\beta}(k))"><span></span><span></span></span>最小
①相当于<span class="equation-text" contenteditable="false" data-index="0" data-equation="f(k)=MSE(\hat{\alpha}(k))"><span></span><span></span></span>最小,对其求偏导,直接算解不出来
②H-K统计量:<span class="equation-text" contenteditable="false" data-index="0" data-equation="\hat{k}=\frac{\hat{\sigma}^2}{max \hat{\alpha}_i}"><span></span><span></span></span>
推导思路:令<span class="equation-text" data-index="0" data-equation="f^{'}(k)<0,得k^*=\frac{\sigma^2}{max \alpha_i}" contenteditable="false"><span></span><span></span></span>,于是当<span class="equation-text" contenteditable="false" data-index="1" data-equation="0<k<k^*"><span></span><span></span></span>时,f(k)总是在下降<br>所以<span class="equation-text" contenteditable="false" data-index="2" data-equation="f(k^*)<f(0),即MSE(\hat{\beta(k^*)})<MSE(\hat{\beta})"><span></span><span></span></span>
③岭迹法
在同一张图中画出<span class="equation-text" contenteditable="false" data-index="0" data-equation="\hat{\beta_1}(k),\hat{\beta_2}(k)......\hat{\beta_p}(k)"><span></span><span></span></span>随k变化的的趋势,选逐渐趋向于平稳时的k值
四、假设检验与预测
引入
<span class="equation-text" data-index="0" data-equation="y=X\beta+e,e\sim N(0,\sigma^2I_n),\theta =(\beta,\sigma^2)" contenteditable="false"><span></span><span></span></span><br>假设<span class="equation-text" data-index="1" data-equation="H_0:H\beta=d,H" contenteditable="false"><span></span><span></span></span>为满秩矩阵(<span class="equation-text" contenteditable="false" data-index="2" data-equation="r(H_{m\times p})=m"><span></span><span></span></span>)
似然比检验
<span class="equation-text" data-index="0" data-equation="H" contenteditable="false"><span></span><span></span></span>下:<span class="equation-text" data-index="1" data-equation="\begin{cases}\hat{\beta}=(X{'}X)^{-1}X{'}Y\\\hat{\sigma^2}=\frac{SSE}{n}=\frac{||y-X\hat{\beta}||^2}{n} \\\end{cases}" contenteditable="false"><span></span><span></span></span>(<span class="equation-text" data-index="2" data-equation="\sigma^2" contenteditable="false"><span></span><span></span></span>的最小二乘估计是<span class="equation-text" contenteditable="false" data-index="3" data-equation="\frac{SSE}{n-p}"><span></span><span></span></span>)
<span class="equation-text" data-index="0" data-equation="H_0" contenteditable="false"><span></span><span></span></span>下:<span class="equation-text" data-index="1" data-equation="\begin{cases}\hat{\beta}_H=\hat{\beta}-(X^{'}X)^{-1}H^{'}(H(X^{'}X)^{-1}H^{'})^{-1}(H\hat{\beta}-d) \\\hat{\sigma^2}_H=\frac{SSE_H}{n}=\frac{||y-X\hat{\beta}_H||^2}{n} \\\end{cases}" contenteditable="false"><span></span><span></span></span>
记忆技巧:可以把<span class="equation-text" contenteditable="false" data-index="0" data-equation="(X^{'}X)^{-1}H^{'}"><span></span><span></span></span>当成一个整体来记
似然比统计量
<span class="equation-text" contenteditable="false" data-index="0" data-equation="\lambda(y)=\frac{L(\hat{\beta},\hat{\sigma}^2,y)}{L(\hat{\beta}_H,\hat{\sigma}^2_H,y)}=(\frac{||y-X\hat{\beta}_H||^2}{||y-X\hat{\beta}||^2})^{-\frac{n}{2}}"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="F=\frac{(SSE_H-SSE)/m}{SSE/(n-p)}=\frac{n-p}{m}(\frac{SSE_H}{SSE}-1)=\frac{n-p}{m}(\lambda(y)^{\frac{2}{n}}-1)"><span></span><span></span></span>,单调递增
4.1 一般线性假设
考虑正态线性回归模型:<span class="equation-text" data-index="0" data-equation="y=X\beta+e,e\sim N(0,\sigma^2I_n),X列满秩(r(X_{n\times p})=p)" contenteditable="false"><span></span><span></span></span><br>假设<span class="equation-text" contenteditable="false" data-index="1" data-equation="H_0:H\beta=d,H行满秩(r(H_{m\times p})=m)"><span></span><span></span></span>
定理4.1.1
①<span class="equation-text" data-index="0" data-equation="\frac{SSE}{\sigma^2}\sim \chi^2_{n-p}" contenteditable="false"><span></span><span></span></span><br>②原假设成立时,<span class="equation-text" data-index="1" data-equation="\frac{SSE_H-SSE}{\sigma^2}\sim\chi^2_m" contenteditable="false"><span></span><span></span></span>(原假设不成立时,<span class="equation-text" data-index="2" data-equation="\frac{SSE_H-SSE}{\sigma^2}\sim\chi^2_{m,\delta},非中心\chi^2" contenteditable="false"><span></span><span></span></span>)<br>③<span class="equation-text" data-index="3" data-equation="SSE_H-SSE与SSE 独立" contenteditable="false"><span></span><span></span></span><br>④原假设成立时,<span class="equation-text" data-index="4" data-equation="F=\frac{(SSE_H-SSE)/m}{SSE/(n-p)}\sim F(m,n-p)" contenteditable="false"><span></span><span></span></span>(消掉了未知参数<span class="equation-text" contenteditable="false" data-index="5" data-equation="\sigma^2"><span></span><span></span></span>)<br>
证明思路:①见定理3.2.4<br>②<span class="equation-text" data-index="0" data-equation="SSE_H=||y-X\hat{\beta}_H||^2=||y-X\hat{\beta}+X\hat{\beta}-X\hat{\beta}_H||^2" contenteditable="false"><span></span><span></span></span><br> <span class="equation-text" data-index="1" data-equation="=SSE+(\hat{\beta}-\hat{\beta}_H)^{'}X^{'}X(\hat{\beta}-\hat{\beta}_H)" contenteditable="false"><span></span><span></span></span><br> <span class="equation-text" data-index="2" data-equation="=SSE+(H\hat{\beta}-d)^{'}(H(X^{'}X)^{-1}H^{'})^{-1}(H\hat{\beta}-d)" contenteditable="false"><span></span><span></span></span><br><span class="equation-text" data-index="3" data-equation="\hat{\beta}\sim N(\beta,\sigma^2(X^{'}X)^{-1})\longrightarrow H\hat{\beta}-d\sim N(H\beta-d,\sigma^2H(X^{'}X)^{-1}H^{'})" contenteditable="false"><span></span><span></span></span><br>原假设成立时,<span class="equation-text" data-index="4" data-equation="H\beta-d=0\longrightarrow H\hat{\beta}-d\sim N(0,\sigma^2H(X^{'}X)^{-1}H^{'})" contenteditable="false"><span></span><span></span></span><br><span class="equation-text" data-index="5" data-equation="对H\hat{\beta}-d" contenteditable="false"><span></span><span></span></span>作标椎化后再平方,即可<br>③<span class="equation-text" data-index="6" data-equation="SSE_H-SSE=(H\hat{\beta}-d)^{'}(H(X^{'}X)^{-1}H^{'})^{-1}(H\hat{\beta}-d)" contenteditable="false"><span></span><span></span></span>,拆开后可化为<span class="equation-text" data-index="7" data-equation="y^{'}By+Cy+\eta" contenteditable="false"><span></span><span></span></span>的形式<br>由定理2.4.6得,<span class="equation-text" data-index="8" data-equation="y^{'}(I-P_x)y\perp y^{'}By" contenteditable="false"><span></span><span></span></span>;由定理2.4.5得<span class="equation-text" contenteditable="false" data-index="9" data-equation="y^{'}(I-P_x)y\perp Cy"><span></span><span></span></span><br>④易证
F计算
(复杂版)引理:(1)<span class="equation-text" data-index="0" data-equation="SSE=||y-X\hat{\beta}||^2=||\hat{e}||^2=SST-SSR=y^{'}y-\hat{\beta}^{'}X{'}y" contenteditable="false"><span></span><span></span></span><br> (2)<span class="equation-text" contenteditable="false" data-index="1" data-equation="SSE_H=||y-X\hat{\beta}_H||^2=||\hat{e}_H||^2=SST-SSR_H=y^{'}y-(\hat{\beta}_H^{'}X{'}y-d^{'}\hat{\lambda})"><span></span><span></span></span>
证明:(2)<span class="equation-text" data-index="0" data-equation="SSE_H=||y-X\hat{\beta}_H||^2=(y-X\hat{\beta}_H)^{'}(y-X\hat{\beta}_H)" contenteditable="false"><span></span><span></span></span><br> ha<span class="equation-text" data-index="1" data-equation="=y^{'}y-\hat{\beta}_H^{'}X{'}y-\hat{\beta}_H^{'}X{'}y+\hat{\beta}_H^{'}X{'}X\hat{\beta}_H" contenteditable="false"><span></span><span></span></span><br> ha<span class="equation-text" data-index="2" data-equation="=y^{'}y-\hat{\beta}_H^{'}X{'}y-\hat{\beta}_H^{'}(X{'}y-X{'}X\hat{\beta}_H)" contenteditable="false"><span></span><span></span></span><br> ha<span class="equation-text" data-index="3" data-equation="=y^{'}y-\hat{\beta}_H^{'}X{'}y-\hat{\beta}_H^{'}H^{'}\hat{\lambda}" contenteditable="false"><span></span><span></span></span>(用到了有约束最小二乘中的拉格朗日函数对<span class="equation-text" data-index="4" data-equation="\beta" contenteditable="false"><span></span><span></span></span>求导的结果,又叫正则方程)<br> ha<span class="equation-text" data-index="5" data-equation="=y^{'}y-(\hat{\beta}_H^{'}X{'}y-d^{'}\hat{\lambda})" contenteditable="false"><span></span><span></span></span>(带入了<span class="equation-text" contenteditable="false" data-index="6" data-equation="\hat{\beta}_H^{'}H^{'}=d"><span></span><span></span></span>)
此时,F统计量为<span class="equation-text" data-index="0" data-equation="F=\frac{(SSE_H-SSE)/m}{SSE/(n-p)}=\frac{[\hat{\beta}^{'}X{'}y-y^{'}y-(\hat{\beta}_H^{'}X{'}y-d^{'}\hat{\lambda})]/m}{SSE/(n-p)}" contenteditable="false"><span></span><span></span></span>,要用到<span class="equation-text" contenteditable="false" data-index="1" data-equation="\beta_H和\lambda"><span></span><span></span></span>的估计,比较复杂
(简化版)约简模型
假设<span class="equation-text" contenteditable="false" data-index="0" data-equation="H_0:\beta_1=\beta_2=\beta_3"><span></span><span></span></span>
Step1.先把模型写成样本形式:<span class="equation-text" contenteditable="false" data-index="0" data-equation="y_i=\beta_0+X_{i1}\beta_1+X_{i2}\beta_2+...+X_{i,p-1}\beta_{p-1}+e_i,其中1\leq i\leq n"><span></span><span></span></span>
Step2.转换原假设:<span class="equation-text" contenteditable="false" data-index="0" data-equation="H_0:\beta_1=\beta_2=\beta_3\iff \begin{bmatrix}0& 1&-1&0\cdots & 0\\0& 1 &0&-1\cdots & 0\end{bmatrix}\begin{bmatrix}\beta_0\\\beta_1 \\ \cdots &\\\beta_{p-1}\end{bmatrix}=\begin{bmatrix}0\\0 \\ \cdots &\\0\end{bmatrix}=d"><span></span><span></span></span>
Step3.带入原模型:<span class="equation-text" data-index="0" data-equation="y_i=\beta_0+(X_{i1}+X_{i2}+X_{i3})\beta_1...+X_{i,p-1}\beta_{p-1}+e_i" contenteditable="false"><span></span><span></span></span>,即<span class="equation-text" contenteditable="false" data-index="1" data-equation="y=X_s\beta_s+e,e\sim N(0,\sigma^2I_n),其中\beta_s=(\beta_0,\beta_1,\beta_4...\beta_{p-1})^{'},X_s=(1,X_1+X_2+X_3,X_4...X_{p-1})^{'}"><span></span><span></span></span>(在原假设成立下)
<span class="equation-text" data-index="0" data-equation="\hat{\beta}_s=(X_s^{'}X_s)^{-1}X_s^{'}y" contenteditable="false"><span></span><span></span></span>(约简模型中参数估计没有<span class="equation-text" contenteditable="false" data-index="1" data-equation="\hat{\beta}_H"><span></span><span></span></span>那么复杂!)
<span class="equation-text" data-index="0" data-equation="SSE_H=y^{'}y-\hat{\beta}_s\hat{X}_sy" contenteditable="false"><span></span><span></span></span>(无需计算<span class="equation-text" data-index="1" data-equation="\hat{\beta}_H" contenteditable="false"><span></span><span></span></span>)<span class="equation-text" contenteditable="false" data-index="2" data-equation="\Longrightarrow F统计量也不用计算\hat{\beta}_H"><span></span><span></span></span>
4.2回归方程的显著性检验<br>(检验模型够不够大,是否要再加<span class="equation-text" contenteditable="false" data-index="0" data-equation="X_i"><span></span><span></span></span>)
原假设<span class="equation-text" contenteditable="false" data-index="0" data-equation="H_0:\beta_1=\beta_2=…=\beta_{p-1}=0\iff \begin{bmatrix}0&1& \cdots & 0\\\vdots &\vdots& \ddots & \vdots \\0& 0&\cdots & 1\end{bmatrix}\begin{bmatrix}\beta_0\\\beta_1 \\ \cdots &\\\beta_{p-1}\end{bmatrix}=\begin{bmatrix}0\\0 \\ \cdots &\\0\end{bmatrix}=d"><span></span><span></span></span>
求SSE
<span class="equation-text" contenteditable="false" data-index="0" data-equation="\hat{\beta}=(X^{'}X)^{-1}X^{'}y\Longrightarrow SSE=y^{'}y-\hat{\beta}^{'}X^{'}y=y^{'}(I-P_x)y"><span></span><span></span></span>
求<span class="equation-text" contenteditable="false" data-index="0" data-equation="SSE_H"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="\hat{\beta}_s=(X_s^{'}X_s)^{-1}X_s^{'}y=\bar{y}\Longrightarrow SSE_H=y^{'}y-\hat{\beta_s}^{'}X_s^{'}y=y^{'}y-\bar{y}^{'}1^{'}y=y^{'}y-n\bar{y}^{2}=\sum(y_i-\bar{y})^2"><span></span><span></span></span>
求F统计量
<span class="equation-text" data-index="0" data-equation="F=\frac{(SSE_H-SSE)/m}{SSE/(n-p)}=\frac{(\hat{\beta}^{'}X^{'}y-n\bar{y}^2)/(p-1)}{(y^{'}y-\hat{\beta}^{'}X^{'}y)/(n-p)}\sim F(p-1,n-p)" contenteditable="false"><span></span><span></span></span>(在<span class="equation-text" contenteditable="false" data-index="1" data-equation="H_0"><span></span><span></span></span>成立下)
求拒绝域
<span class="equation-text" data-index="0" data-equation="F>F_{p-1,n-p}(\alpha)" contenteditable="false"><span></span><span></span></span>(<font color="#ff0000">不是双边</font>!)
4.3回归系数的显著性检验<br>(检验模型够不够小,是否要剔除<span class="equation-text" contenteditable="false" data-index="0" data-equation="X_i"><span></span><span></span></span>)
原假设<span class="equation-text" contenteditable="false" data-index="0" data-equation="H_0:\beta_i=0"><span></span><span></span></span>
求统计量
<span class="equation-text" data-index="0" data-equation="\hat{\beta}=(X^{'}X)^{-1}X^{'}y\sim N(\beta,\sigma^2(X^{'}X)^{-1})\Longrightarrow \hat{\beta}_i\sim N(\beta_i,\sigma^2(X_{i+1}^{'}X_{i+1})^{-1})" contenteditable="false"><span></span><span></span></span><span class="equation-text" contenteditable="false" data-index="1" data-equation="\implies \frac{\hat{\beta}_i}{\sigma\sqrt{(X^{'}_{i+1}X_{i+1})^{-1}}}\sim N(0,1)"><span></span><span></span></span>(在<span class="equation-text" data-index="2" data-equation="H_0" contenteditable="false"><span></span><span></span></span>成立下)
还存在<span class="equation-text" contenteditable="false" data-index="0" data-equation="\sigma"><span></span><span></span></span>,不是随机变量
<span class="equation-text" contenteditable="false" data-index="0" data-equation="t_i=\frac{\frac{\hat{\beta}_i}{\sigma\sqrt{(X^{'}_{i+1}X_{i+1})^{-1}}}}{\sqrt{\frac{SSE}{\sigma^2(n-p)}}}=\frac{\hat{\beta}_i}{\hat{\sigma}\sqrt{(X^{'}_{i+1}X_{i+1})^{-1}}}\sim t(n-p)"><span></span><span></span></span>
利用<span class="equation-text" data-index="0" data-equation="\hat{\sigma}^2=\frac{SSE}{n-p}和\frac{SSE}{\sigma^2}\sim \chi^2_{n-p}" contenteditable="false"><span></span><span></span></span>把 <span class="equation-text" data-index="1" data-equation="\sigma" contenteditable="false"><span></span><span></span></span>换成 <span class="equation-text" data-index="2" data-equation="\hat{\sigma}" contenteditable="false"><span></span><span></span></span>,从而去掉<span class="equation-text" contenteditable="false" data-index="3" data-equation="\sigma"><span></span><span></span></span>
求拒绝域
<span class="equation-text" data-index="0" data-equation="|t_i|>t_{n-p}(\frac{\alpha}{2})" contenteditable="false"><span></span><span></span></span>(<font color="#ff0000">双边</font>!)
4.4异常点检测
1.模型(又叫漂移模型)
假设<span class="equation-text" data-index="0" data-equation="(x_j^{'},y_j)" contenteditable="false"><span></span><span></span></span>是异常的(<font color="#ff0000">只有这一组!</font>),则<span class="equation-text" contenteditable="false" data-index="1" data-equation="\begin{cases}y_i=x_i^{'}\beta+e_i\ \ \ \ \ \ \ \ i\neq j \\y_j=x_j^{'}\beta+\eta+e_j \ \ \ \ \ i=j\\\end{cases},e_i\sim N(0,\sigma^2)\Longrightarrow y=X\beta+d_j\eta+e,其中d_j=(0,0....1,0....0)^{'}"><span></span><span></span></span>
<span class="equation-text" data-index="0" data-equation="y=Z\gamma+e,其中\gamma=\begin{pmatrix}\beta \\\eta \end{pmatrix},Z=(X\ \ \ d_j),e\sim N(0,\sigma^2I_n)" contenteditable="false"><span></span><span></span></span><br>原假设<span class="equation-text" contenteditable="false" data-index="1" data-equation="H_0:H\gamma=d,其中H=(0,0...0,1),d=0(其实就是\eta=0)"><span></span><span></span></span>
注意:<span class="equation-text" data-index="0" data-equation="H_0" contenteditable="false"><span></span><span></span></span>成立,漂移模型为原模型;<span class="equation-text" data-index="1" data-equation="H_0" contenteditable="false"><span></span><span></span></span>不成立时,为切割模型
2.计算<span class="equation-text" contenteditable="false" data-index="0" data-equation="SSE_H"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="\hat{\beta}_H=(X^{'}X)^{-1}X^{'}y"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="SSE_H=y^{'}y-\hat{\beta}^{'}X^{'}y\ ,r(H)=1"><span></span><span></span></span>
3.计算SSE
记漂移模型参数估计为<span class="equation-text" contenteditable="false" data-index="0" data-equation="\beta^*,\eta^*"><span></span><span></span></span>
定理4.4.1 <span class="equation-text" contenteditable="false" data-index="0" data-equation="\beta^*=\hat{\beta}_{(j)}=\hat{\beta}-\frac{\hat{e}_j}{1-P_{jj}}(X^{'}X)^{-1}X_j,\eta^*=\frac{\hat{e}_{j}}{1-P_{jj}}"><span></span><span></span></span>
证明:<span class="equation-text" contenteditable="false" data-index="0" data-equation="\gamma^*=\begin{pmatrix}\beta^* \\\eta^* \end{pmatrix}=(Z^{'}Z)^{-1}Z^{'}y=\begin{pmatrix}X^{'}X & X^{'}d_j\\d_j^{'}X & d_j^{'}d_j\end{pmatrix}^{-1}\begin{pmatrix}X^{'}y \\y_j \end{pmatrix}"><span></span><span></span></span>,然后再拆开就行
SSE<span class="equation-text" contenteditable="false" data-index="0" data-equation="=y^{'}y-\begin{pmatrix}\beta^* \\\eta^* \end{pmatrix}^{'}(X \ \ \ d_j)^{'}y=y^{'}y-\beta^{*'}X^{'}y-\eta^{*'}d_j^{'}y\ ,\ r=n-(p+1)"><span></span><span></span></span>
<span class="equation-text" data-index="0" data-equation="SSE_H-SSE=\beta^{*'}X^{'}y+\eta^{*'}d_j^{'}y-\hat{\beta}^{'}X^{'}y\ \ \ \underrightarrow{代入\beta^*}\ \ \frac{\hat{e}_{j}^2}{1-P_{jj}}" contenteditable="false"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="SSE=SSE-SSE_H+SSE_H=-\frac{\hat{e}_{j}^2}{1-P_{jj}}+y^{'}y-\hat{\beta}^{'}X^{'}y=(n-p)\hat{\sigma}^2-\frac{\hat{e}_{j}^2}{1-P_{jj}}"><span></span><span></span></span>
4.计算F统计量
<span class="equation-text" contenteditable="false" data-index="0" data-equation="F=\frac{(SSE_H-SSE)/1}{SSE/(n-p-1)}=\frac{\frac{\hat{e}_{j}^2}{1-P_{jj}}}{\frac{(n-p)\hat{\sigma}^2-\frac{\hat{e}_{j}^2}{1-P_{jj}}}{(n-p-1)}}\ \ \underrightarrow{r_j=\frac{\hat{e}_{j}}{\hat{\sigma}\sqrt{1-P_{jj}}}}\ \ \frac{(n-p-1)r_j^2}{n-p-r_j^2}\sim F(1,n-p-1)"><span></span><span></span></span>
4.5 因变量的预测
点预测
考虑模型:<span class="equation-text" data-index="0" data-equation="y_i=X_i^{'}\beta+e_i,Ee_i=0,cov(e_i,e_j)=\sigma^2" contenteditable="false"><span></span><span></span></span>,在样本点<span class="equation-text" data-index="1" data-equation="x_0=(1,X_{0,1}…X_{0,p-1})" contenteditable="false"><span></span><span></span></span>处,<span class="equation-text" data-index="2" data-equation="\begin{cases}y_0=x_0^{'}\beta+e_0 \\E(e_0)=0 \\cov(e_0,e_i)=0,D(e_0)=\sigma^2\end{cases}" contenteditable="false"><span></span><span></span></span><span class="equation-text" data-index="3" data-equation="\ \Longrightarrow\ \ 点预测为y_0^*=X_0^{'}\hat{\beta}=X_0^{'}(X^{'}X)^{-1}X^{'}y" contenteditable="false"><span></span><span></span></span>(正好均值估计<span class="equation-text" contenteditable="false" data-index="4" data-equation="\hat{\mu}_0"><span></span><span></span></span>一样了)
性质
a.无偏预测(和无偏估计不同)
<span class="equation-text" contenteditable="false" data-index="0" data-equation="E(\hat{y}_0^*)=E(X_0^{'}\hat{\beta})=X_0^{'}\beta=Ey_0"><span></span><span></span></span>
b.最佳线性无偏预测BLUP
<span class="equation-text" contenteditable="false" data-index="0" data-equation="D(a^{'}y)\geq D(y_0^*)"><span></span><span></span></span>
c.估计效果>预测
证明:估计偏差<span class="equation-text" data-index="0" data-equation="d=\hat{\mu}_0-\mu_0," contenteditable="false"><span></span><span></span></span>预测偏差<span class="equation-text" data-index="1" data-equation="z=y_0^*-y_0" contenteditable="false"><span></span><span></span></span><br> 由于一阶距为0(均无偏),考虑二阶矩<br> <span class="equation-text" data-index="2" data-equation="cov(z)=cov(y_0^*-y_0)=cov(y_0^*)+cov(y_0)-2cov(y_0^*,y_0)" contenteditable="false"><span></span><span></span></span><br> <span class="equation-text" data-index="3" data-equation="=cov(X_0^{'}\hat{\beta})+\sigma^2-2cov(X_0^{'}(X^{'}X)^{-1}X^{'}y,y_0)" contenteditable="false"><span></span><span></span></span>(<span class="equation-text" data-index="4" data-equation="y和y_0的" contenteditable="false"><span></span><span></span></span>随机性部分只有<span class="equation-text" data-index="5" data-equation="e和e_0" contenteditable="false"><span></span><span></span></span>,所以协方差部分为0<span class="equation-text" data-index="6" data-equation="\Longrightarrow y_0与y_0^*独立" contenteditable="false"><span></span><span></span></span>)<br> <span class="equation-text" data-index="7" data-equation="=\sigma^2X_0^{'}(X^{'}X)^{-1}X_0+\sigma^2" contenteditable="false"><span></span><span></span></span><br> <span class="equation-text" contenteditable="false" data-index="8" data-equation="cov(d)=cov(\hat{\mu}_0)=cov(X_0^{'}\hat{\beta})=\sigma^2X_0^{'}(X^{'}X)^{-1}X_0"><span></span><span></span></span>
区间预测
求统计量
<span class="equation-text" data-index="0" data-equation="z=y_0^*-y_0 \sim N(0,\sigma^2X_0^{'}(X^{'}X)^{-1}X_0+\sigma^2)" contenteditable="false"><span></span><span></span></span><span class="equation-text" contenteditable="false" data-index="1" data-equation="\Longrightarrow \frac{y_0^*-y_0}{\sqrt{\sigma^2(X_0(X^{'}X)^{-1}X_0+1)}}\sim N(0,1)"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="\hat{\sigma}^2=\frac{SSE}{n-p},\frac{(n-p)\hat{\sigma}^2}{\sigma^2}=\frac{SSE}{\sigma^2}\sim \chi^2(n-p)"><span></span><span></span></span>
上面两式相除,得<span class="equation-text" contenteditable="false" data-index="0" data-equation="\frac{y_0^*-y_0}{\hat{\sigma}\sqrt{X_0(X^{'}X)^{-1}X_0+1}}\sim t(n-p)"><span></span><span></span></span>
求<span class="equation-text" contenteditable="false" data-index="0" data-equation="1-\alpha"><span></span><span></span></span>置信区间
<span class="equation-text" contenteditable="false" data-index="0" data-equation="y_0^*\pm t_{n-p}^{-1}(\frac{\alpha}{2})\hat{\sigma}\sqrt{X_0(X^{'}X)^{-1}X_0+1}"><span></span><span></span></span>
五、回归方程的选择
包含回归方程的选择和自变量的选择
回归方程:线性回归 or 非线性回归
统计上称之为模型的线性性检验
自变量的选择:从与因变量保持线性关系的自变量集合中,选择一个最优的自变量集合
5.1 变量的选择
1.全模型&选模型
full model : <span class="equation-text" data-index="0" data-equation="y=X\beta+e,E(e)=0,cov(e)=\sigma^2I_n\\" contenteditable="false"><span></span><span></span></span><br> <span class="equation-text" contenteditable="false" data-index="1" data-equation="其中X=(1,X_1…X_{q-1}, X_q…X_{p-1}),\beta=\begin{pmatrix}\beta_q \\\beta_t \end{pmatrix}"><span></span><span></span></span>
selective model : <span class="equation-text" contenteditable="false" data-index="0" data-equation="y=X_q\beta_q+e"><span></span><span></span></span>
2.估计
①<span class="equation-text" data-index="0" data-equation="\hat{\beta}=" contenteditable="false"><span></span><span></span></span><span class="equation-text" data-index="1" data-equation="(X^{'}X)^{-1}X^{'}y=(\hat{\beta}_q^{'},\hat{\beta}_t^{'})" contenteditable="false"><span></span><span></span></span>,<span class="equation-text" contenteditable="false" data-index="2" data-equation="\ \tilde{\beta}_q=(X_q^{'}X_q)^{-1}X_q^{'}y"><span></span><span></span></span>
<b>定理5.1.1 </b>假设全模型正确,则<br> a. <span class="equation-text" data-index="0" data-equation="E(\tilde{\beta}_q)=\beta_q+A\beta_t,A=(X_q^{'}X_q)^{-1}X_q^{'}X_t" contenteditable="false"><span></span><span></span></span>(<font color="#ff0000">有偏</font>!)<br> b. <span class="equation-text" data-index="1" data-equation="cov(\hat{\beta}_q)>cov(\tilde{\beta}_q)" contenteditable="false"><span></span><span></span></span>
证明:(a)<span class="equation-text" data-index="0" data-equation="\tilde{\beta}_q=(X_q^{'}X_q)^{-1}X_q^{'}y,E(\tilde{\beta}_q)=(X_q^{'}X_q)^{-1}X_q^{'}E(y)=(X_q^{'}X_q)^{-1}X_q^{'}\begin{pmatrix}X_q&X_t \end{pmatrix}\begin{pmatrix}\beta_q \\\beta_t \end{pmatrix}=\beta_q+A\beta_t" contenteditable="false"><span></span><span></span></span><br> (b)<span class="equation-text" data-index="1" data-equation="cov(\hat{\beta})=\sigma^2(X^{'}X)^{-1}=\sigma^2\begin{pmatrix}X_q^{'}X_q & X_q^{'}X_t \\X_t^{'}X_q & X_t^{'}X_t\end{pmatrix}^{-1}" contenteditable="false"><span></span><span></span></span><br> <span class="equation-text" data-index="2" data-equation="cov(\hat{\beta}_q)=\sigma^2((X_q^{'}X_q)^{-1}+ADA^{'})" contenteditable="false"><span></span><span></span></span><br> <span class="equation-text" contenteditable="false" data-index="3" data-equation="cov(\tilde{\beta}_q)=\sigma^2(X_q^{'}X_q)^{-1}"><span></span><span></span></span>
注意:两种极端情况<br>①<span class="equation-text" contenteditable="false" data-index="0" data-equation="X_q与X_t"><span></span><span></span></span>无关,此时A=0,选择全模型<br>②<span class="katex"><span class="katex-html" aria-hidden="true"><span class="base"><span class="mord"><span class="mord mathdefault" style="margin-right:0.07847em;">X</span><span class="msupsub"><span class="vlist-t vlist-t2"><span class="vlist-r"><span class="vlist" style="height:0.15139200000000003em;"><span style="top:-2.5500000000000003em;margin-left:-0.07847em;margin-right:0.05em;"><span class="pstrut" style="height:2.7em;"></span><span class="sizing reset-size6 size3 mtight"><span class="mord mathdefault mtight" style="margin-right:0.03588em;">q</span></span></span></span><span class="vlist-s"></span></span><span class="vlist-r"><span class="vlist" style="height:0.286108em;"></span></span></span></span></span><span class="mord cjk_fallback">与</span><span class="mord"><span class="mord mathdefault" style="margin-right:0.07847em;">X</span><span class="msupsub"><span class="vlist-t vlist-t2"><span class="vlist-r"><span class="vlist" style="height:0.2805559999999999em;"><span style="top:-2.5500000000000003em;margin-left:-0.07847em;margin-right:0.05em;"><span class="pstrut" style="height:2.7em;"></span><span class="sizing reset-size6 size3 mtight"><span class="mord mathdefault mtight">t</span></span></span></span><span class="vlist-s"></span></span><span class="vlist-r"><span class="vlist" style="height:0.15em;"></span></span></span></span></span></span></span></span>完全线性,选择选模型
② <span class="equation-text" data-index="0" data-equation="\tilde{\beta}" contenteditable="false"><span></span><span></span></span>有偏<span class="equation-text" contenteditable="false" data-index="1" data-equation="\Longrightarrow MSEM(\tilde{\beta})"><span></span><span></span></span>
<font color="#ff0000">注意:</font><span class="equation-text" data-index="0" data-equation="MSE(\hat{\theta})=E((\hat{\theta}-\theta)^{'}(\hat{\theta}-\theta))" contenteditable="false"><span></span><span></span></span>,是个数(行x列)<br><span class="equation-text" data-index="1" data-equation="MSEM(\hat{\theta})=E((\hat{\theta}-\theta)(\hat{\theta}-\theta)^{'})" contenteditable="false"><span></span><span></span></span>,是个矩阵(列x行)
<b>定理5.1.2</b> 假设全模型正确,则<span class="equation-text" data-index="0" data-equation="cov(\hat{\beta}_t)\geq \beta_t\beta_t^{'}\Longrightarrow MSEM(\hat{\beta}_q)\geq MSEM(\tilde{\beta}_q)" contenteditable="false"><span></span><span></span></span><br>
证明:<span class="equation-text" data-index="0" data-equation="MSEM(\tilde{\beta}_q)=cov(\tilde{\beta}_q)+E(\tilde{\beta}_q-\beta_q)(\tilde{\beta}_q-\beta_q)^{'}=\sigma^2(X_q^{'}X_q)^{-1}+A\beta_t\beta_t^{'}A" contenteditable="false"><span></span><span></span></span><br> <span class="equation-text" data-index="1" data-equation="MSEM(\hat{\beta}_q)=cov(\hat{\beta}_q)=\sigma^2((X_q^{'}X_q)^{-1}+ADA^{'})=\sigma^2(X_q^{'}X_q)^{-1}+Acov(\hat{\beta}_t)A^{'}" contenteditable="false"><span></span><span></span></span>
就是说当不选的变量波动大时,用选模型
3.预测
①已知<span class="equation-text" data-index="0" data-equation="X_0=(X_{0q}^{'},X_{0q}^{'})^{'}," contenteditable="false"><span></span><span></span></span>预测<span class="equation-text" contenteditable="false" data-index="1" data-equation="y_0"><span></span><span></span></span>
<span class="equation-text" data-index="0" data-equation="y_0=X_0^{'}\beta+e_0=(X_{0q}^{'}\ \ X_{0t}^{'})^{'}\begin{pmatrix}\beta_q \\\beta_t \end{pmatrix}+e=X_{0q}^{'}\beta_q+X_{0t}^{'}\beta_t +e_0" contenteditable="false"><span></span><span></span></span><br><span class="equation-text" data-index="1" data-equation="E(e_0)=0,D(e_0)=\sigma^2,cov(e_0,e)=0" contenteditable="false"><span></span><span></span></span><br>full : <span class="equation-text" data-index="2" data-equation="\hat{y}_0=X_0\hat{\beta}\ ,\ Z=\hat{y}_0-y_0=X_0^{'}\hat{\beta}-y_0" contenteditable="false"><span></span><span></span></span>(全模型正确时Z=0)<br>selective : <span class="equation-text" contenteditable="false" data-index="3" data-equation="\tilde{y}_{0q}=X_{0q}^{'}\tilde{\beta}_q\ ,\ Z_q=\tilde{y}_{0q}-y_0=X_{0q}^{'}\tilde{\beta}_q-y_0"><span></span><span></span></span>
<b>定理5.1.3</b> 若全模型正确,则<span class="equation-text" data-index="0" data-equation="(a)EZ_q=X_{0q}^{'}A\beta_t-X_{0t}^{'}\beta_t,A=" contenteditable="false"><span></span><span></span></span><span class="equation-text" data-index="1" data-equation="(X_q^{'}X_q)^{-1}X_q^{'}X_t" contenteditable="false"><span></span><span></span></span><br> <span class="equation-text" contenteditable="false" data-index="2" data-equation=" (b)var(Z)\geq var(Z_q)"><span></span><span></span></span>
证明:(a)<span class="equation-text" data-index="0" data-equation="EZ_q=E(\tilde{y}_{0q}-y_0)=E(X_{0q}^{'}\tilde{\beta}_q)-E(X_{0q}^{'}\beta_q+X_{0t}^{'}\beta_t +e_0)=X_{0q}^{'}(\beta_q+A\beta_t)-X_{0q}^{'}\beta_q-X_{0t}^{'}\beta_t =X_{0q}^{'}A\beta_t-X_{0t}^{'}\beta_t " contenteditable="false"><span></span><span></span></span><br> (b)<span class="equation-text" data-index="1" data-equation="var(Z)=var(X_0^{'}\hat{\beta}-y_0)=var(X_0^{'}\hat{\beta})+var(y_0)-2cov(X_0^{'}\hat{\beta},y_0)=\sigma^2X_0^{'}(X^{'}X)^{-1}X_0+\sigma^2-0" contenteditable="false"><span></span><span></span></span><br> <span class="equation-text" data-index="2" data-equation=" var(Z_q)=var(X_{0q}^{'}\tilde{\beta}_q-y_0)=\sigma^2X_{0q}^{'}(X_q^{'}X_q)^{-1}X_{0q}+\sigma^2-0" contenteditable="false"><span></span><span></span></span><br> (转化为了两个二次型的比较)<br> 然后,计算<span class="equation-text" data-index="3" data-equation="var(Z)-var(Z_q)," contenteditable="false"><span></span><span></span></span>把<span class="equation-text" data-index="4" data-equation="X=\begin{pmatrix}X_q \\X_t \end{pmatrix}" contenteditable="false"><span></span><span></span></span>代入,重点是分块矩阵求逆,<br> 最终化简得到<span class="equation-text" contenteditable="false" data-index="5" data-equation="var(Z)-var(Z_q)=\sigma^2(X_{0q}^{'}A-X_{0t}^{'})D(X_{0q}^{'}A-X_{0t}^{'})^{'}\geq 0"><span></span><span></span></span><br>
②<span class="equation-text" contenteditable="false" data-index="0" data-equation="\tilde{y}_0\ ,\ bias"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="MSEP(\tilde{y}_0)=E(\tilde{y}_0-y_0)=E(Z_q^2)=var(Z_q)+E(Z_q)^2"><span></span><span></span></span>
<b>定理5.1.4</b> 若全模型正确,<span class="equation-text" data-index="0" data-equation="cov(\hat{\beta}_t)\geq\beta_t\beta_t^{'}\Longrightarrow MSEP(\hat{y}_0)\geq MSEP(\tilde{y}_0)" contenteditable="false"><span></span><span></span></span>,选择选模型
证明:<span class="equation-text" data-index="0" data-equation="MSEP(\hat{y}_0)=var(Z) +0=var(Z)" contenteditable="false"><span></span><span></span></span><br> <span class="equation-text" data-index="1" data-equation="MSE(\tilde{y}_0)=var(Z_q)+(EZ_q)^2=var(Z_q)+(X_{0q}^{'}A\beta_t-X_{0t}^{'}\beta_t)(X_{0q}^{'}A\beta_t-X_{0t}^{'}\beta_t)^{'}" contenteditable="false"><span></span><span></span></span><br> ha<span class="equation-text" data-index="2" data-equation="=var(Z_q)+(X_{0q}^{'}A-X_{0t}^{'})\beta_t\beta_t^{'}(X_{0q}^{'}A-X_{0t}^{'})^{'}" contenteditable="false"><span></span><span></span></span><br> ha<span class="equation-text" contenteditable="false" data-index="3" data-equation="\leq var(Z_q)+(X_{0q}^{'}A-X_{0t}^{'})cov(\beta_t)(X_{0q}^{'}A-X_{0t}^{'})^{'}=var(Z_q)+var(Z)-var(Z_q)=var(Z_q)=MSE(Z_q)"><span></span><span></span></span>
4.评判准则
①<span class="equation-text" contenteditable="false" data-index="0" data-equation="RMS_q=\frac{SSE}{n-q}=\frac{y^{'}(I-P_{x_q})y}{n-q}=\hat{\sigma^2}_q"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="q\uparrow,SSE\downarrow"><span></span><span></span></span>
证明:<span class="equation-text" data-index="0" data-equation="X_q=(1,X_{(1)}...X_{(q-1)}),X_{q+1}=(1,X_{(1)}...X_{(q-1)},X_{(q)})=(X_{(q-1)},X_{(q)})" contenteditable="false"><span></span><span></span></span><br> <span class="equation-text" data-index="1" data-equation="P_{X_{q+1}}=(X_q\ X_{(q)})\begin{pmatrix}X_q^{'}X_q& X_q^{'}X_{(q)}\\X_{(q)}^{'}X_q & X_{(q)}^{'}X_{(q)}\end{pmatrix}^{-1}\begin{pmatrix}X_q\\X_{(q)}\end{pmatrix}=(X_q\ X_{(q)})\begin{pmatrix}(X_q^{'}X_q)^{-1}+ada^{'}& -ad\\-d^{'}a^{'} & d\end{pmatrix}\begin{pmatrix}X_q\\X_{(q)}\end{pmatrix}" contenteditable="false"><span></span><span></span></span><br> <span class="equation-text" data-index="2" data-equation="\Longrightarrow P_{X_{q+1}}-P_{X_{q}}=X_qada^{'}X_q^{'}-2X_qadX_{(q)}^{'}+X_{(q)}dX_{(q)}^{'}=(X_{q}a-X_{(q)})d(X_{q}a-X_{(q)})^{'}\geq0" contenteditable="false"><span></span><span></span></span><br> <span class="equation-text" contenteditable="false" data-index="3" data-equation="\Longrightarrow SSE_q-SSE_{q+1}=y^{'}( P_{X_{q+1}}- P_{X_{q}})y\geq0"><span></span><span></span></span>
<span class="equation-text" data-index="0" data-equation="\frac{1}{n-q}" contenteditable="false"><span></span><span></span></span>相当于成了个惩罚因子,最后<span class="equation-text" data-index="1" data-equation="RMS_q" contenteditable="false"><span></span><span></span></span>先降后升,<span class="equation-text" contenteditable="false" data-index="2" data-equation="q=arc\ minRMS_q"><span></span><span></span></span>
②<span class="equation-text" data-index="0" data-equation="C_p=\frac{SSE_q}{\hat{\sigma}^2}-(n-2q)" contenteditable="false"><span></span><span></span></span>,其中<span class="equation-text" contenteditable="false" data-index="1" data-equation="\hat{\sigma}^2=\frac{SSE}{n-p}"><span></span><span></span></span>
从预测精度出发,<span class="equation-text" contenteditable="false" data-index="0" data-equation="MSEP(\hat{y})=E(\hat{y}-y)^2"><span></span><span></span></span>,MSEP愈小愈好
<span class="equation-text" contenteditable="false" data-index="0" data-equation="C_p"><span></span><span></span></span>愈小愈好
③<span class="equation-text" data-index="0" data-equation="AIC=-2lnL(\beta_q,\sigma^2_q|y)+2q" contenteditable="false"><span></span><span></span></span><br> <span class="equation-text" contenteditable="false" data-index="1" data-equation="BIC=nln(SSE_q)+2(lnn)q"><span></span><span></span></span>
从极大似然原理出发
选择使AIC达到最小的自变量子集
5.2 计算所有可能的回归
对p-1个自变量的线性回归,所有可能的回归有<span class="equation-text" contenteditable="false" data-index="0" data-equation="C_{p-1}^1+C_{p-1}^2+...C_{p-1}^{p-1}=2^{p-1}-1"><span></span><span></span></span>个
2. <span class="equation-text" contenteditable="false" data-index="0" data-equation="S_p"><span></span><span></span></span>顺序
<span class="equation-text" contenteditable="false" data-index="0" data-equation="u_i=\begin{cases}1,\ \ \ 选中X_i \\0 \ \ \ 不选X_i \\\end{cases}"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="S _p=\left\{S_{p-1},p,T_{p-1}\right\}"><span></span><span></span></span>
i 表示引入变量<span class="equation-text" data-index="0" data-equation="X_i" contenteditable="false"><span></span><span></span></span> , -i 表示剔除变量<span class="equation-text" data-index="1" data-equation="X_i" contenteditable="false"><span></span><span></span></span> , <span class="equation-text" contenteditable="false" data-index="2" data-equation="S_p"><span></span><span></span></span>里是 i 与-i 的集合。
3. 消去变换
消去变换
<span class="equation-text" contenteditable="false" data-index="0" data-equation="A=(a_{ij})_{m\times n},T_iA=B=\begin{cases}b_{ii}=\frac{1}{a_{ii}} \\b_{ij}=\frac{a_{ij}}{a_{ii}} ,\ \ \ i\neq j\\b_{ji}=-\frac{a_{ji}}{a_{ii}} ,\ \ \ i\neq j\\b_{kl}=a_{kl}-\frac{a_{ki}a_{il}}{a_{ii}},k\neq i,l\neq i\end{cases}"><span></span><span></span></span>
性质
①<span class="katex"><span class="katex-html" aria-hidden="true"><span class="base"><span class="mord"><span class="mord mathdefault" style="margin-right:0.13889em;">T</span><span class="msupsub"><span class="vlist-t vlist-t2"><span class="vlist-r"><span class="vlist" style="height:0.31166399999999994em;"><span style="top:-2.5500000000000003em;margin-left:-0.13889em;margin-right:0.05em;"><span class="pstrut" style="height:2.7em;"></span><span class="sizing reset-size6 size3 mtight"><span class="mord mathdefault mtight">i</span></span></span></span><span class="vlist-s"></span></span><span class="vlist-r"><span class="vlist" style="height:0.15em;"></span></span></span></span></span><span class="mord"><span class="mord mathdefault" style="margin-right:0.13889em;">T</span><span class="msupsub"><span class="vlist-t vlist-t2"><span class="vlist-r"><span class="vlist" style="height:0.31166399999999994em;"><span style="top:-2.5500000000000003em;margin-left:-0.13889em;margin-right:0.05em;"><span class="pstrut" style="height:2.7em;"></span><span class="sizing reset-size6 size3 mtight"><span class="mord mathdefault mtight">i</span></span></span></span><span class="vlist-s"></span></span><span class="vlist-r"><span class="vlist" style="height:0.15em;"></span></span></span></span></span><span class="mord mathdefault">A</span><span class="mspace" style="margin-right:0.2777777777777778em;"></span><span class="mrel">=</span><span class="mspace" style="margin-right:0.2777777777777778em;"></span></span><span class="base"><span class="strut" style="height:0.68333em;vertical-align:0em;"></span><span class="mord mathdefault">A</span></span></span></span>
② <span class="equation-text" contenteditable="false" data-index="0" data-equation="T_iT_jA=T_jT_iA"><span></span><span></span></span>
③ <span class="equation-text" data-index="0" data-equation="A=\begin{pmatrix}A_{11} & A_{12} \\A_{21} & A_{22}\end{pmatrix}_{n\times n}" contenteditable="false"><span></span><span></span></span>,对<span class="equation-text" data-index="1" data-equation="(A_{11})_{q\times q}" contenteditable="false"><span></span><span></span></span>做消去变化,<span class="equation-text" contenteditable="false" data-index="2" data-equation="T_1T_2...T_qA=\begin{pmatrix}A_{11}^{-1}& A_{11}^{-1}A_{12} \\-A_{21}A_{11}^{-1} & A_{22.1}\end{pmatrix}"><span></span><span></span></span>(与求逆不完全一样,看左上角)
<span class="equation-text" contenteditable="false" data-index="0" data-equation="T_1T_2...T_nA=A^{-1}"><span></span><span></span></span>
模型选择上的应用
引入增广矩阵Z,<span class="equation-text" contenteditable="false" data-index="0" data-equation="Z=(X\ \ Y)"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="B=X^{'}X=\begin{pmatrix}X_q^{'} \\X_t^{'} \end{pmatrix}(X_q X_t)=\begin{pmatrix}B_{11} & B_{12} \\B_{21} & B_{22}\end{pmatrix}"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="A_{p+1}=Z^{'}Z=\begin{pmatrix}X^{'} \\Y^{'} \end{pmatrix}\begin{pmatrix}X & Y \end{pmatrix}=\begin{pmatrix}X_q^{'} \\X_t^{'} \\y^{'}\end{pmatrix}\begin{pmatrix}X_q & X_t&y\end{pmatrix}=\begin{pmatrix}B_{11} & B_{12} & X_q^{'}y\\B_{21} & B_{22} &X_t^{'}y \\y^{'}X_q&y^{'}X_t&y^{'}y\end{pmatrix}"><span></span><span></span></span>
对A做消去变换,得<span class="equation-text" contenteditable="false" data-index="0" data-equation="T_1T_2...T_qA=\begin{pmatrix}B_{11}^{-1} & *& B_{11}^{-1}X_q^{'}y\\* & *&*\\*&*&y^{-1}y-y^{'}X_qB_{11}^{-1}X_q^{'}y\end{pmatrix}"><span></span><span></span></span>
应用:①<span class="equation-text" data-index="0" data-equation="B_{11}^{-1}X_q^{'}y" contenteditable="false"><span></span><span></span></span>正好是<span class="equation-text" data-index="1" data-equation="\beta_q" contenteditable="false"><span></span><span></span></span>的最小二乘估计。<br> ②<span class="equation-text" data-index="2" data-equation="y^{-1}y-y^{'}X_qB_{11}^{-1}X_q^{'}y" contenteditable="false"><span></span><span></span></span>正好是选模型的<span class="equation-text" data-index="3" data-equation="SSE_q" contenteditable="false"><span></span><span></span></span><br> ③增加或去掉<span class="equation-text" contenteditable="false" data-index="4" data-equation="X_i"><span></span><span></span></span>,都只需再对应做一次消去变换
①②就是说如果要求选模型的<span class="equation-text" contenteditable="false" data-index="0" data-equation="\beta_q和SSE_q"><span></span><span></span></span>,只需对其作连续的消去变换即可
4.步骤
Step1. 确定<span class="equation-text" contenteditable="false" data-index="0" data-equation="S_p=\left\{S_{p-1},p,T_{p-1} \right\}"><span></span><span></span></span>
Step2. 计算Z,然后计算<span class="equation-text" contenteditable="false" data-index="0" data-equation="A=Z^{'}Z"><span></span><span></span></span>
Step3. 做消去变换,一个一个算,<span class="equation-text" contenteditable="false" data-index="0" data-equation="T_1T_2...T_{p-1}A"><span></span><span></span></span>
Step4. min <span class="equation-text" data-index="0" data-equation="C_p/RMS_q/AIC/BIC" contenteditable="false"><span></span><span></span></span>,确定p,然后确定<span class="equation-text" contenteditable="false" data-index="1" data-equation="\hat{\beta}_{p-1}"><span></span><span></span></span>
5.3 计算最优子集回归
根据 5.2 可确定最优自变量子集
<span class="equation-text" contenteditable="false" data-index="0" data-equation="C_p"><span></span><span></span></span>性质
性质1 若全模型误差<span class="equation-text" data-index="0" data-equation="e\sim N(0,\sigma^2I),则E(C_p)=E(\frac{SSE_q}{\hat{\sigma}^2}-(n-2q))=q+\frac{2r+(n-p)\lambda^2}{n-p-2}," contenteditable="false"><span></span><span></span></span><br> <span class="equation-text" contenteditable="false" data-index="1" data-equation="其中r=p-q,\lambda^2=\frac{\beta_t^{'}D^{-1}\beta_t}{\sigma^2},D^{-1}=X_t^{'}X_t-X_t^{'}X_q(X_q^{'}X_q)^{-1}X_q^{'}X_t=X_t^{'}(I-P)X_t"><span></span><span></span></span>
性质2 条件同性质1,若<span class="equation-text" contenteditable="false" data-index="0" data-equation="\beta_t=0,则E(C_p)=q+\frac{2r}{n-q-2}\approx q"><span></span><span></span></span>
性质3 条件同性质1,若<span class="equation-text" contenteditable="false" data-index="0" data-equation="\beta_t=0,则C_p=q+r(F_r-1),其中F_r\sim F_{r,n-p}"><span></span><span></span></span>
注意:<span class="equation-text" contenteditable="false" data-index="0" data-equation="\beta_t\neq0+n-p\gg r\Longrightarrow E(C_p)\approx q+\lambda^2>q"><span></span><span></span></span>
5.4 逐步回归
引入
当自变量个数>40个,上述方法很麻烦,逐步回归算法是应用最普遍的<b>不用计算所有可能</b>子集回归的变量选择算法
基本思想
若变量偏回归平方和显著——加入,然后所有新老变量逐个检验,不显著的去掉,显著的留下。
步骤
假设已有q个自变量入选,且为前q个,记为<span class="equation-text" contenteditable="false" data-index="0" data-equation="X_{(1)}、X_{(2)}...X_{(q)}"><span></span><span></span></span>
①考虑是否剔除<span class="equation-text" contenteditable="false" data-index="0" data-equation="X_{(q)}"><span></span><span></span></span>
<span class="equation-text" contenteditable="false" data-index="0" data-equation="y=X_{q+1}\beta_{q+1}+e,其中X_{q+1}=(1,X_{(1)},X_{(2)}...X_{(q)})=(X_q,X_{(q)}),\beta_{q+1}=\begin{pmatrix}\beta_q^{'} \\\beta_{(q)}^{'} \end{pmatrix}"><span></span><span></span></span>
<span class="equation-text" data-index="0" data-equation="H_{i_0}:\beta_{(q)}=0" contenteditable="false"><span></span><span></span></span>,做 t 检验,<span class="equation-text" contenteditable="false" data-index="1" data-equation="t_{(q)}=\frac{\hat{\beta}_{(q)}}{\hat{\sigma}^2\sqrt{C_{q+1,q+1}}}\sim t(n-q-1)\Longrightarrow\frac{(n-q-1)\hat{\beta}^2_{(q)}(X_{(q)}^{'}(I-P_{x_q}X_{(q)})}{SSE_{q+1}}\sim F(1,n-q-1)"><span></span><span></span></span>(均为原假设成立情况下,且F统计量不完全是t统计量的平方)
拒绝域:
②计算每个回归自变量的F统计量,排序,一个一个地扔
六、方差分析模型
单因素方差分析
子主题
子主题
七、
评论
0 条评论
下一页