Further research on simple projection prediction under a general linear model

Bo JIANG , Luping WANG , Yongge TIAN

Front. Math. China ›› 2024, Vol. 19 ›› Issue (5) : 255 -275.

PDF (565KB)
Front. Math. China ›› 2024, Vol. 19 ›› Issue (5) : 255 -275. DOI: 10.3868/s140-DDD-024-0016-x
RESEARCH ARTICLE

Further research on simple projection prediction under a general linear model

Author information +
History +
PDF (565KB)

Abstract

Linear model is a kind of important model in statistics. The estimation and prediction of unknown parameters in the model is a basic problem in statistical inference. According to different evaluation criteria, different forms of predictors are obtained. The simple projection prediction (SPP) boasts simple form while the best linear unbiased prediction (BLUP) enjoys very excellent properties. Naturally, the relationship between the two predictors are considered. We give the necessary and sufficient conditions for the equality of SPP and BLUP. Finally, we study the robustness of SPP and BLUP with respect to covariance matrix and illustrate the application of equivalence conditions by use of error uniform correlation models.

Keywords

Linear model / SPP / BLUP / equivalence / robustness

Cite this article

Download citation ▾
Bo JIANG, Luping WANG, Yongge TIAN. Further research on simple projection prediction under a general linear model. Front. Math. China, 2024, 19(5): 255-275 DOI:10.3868/s140-DDD-024-0016-x

登录浏览全文

4963

注册一个新账户 忘记密码

1 Introduction

Linear model is very important in statistics and have important applications in many fields, such as agriculture, medicine, economics, education, and many others. The estimation and prediction of unknown parameters in models are a fundamental problem in statistical inference. Based on different optimal criteria, statisticians can define different estimators and predictors of unknown parameters. Among these estimators and predictors, classic ones include least squares estimation, weighted least squares estimation, simple projection prediction (SPP), optimal linear unbiased prediction (BLUP), etc. Among them, simple projection prediction boasts a simple form while optimal linear unbiased prediction enjoys excellent properties. Naturally, we consider the topic of equality between them, and present the equivalent conditions for equality, to provide a theoretical basis for further statistical inference.

We consider the following classic linear model

M1:y=Xβ+ε,E(ε)=0,D(ε)=Σ11.

Here, y is an observable random vector of order n×1, X is a model matrix of order n×p under random rank, β is an unknown parameter vector of order p×1, ε is an error vector of order n×1, E(ε)=0Rn×1 is a zero mean vector, Σ11Rn×n is a known non-negative definite matrix, denoted by Σ110, where E(·) and D(·) represent the expectation and variance of the random vector, respectively.

In statistical research, it is sometimes necessary to study the problem of predictor of censored data in future observations. Let the future observation vector model of model (1.1) be

M2:y=Xβ+ε,E(ε)=0,D(ε)=Σ22.

Here, y is the future observation vector of order m×1, X is the model matrix of order m×p under random rank, ε is the error vector of order m×1, E(ε)=0Rm×1 is the zero mean vector, and Σ22Rm×m is a known non-negative definite matrix.

The simultaneous model of models (1.1) and (1.2) is

M:y^=Zβ+ε^,E(ε^)=0,D(ε^)=[Σ11Σ12Σ21Σ22]V,

where y^=[yy], Z=[XX], ε^=[εε].

Due to the presence of unknown parameter vectors and error terms in linear models, people usually only focus on the estimation of unknown parameter β, and there have been scarce discussions on the joint prediction of unknown parameter vectors and error terms. Here, we consider the joint prediction problem of unknown parameter vectors and error terms

ϕ=Fβ+Gε+Hε.

Here, FRs×p, GRs×n and HRs×m are known matrices. Since F, G, and H can be any matrix, ϕ contains all linear combinations of unknown parameters β, error terms ε and ε.

(i) If F=Z, G=[In0], H=[0Im], then ϕ is the joint observation vector y^ in model (1.3);

(ii) If F=QZ, G=Q[In0], H=Q[0Im], then ϕ is random linear function of the joint vector in model (1.3), where QRt×(m+n) is the known matrix;

(iii) If F=1m+nZ, G=1n, H=1m, then ϕ is the sum of the observed values of the joint vector y^ in model (1.3), i.e., ϕ=iyi, where 1k represents vectors of order 1×k with all components being 1;

(iv) If F=X, G=0, H=Im, then ϕ is the future observation vector y in model (1.2);

(v) If F=X, G=In, H=0, then ϕ is the known observation vector y in model (1.1).

Reference [17] gave an important lemma for optimizing matrix-valued functions in the sense of Löwner partial order, and studied the joint prediction problem under mixed effect model. In recent years, references [19, 20] presented the BLUP analytical expressions for the joint prediction of fixed effect, random effect, and error term under random effect model by solving the constrained quadratic matrix-valued function optimization problem in the sense of Löwner partial order. This work can be referred to in references [3-5, 7-11].

With (1.3) and (1.4), we can get

E(ϕ)=Fβ,Cov(ϕ)=JVJ,Cov(ϕ,y)=JVT,Cov(y)=Σ11=TVT,

where J=[G,H], T=[In,0].

This paper uses a novel method to solve the constrained quadratic matrix-valued function optimization problem, establishes a complete theoretical system for optimal linear unbiased prediction (BLUP) with respect to ϕ, and presents an analytical expression for BLUP and related algebraic and statistical properties. The main work is divided into the following four parts:

(I) Provide model (1.1) consistency definition and definition of ϕ and Qy^ predictability under model (1.1);

(II) Establish analytical expressions for BLUPM1(ϕ), BLUPM1(Qy^) and SPPM(Qy^)and their algebraic and statistical properties;

(III) Present the equivalence conditions for BLUPM1(Qy^)=SPPM(Qy^);

(IV) Study the robustness of covariance matrices with respect to SPP(Qy^) and BLUP(Qy^).

2 Preliminaries

In this paper, A, r(A) and R(A) are used to indicate the transpose, rank of ARm×n matrix, and linear subspace generated by column vector of matrix A, respectively. Im represents the identity matrix of order m. The Moore-Penrose inverse of matrix A, denoted by A, is defined as the unique solution G that simultaneously satisfies the following four matrix equations AGA=A, GAG=G, (AG)=AG and (GA)=GA. PA, EA, and PA represent orthogonal projection matrices (symmetric idempotent matrices), and PA=AA, EA=A=ImAA and FA=InAA. Given two symmetric matrices A and B of the same order, if AB0, they satisfy Löwner partial order AB.

Matrix theory, as a mathematical operation tool, has been widely used in linear statistical models (see monograph [14]). The rank of a matrix is a rather basic concept and an important algebraic tool in matrix algebra, which can be used to establish and simplify various equations containing generalized inverse matrix operations. In recent decades, researchers at home and abroad have established numerous analytical calculation formulas for matrix operation ranks, used these formulas to handle the construction, calculation, and comparison of a large number of estimators in regression analysis, and derived many simple and practical criteria for determining the equivalence of different estimators, laying an important theoretical basis for subsequent research on statistical inference of more complex statistical models. Some basic formulas for block matrix rank were used in some cases, such as references [12, 14].

Below are some basic formulas for calculating the rank of block matrices. They will be used in the proof of the theorem later.

Lemma 2.1  Assume A is a random matrix, then A=0r(A)=0.

Lemma 2.2 [12]  Let ARm×n, BRm×k, CRl×n, DRl×k. Then

r[A,B]=r(A)+r(EAB)=r(B)+r(EBA),

r[AC]=r(A)+r(CFA)=r(C)+r(AFC),

r[ABCD]=r(A)+r(DCAB),ifR(B)R(A),R(C)R(A).

And the following results hold:

(a) r[A,B]=r(A)R(B)R(A)AAB=BEAB=0.

(b) r[A]=r(A)R(C)R(A)CAA=CCFA=0.

Lemma 2.3 [22]  Let ARm×n, BRm×k, and assume R(A)=R(B). Then

XA=0XB=0.

Lemma 2.4 [13]  The necessary and sufficient condition for the linear matrix equation AX=B to be solvable is r[A,B]=r(A), or AAB=B. Then the general solution of the equation can be expressed as X=AB+(IAA)U, where U is an arbitrary matrix.

Below is the definition of model (1.1) consistency. With Moore-Penrose inverse definition, equation

[X,Σ11][X,Σ11][X,Σ11]=[X,Σ11]

always holds. Thus,

E((In[X,Σ11][X,Σ11])y)=(In[X,Σ11][X,Σ11])Xβ=0,Cov((In[X,Σ11][X,Σ11])y)=(In[X,Σ11][X,Σ11])Σ11(In[X,Σ11][X,Σ11])=0

is established. These two equations indicate that [X,Σ11][X,Σ11]y=y holds true for a probability of 1, or yR[X,Σ11] holds true for a probability of 1. For the definition of model consistency, please refer to [15, 16].

Definition 2.1 If yR[X,Σ11] holds true for a probability of 1, then model M1 is consistent or compatible.

Below is a classic definition of predictability (see [1, 6]).

Definition 2.2 If there exists a linear estimator Ly that satisfies

E(Lyϕ)=0,

then vector ϕ is linearly predictable under model M1.

Theorem 2.1  ϕ is linearly estimable under model M1Fβ is linearly estimable under model M1R(F)R(X). In particular,

(a) y is always linearly predictable under model, and Xβ is always linearly estimable under model M1;

(b) y^ and y are linearly predictable under model M1Zβ,Xβ is linearly estimable under model M1R(X)R(X);

(c) Gy^ is linearly predictable under model M1GZβ is linearly estimable under model M1R((GZ))R(X).

Proof Through Definition 2.2, Φ is linearly predictable under model M1Fβ is linearly estimable under model M1LXβ=FβLX=F.

With Lemma 2.4, the result holds. □

Many topics in statistics are closely related to optimization problems in mathematics. References [1921] provided an analytical solution for the optimization problem of quadratic matrix-valued functions with constraints, and used the following lemma to present an analytical expression for joint prediction of BLUP under the most general assumptions of the model, as well as algebraic and statistical properties.

Lemma 2.5  Let ARp×q, BRn×q, CRp×m, DRn×m, MRm×m be non-negative definite matrices. Assume matrix equation GA=B is solvable with respect to GRn×p, then in the sense of Löwner partial order, the quadratic matrix-valued function with constraint

f(G)=(GC+D)M(GC+D)f(G0)s.t.GA=B

is always solvable with respect to G, and the solution G0 of (2.4) satisfies the following matrix equation

G0[A,CMCA]=[B,DMCA].

Then, the analytical expression of G0, f(G0), and f(G) can be expressed as

G0=argminGA=Bf(G)=[B,DMCA][A,CMCA]+U[A,CMC],

f(G0)=minGA=Bf(G)=WMWWMC(ACMCA)CMW,

f(G)=f(G0)+(GC+D)MC(ACMCA)CM(GC+D),

where W=BAC+D,URn×p is a random matrix.

3 Predictable vector φ under model M1 and its special cases SPP and BLUP

Prediction in linear models has always been a fundamental problem in statistics. Based on different evaluation criteria, different expression forms of predictor can be obtained. The intuitive idea is to seek a concise form of prediction. Särndal and Wright [18] emphasized that prediction should have a concise and intuitive form, and the most concise and intuitive prediction was SPP; another commonly used optimal criterion was to find the prediction with the smallest covariance among all unbiased predictors of unknown parameters in the sense of Löwner partial order, which was the famous BLUP. This paper mainly considers the topic of equality between SPPM(Qy^) and BLUPM1(Qy^). The work in this regard can be seen in [2, 18, 23-25].

Firstly, we provide the analytical expressions for these two basic predictors.

Definition 3.1 If there exists a matrix L^ that under Löwner partial order satisfies

Cov(Lyϕ)Cov(L^yϕ)s.t.E(Lyϕ)=0,

then linear statistic L^y is defined as BLUP of ϕ, denoted by

L^y=BLUPM1(ϕ)=BLUPM1(Fβ+Gε+Hε).

When G=0,H=0,L^ is defined as BLUE of Fβ, denoted by

L^y=BLUEM1(Fβ).

When F=QZ, G=Q[In0], H=Q[0Im], L^ is defined as BLUP of Qy^, denoted by

L^y=BLUPM1(Qy^).

The derivation of the analytical expressions for BLUP/BLUE under model M1 with respect to ϕ and its special cases can be found in [19].

Lemma 3.1  Assume ϕ is predictable under model M1, that is, R(F)R(X), then matrix equation

L[X,Cov(y)X]=[F,Cov{ϕ,y}X]

is solvable, and the general solution of the equation and the BLUP analytical expression of ϕ are

BLUPM1(ϕ)=Ly=([F,CX][X,Σ11X]+U[X,Σ11X])y,

where C=Cov{ϕ,y}, Cov(y)=Σ11, while URs×n is a random matrix.

Assume Qy^ is predictable under model M1, QZβ is estimable under model M1, that is, R((QZ))R(X), then matrix equation

L[X,Cov(y)X]=[QZ,Cov{Qy^,y}X]

is solvable, so that the BLUP/BLUE analytical expressions of Qy^, QZβ, and Qε^ are

BLUPM1(Qy^)=([QZ,QVTX][X,Σ11X]+U1[X,Σ11X])y,BLUEM1(QZβ)=([QZ,0][X,Σ11X]+U2[X,Σ11X])y,BLUEM1(Qε^)=([0,QVTX][X,Σ11X]+U3[X,Σ11X])y,

where T=[In,0], Cov{Qy^,y}=QVT, while UiRt×n is a random matrix. And the following results are established:

(a) R[X,Σ11X]=R[X,XΣ11]=R[X,Σ11].

(b) The only necessary and sufficient condition of coefficient matrix L is r[X,Σ11X]=n.

(c) The only necessary and sufficient condition yR[X,Σ11] of BLUPM1(ϕ) holds true with a probability of 1, namely model M1 is compatible.

(d) The following formula holds:

Cov(BLUPM1(ϕ))=[F,|CX][X,Σ11X]Σ11([F,CX][X,Σ11X]),Cov{BLUPM1(ϕ),ϕ}=([F,CX][X,Σ11X]+U[X,Σ11X])C=([F,CX][X,TVTX]+U[X,TVTX])TVJ=([F,CX][X,TVTX]+U[X,TV])TVJ=[F,CX][X,TVTX]TVJ=[F,CX][X,Σ11X]C,

and

Cov(ϕ)Cov(BLUPM1(ϕ))=JVJ[F,CX][X,Σ11X]Σ11([F,CX][X,X]),

Cov(ϕBLUPM1(ϕ))=([F,CX][X,Σ11X]TJ)V([F,CX][X,Σ11X]TJ),

where T=[In,0].

(e) BLUPM1(ϕ) can be decomposed into

BLUPM1(ϕ)=BLUEM1(Fβ)+BLUPM1(Gε)+BLUPM1(Hε),Cov{BLUEM1(Fβ),BLUPM1(Gε+Hε)}=0,Cov(BLUPM1(ϕ))=Cov(BLUEM1(Fβ))+Cov(BLUPM1(Gε+Hε)).

When the covariance matrix V is a special form of matrix, such as V=diag(V1,V2) or V=diag(σ12In,σ22Im), the above results about BLUP/BLUE can be further simplified. The above results are given under the most general assumptions of the model, which do not require conditions like full rank of the model matrix columns and invertibility of the covariance matrix, and there are no assumptions about the distribution of error terms. Therefore, the above lemma displays generality.

Next, we define the simple projection prediction of Qy^ under model (1.3).

Definition 3.2 The SPP of vector Qy^ under model (1.3) is defined as

SPPM(Qy^)=BLUEM1(QZβ)=([QZ,0][X,Σ11X]+U[X,Σ11X])y,

and

E(SPPM(Qy^))=E(BLUEM1(QZβ))=E(Qy^)=QZβ.

4 Research on the relationship between BLUP and SPP of Qy^

BLUP and SPP, as two basic predictors in mathematical statistics, are based on different optimal criteria and therefore have different analytical expressions. BLUP enjoys many excellent algebraic and statistical properties, and it has the minimum variance in unbiased predictors. Statisticians often use BLUP to measure the effectiveness of other estimators. In order to study the relationship between the two linear estimators G1y and G2y, the following definition is given first.

Definition 4.1 Let y be a random vector, G1 and G2 be two matrices with appropriate order.

(a) If G1=G2, then the two random vectors G1y and G2y are absolutely equal, namely G1y=G2y.

(b) If E(G1yG2y)=0 and Cov(G1yG2y)=0 hold simultaneously, then the two random vectors G1y and G2y are equal with a probability of 1.

The main conclusions of this paper are presented below.

Theorem 4.1  Assume Qy^ is linearly predictable under model M1, then the following conclusions are equivalent:

(i) BLUPM1(Qy^)=SPPM(Qy^).

(ii) BLUPM1(Qy^) and SPPM(Qy^) are equal with a probability of 1.

(iii) D(BLUPM1(Qε^))=0.

(iv) D(BLUPM1(Qy^))=D(SPPM(Qy^)).

(v) BLUPM1(Qε^)=0.

(vi) BLUPM1(Qε^)=0 holds with a probability of 1.

(vii) QVTX=0.

(viii) r[TVQ,X]=r(X).

(ix) R(Σ11Q1+Σ12Q2)R(X), where Q=[Q1,Q2].

Proof With Definition 4.1(a), BLUPM1(Qy^)=SPPM(Qy^), that is, the coefficient matrices of the two estimators are equal. Hence

[QZ,0][X,Σ11X]+U1[X,Σ11X]=[QZ,QVTX][X,Σ11X]+U2[X,Σ11X]U1[X,Σ11X]U2[X,Σ11X]=[QZ,QVTX][X,Σ11X][QZ,0][X,Σ11X](U1U2)[X,Σ11X]=[0,QVTX][X,Σ11X].

There exist U1 and U2 that satisfy the necessary and sufficient condition of (4.1)

r[[0,QVTX][X,Σ11X][X,Σ11X]]=r([X,Σ11X]).

After simplification of both sides of (4.2) through elementary transformation of (2.1) and block matrix, we can get

r[[0,QVXX][X,Σ11X][X,Σ11X]]=r[X[0,QVTX][X,Σ11X]0In[X,Σ11X]]r[X,Σ11X]=r[0[0,QVTX]In[X,Σ11X]]r[X,Σ11]=r[0[0,QVTX]In0]r[X,Σ11]=r(QVTX)+nr[X,Σ11],

and

r([X,Σ11X])=nr[X,Σ11].

Based on the above results, as well as (2.2), Lemma 2.1 and Lemma 2.2(b), we can get

r(QVTX)=0r[QVTX]=r(X)QVTX=0,

which indicates results (i), (vii), (viii), and (ix) are equivalent.

Through Definition 4.1(b),

BLUPM1(Qy^) and SPPM(Qy^) are equal with a probability of 1

[0,QVTX][X,Σ11X]Σ11=0D[BLUPM1(Qε^)]=0.

From (2.3),

r([0,QVTX][X,Σ11X]Σ11)=r[X,Σ11X]Σ11[0,QVTX]0r[X,Σ11X]=r[X0Σ110QVTX0]r[X,Σ11]=r(QVTX)+r[X,Σ11]r[X,Σ11]=r(QVTX).

Through (2.2), Lemma 2.1 and Lemma 2.2(b), we can obtain

r(QVTX)=0r[QVTX]=r(X)QVTX=0,

which indicate results (ii), (iii), (vi), (vii), (viii), and (ix) are equivalent. From Lemma 3.1(e),

BLUPM1(Qy^)=BLUEM1(QZβ)+BL|UPM1(Qε^)=SPPM(Qy^)+BLUPM1(Qε^),BLUPM1(Qy^)=SPPM(Qy^)BLUPM1(Qε^)=0,D(BLUPM1(Qy^))=D(BLUEM1(QZβ))+D(BLUPM1(Qε^))=D(SPPM(Qy^))+D(BLUPM1(Qε^)),D(BLUPM1(Qε^))=0D(BLUPM1(Qy^))=D(SPPM(Qy^)),

which indicate results (i), (iii), (iv), and (v) are equivalent. □

From Theorem 4.1, the following corollaries can be obtained.

Corollary 4.1  Assume y^ and 1m+ny^ are predictable under model M1, then the following conclusions hold.

(a) The following conclusions are equivalent:

(i) BLUPM1(y^)=SPPM(y^)(with a probability of 1).

(ii) R(TV)R(X).

(b) The following conclusions are equivalent:

(i) BLUPM1(y)=SPPM(y)(with a probability of 1).

(ii) R(Σ12)R(X).

(c) The following conclusions are equivalent:

(i) BLUPM1(y)=SPPM(y)(with a probability of 1).

(ii) Σ111n+Σ121mR(X).

5 Robustness of SPP and BLUP with respect to covariance matrix

Robustness refers to the relative stability of statistical inference regarding statistical models, namely assumptions. When there is a slight change in model conditions, statistical inference also undergoes a slight change accordingly, indicating the robustness of statistical inference regarding such slight changes. For example, assuming that the variance of the error term in the Gauss-Markov model y=Xβ+ε is Cov(ε)=σ2In, but in practical problems, it is sometimes impossible to fully satisfy this assumption. We can only judge whether Cov(ε)=σ2In is generally acceptable through analysis or test. We hope that when the difference between the true Cov(ε) and σ2In is not too large, the estimator of unknown parameters or their functions will still maintain their original optimality or even if it is not optimal, it will not become too bad, that is, this estimator is robust regarding the covariance matrix. Robustness is a property that every statistical inference should possess. This Section mainly considers the robustness of SPP(Qy^) and BLUP(Qy^) as well as their special cases regarding covariance matrix. Finally, a specific example is used to illustrate the specific application of the equivalent condition using the error uniform correlation model.

We consider the following two linear models:

Ni:y=Xβ+εi,E(εi)=0,D(ε1)=V11,D(ε2)=W11,i=1,2.

The future observation vector model of model (5.1) is

Ni:y=Xβ+εi,E(εi)=0,D(ε1)=V22,D(ε2)=W22,i=1,2.

The simultaneous model of models (5.1) and (5.2) is

N1^:y^=Zβ+ε^1,E(ε^1)=0,D(ε^1)=[V11V12V21V22]V,

N2^:y^=Zβ+ε^2,E(ε^2)=0,D(ε^2)=[W11W12W21W22]W.

Then the simple projection prediction of linearly predictable vector Qy^ under model Ni^ is

SPPN1^(Qy^)=([QZ,0][X,V11X]+U1[X,V11X])y,

SPPN2^(Qy^)=([QZ,0][X,W11X]+U2[X,W11X])y.

The best linear unbiased prediction of linearly predictable vector Qy^ under modelNi is

BLUPN1(Qy^)=([QZ,QVTX][X,V11X]+U3[X,V11X])y,

BLUPN2(Qy^)=([QZ,QWTX][X,W11X]+U4[X,W11X])y.

Assuming the true model is N2^, and we mistakenly select N1^, we consider the robustness of the simple projection prediction and best linear unbiased prediction of the linearly predictable vector Qy^ and its special cases regarding the covariance matrix under these two models.

Theorem 5.1  Assume Qy^ is linearly predictable under model, then the following result is established:

(a) SPPN1^(Qy^)=SPPN2^(Qy^);

(b) SPPN1^(Qy^) and SPPN2^(Qy^) are equal with a probability of 1;

(c) r[QZ00XW11XV11X]=r[X,W11X,V11X];

(d) r[QZ00XW11V110X000X]=r[XW11V110X000X];

(e) R([QZ,0,0])R([X,W11X,V11X]).

Proof Through Definition 4.1(a), result (a) holds if and only if the coefficient matrices of the two projection predictors are equal:

[QZ,0][X,V11X]+U1[X,V11X]=[QZ,0][X,W11X]+U2[X,W11X]U1[X,V11X]U2[X,W11X]=[QZ,0][X,W11X][QZ,0][X,V11X][U1,U2][[X,V11X][X,W11X]]=[QZ,0][X,W11X][QZ,0][X,V11X].

There exist U1 and U2 that satisfy the necessary and sufficient condition of (5.3)

r[[QZ,0][X,W11X][QZ,0][X,V11X][X,V11X][X,W11X]]=r[[X,V11X][X,W11X]].

After simplification of both sides of (5.4) through elementary transformation of (2.1) and block matrix, we can get

r[[QZ,0][X,W11X][QZ,0][X,V11X][X,V11X][X,W11X]]=r[[QZ,0][X,W11X][QZ,0][X,V11X]00In[X,V11X]0In0[X,W11X]]r[X,V11X]r[X,W11X]=r[0[QZ,0][QZ,0]In[X,V11X]0In0[X,W11X]]r[X,V11]r[X,W11]

=r[0[QZ,0][QZ,0]In[X,V11X][X,W11X]In00]r[X,V11]r[X,W11]=r[0[QZ,0][QZ,0]0[X,V11X][X,W11X]In00]r[X,V11]r[X,W11]=r[[QZ,0][QZ,0][X,V11X][X,W11X]]+nr[X,V11]r[X,W11]=r[QZ0QZ0XV11XXW11X]+nr[X,V11]r[X,W11]=r[QZ00XV11XW11X]+nr[X,V11]r[X,W11],

and

r[[X,V11X][X,W11X]]=r[In[X,V11X]0In0[X,W11X]]r[X,V11X]r[X,W11X]=r[In[X,V11X][X,W11X]In00]r[X,V11]r[X,W11]=r[0[X,V11X][X,W11X]In00]r[X,V11]r[X,W11]=r([X,V11X],[X,W11X])+nr[X,V11]r[X,W11]=r[X,V11X,W11X]+nr[X,V11]r[X,W11].

Combining the above two results, and through (2.2) and Lemma 2.2(b), we can obtain

r[QZ00XW11XV11X]=r[X,W11X,V11X]r[QZ00XW11V110X000X]=r[XW11V110X000X],

which indicates results (a), (c), (d), and (e) are equivalent.

From Definition 4.1(b),

SPPM1^(Qy^)andSPPM2^(Qy^) are equal with a probability of 1

([QZ,0][X,V11X]+U1[X,V11X][QZ,0][X,W11X]U2[X,W11X])[X,W11]=0([QZ,0][X,V11X]+U1[X,V11X][QZ,0][X,W11X])[X,W11]=0.

With Lemma 2.3, (5.5) is further equivalent to

([QZ,0][X,V11X]+U1[X,V11X][QZ,0][X,W11X])[X,W11X]=0[QZ,0][X,V11X][X,W11X]+U1[X,V11X][X,W11X]=[QZ,0]U1[X,V11X][X,W11X]=[QZ,0][QZ,0][X,V11X][X,W11X].

There exists U1 that satisfies the necessary and sufficient condition of (5.6)

r[[QZ,0][QZ,0][X,V11X][X,W11X][X,V11X][X,W11X]]=r([X,V11X][X,W11X]).

After simplification of both sides of (5.7) through elementary transformation of (2.1) and block matrix, we can get

r[[QZ,0][QZ,0][X,V11X][X,W11X][X,V11X][X,W11X]]=r[[QZ,0][QZ,0][X,V11X][X,W11X]0[X,W11X][X,V11X]]r[X,V11X]=r[[QZ,0][QZ,0][X,W11X][X,V11X]]r[X,V11]=r[QZ0QZ0XW11XXV11X]r[X,V11]=r[QZ00XW11XV11X]r[X,V11],

and

r([X,V11X][X,W11X])=r([X,V11X],[X,W11X])r[[X,V11X]=r[X,V11X,X,W11X]r[X,V11]=r[X,V11X,W11X]r[X,V11].

Combining the above results, it can be seen that (b) and (c) are equivalent. □

Theorem 5.2  Assume Qy^ is linearly predictable under model N2, then the following results are established:

(a) BLUPN1(Qy^)=BLUPN2(Qy^);

(b) BLUPN1(Qy^) and BLUPN2(Qy^) are equal with a probability of 1;

(c) r[QZQWTXQVTXXW11XV11X]=r[X,W11X,V11X];

(d) r[QZQWTQVTXW11V110X000X]=r[XW11V110X000X];

(e) R([QZ,QWTX,QVTX])R([X,W11X,V11X]).

Proof From Definition 4.1(a), result (a) holds if and only if the coefficient matrices of BLUPN1(Qy^) and BLUPN2(Qy^)) are equal, that is

[QZ,QVTX][X,V11X]+U3[X,V11X]=[QZ,QWTX][X,W11X]+U4[X,W11X]U3[X,V11X]U4[X,W11X]=[QZ,QWTX][X,W11X][QZ,QVTX][X,V11X][U3,U4][[X,V11X][X,W11X]].

There exist U3 and U4 that satisfy the necessary and sufficient condition of (5.8)

r[[QZ,QVTX][X,W11X][QZ,QVTX][X,V11X][X,V11X][X,W11X]]=r[[X,V11X][X,W11X]].

After simplification of both sides of (5.9) through elementary transformation of (2.1) and block matrix, we can get

r[[QZ,QVTX][X,W11X][QZ,QVTX][X,V11X][X,V11X][X,W11X]]=r[[QZ,QVTX][X,W11X]00[QZ,QVTX][X,V11X]00In[X,V11X]0In0[X,W11X]]r[X,V11X]r[X,W11X]=r[0[QZ,QVTX][QZ,QVTX]In[X,V11X]0In0[X,W11X]]r[X,V11]r[X,W11]=r[0[QZ,QVTX][QZ,QWTX]In[X,V11X][X,W11X]In00]r[X,V11]r[X,W11]=r[0[QZ,QVTX][QZ,QWTX]0[X,V11X][X,W11X]In00]r[X,V11]r[X,W11]

=r[[QZ,QVTX][QZ,QWTX][X,V11X][X,W11X]]+nr[X,V11]r[X,W11]=r[QZQVTXQZQWTXXV11XXW11X]+nr[X,V11]r[X,W11]=r[QZQVTXQWTXXV11XW11X]+nr[X,V11]r[X,W11],

and

r[[X,V11X][X,W11X]]=r[In[X,V11X]0In0[X,W11X]]r[X,V11X]r[X,W11X]=r[In[X,V11X][X,W11X]In00]r[X,V11]r[X,W11]=r[0[X,V11X][X,W11X]In00]r[X,V11]r[X,W11]=r([X,V11X],[X,W11X])+nr[X,V11]r[X,W11]=r[X,V11X,W11X]+nr[X,V11]r[X,W11].

Combining the above two results, and with (2.2) and Lemma 2.2(b), we can obtain

r[QZQWTXQVTXXW11XV11X]=r[X,V11X,W11X]r[QZQWTQVTXW11V110X000X]=r[XV11V110X000X],

which indicates results (a), (c), (d), and (e) are equivalent.

From Definition 4.1(b),

BLUPN1(Qy^) and BLUPN2(Qy^) are equal with a probability of 1

([QZ,QVTX][X,V11X]+U3[X,V11X][QZ,QWTX][X,W11X]U4[X,W11X])[X,W11]=0([QZ,QVTX][X,V11X]+U3[X,V11X][QZ,QWTX][X,W11X])[X,W11]=0.

With Lemma 2.3, (5.10) is further equivalent to

([QZ,QVTX][X,V11X]+U3[X,V11X][QZ,QWTX][X,W11X])[X,W11X]=0[QZ,QVTX][X,V11X][X,W11X]+U3[X,V11X][X,W11X]=[QZ,QWTX]U3[X,V11X][X,W11X]=[QZ,QWTX][QZ,QVTX][X,V11X][X,W11X].

There exists U3 that satisfies the necessary and sufficient condition of (5.11)

r[[QZ,QWTX][QZ,QVTX][X,V11X][X,W11X][X,V11X][X,W11X]]=r([X,V11X][X,W11X]).

After simplification of both sides of (5.12) through elementary transformation of (2.1) and block matrix, we can obtain

r[[QZ,QWTX][QZ,QVTX][X,V11X][X,W11X][X,V11X][X,W11X]]=r[[QZ,QWTX][QZ,QVTX][X,V11X][X,W11X]0[X,W11X][X,V11X]]r[X,V11X]=r[[QZ,QWTX][QZ,QVTX][X,W11X][X,V11X]]r[X,V11]=r[QWTXQZQVTXXW11XXV11X]r[X,V11]=r[QWTXQVTXXW11XV11X]r[X,V11],

and

r([X,V11X][X,W11X])=r([X,V11X],[X,W11X])r[X,V11X]=r[X,V11X,X,W11X]r[X,V11]=r[X,V11X,W11X]r[X,V11].

Combining the above results, it can be seen that (b) and (c) are equivalent. □

Corollary 5.1  Assume y^ and 1m+ny^ are linearly predictable under model N2, then the following conclusions are established.

(a) The following results are equivalent:

(i) SPPN1^(y^)=SPPN2^(y^) (with a probability of 1).

(ii) r[Z00XW11V110X000X]=r[XW11V110X000X].

(b) The following results are equivalent:

(i) SPPN1^(1m+ny^)=SPPN2^(1m+ny^) (with a probability of 1).

(ii) r[1m+nZ00XW11V110X000X]=r[XW11V110X000X].

Corollary 5.2  Assume y^ and 1m+ny^are linearly predictable under model N2, and let V1=[V11V21], W1=[W11W21]. Then the following conclusions hold.

(a) The following two conditions are equivalent:

(i) BLUPN1(y^)=BLUPN2(y^)(with a probability of 1).

(ii) r[ZW1V10X000X]=r[XW11V110X000X].

(b) The following two conditions are equivalent:

(i) BLUPN1(1m+ny^)=BLUPN2(1m+ny^)(with a probability of 1).

(ii) r[1m+nZ1nW11+1mW211nV11+1mV21XW11V110X000X]=r[XW11V110X000X].

Next, we use the error uniform correlation models in statistics to illustrate the specific application of equivalent conditions.

Assume

N1^:y^=1m+nβ+ε^1,E(ε^1)=0,D(ε^1)=σ2Im+n,

M2^:y^=1m+nβ+ε^2,E(ε^2)=0,D(ε^2)=σ2[1ρρρ1ρρρ1],

where 1m+n=[1,1,,1] are vectors of (m+n)×1 order with all components being 1, and Im+nare unit matrices of m+n order. The covariance matrix of model N2^ can be rewritten as

N2^:y^=1m+nβ+ε^2,E(ε^2)=0,D(ε^2)=σ2[ρ1m+n1m+n+(1ρ)Im+n].

That is, there is equal variance σ2 in all observations, and the same correlation coefficient as well. Assume Q=Im+n, and Qy^=y^ is linearly predictable under model N2,

Z=1m+n,X=1n,W11=σ2[ρ1n1n+(1ρ)In],V11=σ2In.

From the equivalent condition (ii) in Corollary 5.1(a), the left side of the equation is

r[Z00XW11V110X000X]=r[1m+n001nσ2[ρ1n1n+(1ρ)In]σ2In01n0001n]=r[1m+n001nσ2ρ1n1nσ2In01n00(ρ1)1n1n]=r[1m+n0000σ2In01n0001n]=r[1m+n0000σ2In01n0]=n+2,

and the right side of the equation is

r[XW11V110X000X]=r[1nσ2[ρ1n1n+(1ρ)In]σ2In01n0001n]=r[1nσ2ρ1n1nσ2In01n00(ρ1)1n1n]=r[1n0σ2In01n0σ2n00]=r[00σ2In01n0σ2n00]=n+2.

Hence

r[Z00XW11V110X000X]=r[XW11V110X000X].

Then, the equivalent condition (ii) in Corollary 5.1(a) is satisfied, namely the SPP is robust under this model.

6 Conclusions

SPP and BLUP are two very basic predictors of unknown parameters in linear models. SPP boasts a simple form and is very convenient to use in statistical inference; BLUP, on the other hand, is the predictor with the smallest covariance among all linear unbiased predictors of unknown parameters under Löwner partial order. It is often used as a comparison standard for the soundness of other predictors for excellent algebraic and statistical properties. The study of the relationship between these two predictors is an important work. This paper mainly uses a new method - the basic block matrix rank formula in matrix algebra theory, i.e. the matrix rank and block matrix method, to provide various necessary and sufficient conditions for the rank and spatial inclusion relationship between them. Finally, the robustness of SPP and BLUP with respect to the covariance matrix is studied separately, and specific applications of equivalent conditions are illustrated with error uniform correlation models in statistics, lending important theoretical references for subsequent statistical inference and research on predictor relationships under other complex statistical models.

References

[1]

Alalouf I S, Styan G P H. Characterizations of estimability in the general linear model. Ann Statist 1979; 7(1): 194–200

[2]

Bolfarine H, Rodrigues J. On the simple projection predictor in finite populations. Austral J Statist 1988; 30(3): 338–341

[3]

Bolfarine H, Zacks S, Elian S N, Rodrigues J. Optimal prediction of the finite population regression coefficient. Sankhya Ser B 1994; 56(1): 1–10

[4]

Drygas H. Estimation and prediction for linear models in general spaces. Math Operationsforsch Statist 1975; 6(2): 301–324

[5]

Gan S J, Sun Y Q, Tian Y G. Equivalence of predictors under real and over-parameterized linear models. Comm Statist Theory Methods 2017; 46(11): 5368–5383

[6]

Goldberger A S. Best linear unbiased prediction in the generalized linear regression models. J Amer Statist Assoc 1962; 57: 369–375

[7]

Henderson C R. Best linear unbiased estimation and prediction under a selection model. Biometrics 1975; 31(2): 423–447

[8]

Jiang J M. A derivation of BLUP—best linear unbiased predictor. Statist Probab Lett 1997; 32(3): 321–324

[9]

Liu X Q, Rong J Y, Liu X Y. Best linear unbiased prediction for linear combinations in general mixed linear models. J Multivariate Anal 2008; 99(8): 1503–1517

[10]

Lu C L, Gan S J, Tian Y G. Some remarks on general linear model with new regressors. Statist Probab Lett 2015; 97: 16–24

[11]

Lu C L, Sun Y Q, Tian Y G. A comparison between two competing fixed parameter constrained general linear models with new regressors. Statistics 2018; 52(4): 769–781

[12]

Marsaglia G, Styan G P H. Equalities and inequalities for ranks of matrices. Linear Multilinear Algebra 1974/1975; 2: 269–292

[13]

Penrose R. A generalized inverse for matrices. Proc Cambridge Philos Soc 1955; 51: 406–413

[14]

PuntanenSStyanG P HIsotaloJ. Matrix Tricks for Linear Statistical Models: Our Personal Top Twenty. Heidelberg: Springer, 2011

[15]

Rao C R. Unified theory of linear estimation. Sankhya Ser A 1971; 33: 371–394

[16]

Rao C R. Representations of best linear unbiased estimators in the Gauss-Markoff model with a singular dispersion matrix. J Multivariate Anal 1973; 3: 276–292

[17]

Rao C R. A lemma on optimization of a matrix function and a review of the unified theory of linear estimation, In: Statistical Data Analysis and Inference (Neuchatel, 1989). Amsterdam: North-Holland 1989; 397–417

[18]

Särndal C-E, Wright R L. Cosmetic form of estimators in survey sampling. Scand J Statist 1984; 11(3): 146–156

[19]

Tian Y G. A new derivation of BLUPs under random-effects model. Metrika 2015; 78(8): 905–918

[20]

Tian Y G. A matrix handling of predictions under a general linear random-effects model with new observations. Electron J Linear Algebra 2015; 29: 30–45

[21]

Tian Y G. Solutions of a constrained Hermitian matrix-valued function optimization problem with applications. Oper Matrices 2016; 10(4): 967–983

[22]

Tian Y G, Jiang B. A new analysis of the relationships between a general linear model and its mis-specified forms. J Korean Statist Soc 2017; 46(2): 182–193

[23]

Yu S H, He C Z. Comparison of general Gauss-Markov models in estimable subspace. Acta Math Appl Sinica 1997; 20(4): 580–586

[24]

Yu S H, He C Z. Optimal prediction in finite populations. Appl Math J Chinese Univ Ser A 2000; 15(2): 199–205

[25]

Yu S H, Liang X L. The simple projection predictor in finite populations with arbitrary rank. Mathematics in Economics 2001; 18(4): 49–52

RIGHTS & PERMISSIONS

Higher Education Press 2024

AI Summary AI Mindmap
PDF (565KB)

312

Accesses

0

Citation

Detail

Sections
Recommended

AI思维导图

/