1. College of Mathematics and Information Science, Shandong Technology and Business University, Yantai 264005, China
2. College of Business and Economics, Shanghai Business School, Shanghai 201400, China
yongge.tian@gmail.com
Show less
History+
Received
Accepted
Published
Issue Date
Revised Date
2024-10-15
PDF
(565KB)
Abstract
Linear model is a kind of important model in statistics. The estimation and prediction of unknown parameters in the model is a basic problem in statistical inference. According to different evaluation criteria, different forms of predictors are obtained. The simple projection prediction (SPP) boasts simple form while the best linear unbiased prediction (BLUP) enjoys very excellent properties. Naturally, the relationship between the two predictors are considered. We give the necessary and sufficient conditions for the equality of SPP and BLUP. Finally, we study the robustness of SPP and BLUP with respect to covariance matrix and illustrate the application of equivalence conditions by use of error uniform correlation models.
Bo JIANG, Luping WANG, Yongge TIAN.
Further research on simple projection prediction under a general linear model.
Front. Math. China, 2024, 19(5): 255-275 DOI:10.3868/s140-DDD-024-0016-x
Linear model is very important in statistics and have important applications in many fields, such as agriculture, medicine, economics, education, and many others. The estimation and prediction of unknown parameters in models are a fundamental problem in statistical inference. Based on different optimal criteria, statisticians can define different estimators and predictors of unknown parameters. Among these estimators and predictors, classic ones include least squares estimation, weighted least squares estimation, simple projection prediction (SPP), optimal linear unbiased prediction (BLUP), etc. Among them, simple projection prediction boasts a simple form while optimal linear unbiased prediction enjoys excellent properties. Naturally, we consider the topic of equality between them, and present the equivalent conditions for equality, to provide a theoretical basis for further statistical inference.
We consider the following classic linear model
Here, is an observable random vector of order , is a model matrix of order n×p under random rank, is an unknown parameter vector of order , is an error vector of order , is a zero mean vector, is a known non-negative definite matrix, denoted by , where E(·) and D(·) represent the expectation and variance of the random vector, respectively.
In statistical research, it is sometimes necessary to study the problem of predictor of censored data in future observations. Let the future observation vector model of model (1.1) be
Here, is the future observation vector of order m×1, is the model matrix of order under random rank, is the error vector of order m×1, is the zero mean vector, and is a known non-negative definite matrix.
The simultaneous model of models (1.1) and (1.2) is
where , , .
Due to the presence of unknown parameter vectors and error terms in linear models, people usually only focus on the estimation of unknown parameter , and there have been scarce discussions on the joint prediction of unknown parameter vectors and error terms. Here, we consider the joint prediction problem of unknown parameter vectors and error terms
Here, , and are known matrices. Since , , and can be any matrix, contains all linear combinations of unknown parameters , error terms and .
(i) If , , , then is the joint observation vector in model (1.3);
(ii) If , , , then is random linear function of the joint vector in model (1.3), where is the known matrix;
(iii) If , , , then is the sum of the observed values of the joint vector in model (1.3), i.e., , where represents vectors of order with all components being 1;
(iv) If , , , then is the future observation vector in model (1.2);
(v) If , , , then is the known observation vector in model (1.1).
Reference [17] gave an important lemma for optimizing matrix-valued functions in the sense of Löwner partial order, and studied the joint prediction problem under mixed effect model. In recent years, references [19, 20] presented the BLUP analytical expressions for the joint prediction of fixed effect, random effect, and error term under random effect model by solving the constrained quadratic matrix-valued function optimization problem in the sense of Löwner partial order. This work can be referred to in references [3-5, 7-11].
With (1.3) and (1.4), we can get
where , .
This paper uses a novel method to solve the constrained quadratic matrix-valued function optimization problem, establishes a complete theoretical system for optimal linear unbiased prediction (BLUP) with respect to , and presents an analytical expression for BLUP and related algebraic and statistical properties. The main work is divided into the following four parts:
(I) Provide model (1.1) consistency definition and definition of and predictability under model (1.1);
(II) Establish analytical expressions for , and and their algebraic and statistical properties;
(III) Present the equivalence conditions for ;
(IV) Study the robustness of covariance matrices with respect to and .
2 Preliminaries
In this paper, , and are used to indicate the transpose, rank of matrix, and linear subspace generated by column vector of matrix A, respectively. Im represents the identity matrix of order m. The Moore-Penrose inverse of matrix A, denoted by A†, is defined as the unique solution G that simultaneously satisfies the following four matrix equations , , and . PA, EA, and PA represent orthogonal projection matrices (symmetric idempotent matrices), and , and . Given two symmetric matrices A and B of the same order, if , they satisfy Löwner partial order .
Matrix theory, as a mathematical operation tool, has been widely used in linear statistical models (see monograph [14]). The rank of a matrix is a rather basic concept and an important algebraic tool in matrix algebra, which can be used to establish and simplify various equations containing generalized inverse matrix operations. In recent decades, researchers at home and abroad have established numerous analytical calculation formulas for matrix operation ranks, used these formulas to handle the construction, calculation, and comparison of a large number of estimators in regression analysis, and derived many simple and practical criteria for determining the equivalence of different estimators, laying an important theoretical basis for subsequent research on statistical inference of more complex statistical models. Some basic formulas for block matrix rank were used in some cases, such as references [12, 14].
Below are some basic formulas for calculating the rank of block matrices. They will be used in the proof of the theorem later.
Lemma 2.4 [13] The necessary and sufficient condition for the linear matrix equationto be solvable is , or . Then the general solution of the equation can be expressed as , whereUis an arbitrary matrix.
Below is the definition of model (1.1) consistency. With Moore-Penrose inverse definition, equation
always holds. Thus,
is established. These two equations indicate that holds true for a probability of 1, or holds true for a probability of 1. For the definition of model consistency, please refer to [15, 16].
Definition 2.1 If holds true for a probability of 1, then model is consistent or compatible.
Below is a classic definition of predictability (see [1, 6]).
Definition 2.2 If there exists a linear estimator Ly that satisfies
then vector is linearly predictable under model .
Theorem 2.1is linearly estimable under modelis linearly estimable under model . In particular,
(a) yis always linearly predictable under model, andis always linearly estimable under model ;
(b) andare linearly predictable under modelis linearly estimable undermodel ;
(c) is linearly predictable under modelis linearly estimable under model .
Proof Through Definition 2.2, is linearly predictable under model is linearly estimable under model .
With Lemma 2.4, the result holds. □
Many topics in statistics are closely related to optimization problems in mathematics. References [19‒21] provided an analytical solution for the optimization problem of quadratic matrix-valued functions with constraints, and used the following lemma to present an analytical expression for joint prediction of BLUP under the most general assumptions of the model, as well as algebraic and statistical properties.
Lemma 2.5Let , , , , be non-negative definitematrices. Assume matrix equationis solvable with respect to , then in the sense of Löwner partial order, the quadratic matrix-valued function with constraint
is always solvable with respect toG, and the solution G0of (2.4) satisfies the following matrixequation
Then, the analytical expression of G0, f(G0), and f(G) can be expressed as
where is a random matrix.
3 Predictable vector under model and its special cases SPP and BLUP
Prediction in linear models has always been a fundamental problem in statistics. Based on different evaluation criteria, different expression forms of predictor can be obtained. The intuitive idea is to seek a concise form of prediction. Särndal and Wright [18] emphasized that prediction should have a concise and intuitive form, and the most concise and intuitive prediction was SPP; another commonly used optimal criterion was to find the prediction with the smallest covariance among all unbiased predictors of unknown parameters in the sense of Löwner partial order, which was the famous BLUP. This paper mainly considers the topic of equality between and . The work in this regard can be seen in [2, 18, 23-25].
Firstly, we provide the analytical expressions for these two basic predictors.
Definition 3.1 If there exists a matrix that under Löwner partial order satisfies
then linear statistic is defined as BLUP of , denoted by
When is defined as BLUE of , denoted by
When F=QZ, , , is defined as BLUP of Q, denoted by
The derivation of the analytical expressions for BLUP/BLUE under model with respect to and its special cases can be found in [19].
Lemma 3.1Assume is predictable under model , that is, , then matrix equation
is solvable, and the general solution of the equation and the BLUP analytical expression ofare
where , , whileis a random matrix.
Assume Q is predictable under model , QZβ is estimable under model , that is, , then matrix equation
is solvable, so that the BLUP/BLUE analytical expressions of , , and Q are
where , , while is a random matrix. And the following results are established:
(a)
(b) The only necessary and sufficient condition of coefficient matrix L is .
(c) The only necessary and sufficient condition of holds true with a probability of 1, namely model is compatible.
(d) The following formula holds:
and
where .
(e) can be decomposed into
When the covariance matrix V is a special form of matrix, such as or , the above results about BLUP/BLUE can be further simplified. The above results are given under the most general assumptions of the model, which do not require conditions like full rank of the model matrix columns and invertibility of the covariance matrix, and there are no assumptions about the distribution of error terms. Therefore, the above lemma displays generality.
Next, we define the simple projection prediction of under model (1.3).
Definition 3.2 The SPP of vector under model (1.3) is defined as
and
4 Research on the relationship between BLUP and SPP of Q
BLUP and SPP, as two basic predictors in mathematical statistics, are based on different optimal criteria and therefore have different analytical expressions. BLUP enjoys many excellent algebraic and statistical properties, and it has the minimum variance in unbiased predictors. Statisticians often use BLUP to measure the effectiveness of other estimators. In order to study the relationship between the two linear estimators G1y and G2y, the following definition is given first.
Definition 4.1 Let y be a random vector, G1 and G2 be two matrices with appropriate order.
(a) If , then the two random vectors G1y and G2y are absolutely equal, namely .
(b) If and hold simultaneously, then the two random vectors G1y and G2y are equal with a probability of 1.
The main conclusions of this paper are presented below.
Theorem 4.1AssumeQis linearly predictable under model , then the following conclusions are equivalent:
(i)
(ii) and are equal with a probability of 1.
(iii)
(iv)
(v)
(vi) holds with a probability of 1.
(vii)
(viii)
(ix) , where .
Proof With Definition 4.1(a), , that is, the coefficient matrices of the two estimators are equal. Hence
There exist U1 and U2 that satisfy the necessary and sufficient condition of (4.1)
After simplification of both sides of (4.2) through elementary transformation of (2.1) and block matrix, we can get
and
Based on the above results, as well as (2.2), Lemma 2.1 and Lemma 2.2(b), we can get
which indicates results (i), (vii), (viii), and (ix) are equivalent.
Through Definition 4.1(b),
and are equal with a probability of 1
From (2.3),
Through (2.2), Lemma 2.1 and Lemma 2.2(b), we can obtain
which indicate results (ii), (iii), (vi), (vii), (viii), and (ix) are equivalent. From Lemma 3.1(e),
which indicate results (i), (iii), (iv), and (v) are equivalent. □
From Theorem 4.1, the following corollaries can be obtained.
Corollary 4.1Assumeandare predictable under model , then the following conclusions hold.
(a) The following conclusions are equivalent:
(i) (with a probability of 1).
(ii)
(b) The following conclusions are equivalent:
(i) (with a probability of 1).
(ii)
(c) The following conclusions are equivalent:
(i) (with a probability of 1).
(ii)
5 Robustness of SPP and BLUP with respect to covariance matrix
Robustness refers to the relative stability of statistical inference regarding statistical models, namely assumptions. When there is a slight change in model conditions, statistical inference also undergoes a slight change accordingly, indicating the robustness of statistical inference regarding such slight changes. For example, assuming that the variance of the error term in the Gauss-Markov model is , but in practical problems, it is sometimes impossible to fully satisfy this assumption. We can only judge whether is generally acceptable through analysis or test. We hope that when the difference between the true and is not too large, the estimator of unknown parameters or their functions will still maintain their original optimality or even if it is not optimal, it will not become too bad, that is, this estimator is robust regarding the covariance matrix. Robustness is a property that every statistical inference should possess. This Section mainly considers the robustness of SPP(Q) and BLUP(Q) as well as their special cases regarding covariance matrix. Finally, a specific example is used to illustrate the specific application of the equivalent condition using the error uniform correlation model.
We consider the following two linear models:
The future observation vector model of model (5.1) is
The simultaneous model of models (5.1) and (5.2) is
Then the simple projection prediction of linearly predictable vector Q under model is
The best linear unbiased prediction of linearly predictable vector Q under model is
Assuming the true model is , and we mistakenly select , we consider the robustness of the simple projection prediction and best linear unbiased prediction of the linearly predictable vector Q and its special cases regarding the covariance matrix under these two models.
Theorem 5.1AssumeQis linearly predictable under model, then the following result is established:
(a)
(b) andare equal with a probability of 1;
(c)
(d)
(e)
Proof Through Definition 4.1(a), result (a) holds if and only if the coefficient matrices of the two projection predictors are equal:
There exist U1 and U2 that satisfy the necessary and sufficient condition of (5.3)
After simplification of both sides of (5.4) through elementary transformation of (2.1) and block matrix, we can get
and
Combining the above two results, and through (2.2) and Lemma 2.2(b), we can obtain
which indicates results (a), (c), (d), and (e) are equivalent.
From Definition 4.1(b),
are equal with a probability of 1
With Lemma 2.3, (5.5) is further equivalent to
There exists U1 that satisfies the necessary and sufficient condition of (5.6)
After simplification of both sides of (5.7) through elementary transformation of (2.1) and block matrix, we can get
and
Combining the above results, it can be seen that (b) and (c) are equivalent. □
Theorem 5.2AssumeQis linearly predictable under model , then the following results are established:
(a) ;
(b) and are equal with a probability of 1;
(c) ;
(d) ;
(e)
Proof From Definition 4.1(a), result (a) holds if and only if the coefficient matrices of and ) are equal, that is
There exist U3 and U4 that satisfy the necessary and sufficient condition of (5.8)
After simplification of both sides of (5.9) through elementary transformation of (2.1) and block matrix, we can get
and
Combining the above two results, and with (2.2) and Lemma 2.2(b), we can obtain
which indicates results (a), (c), (d), and (e) are equivalent.
From Definition 4.1(b),
and are equal with a probability of 1
With Lemma 2.3, (5.10) is further equivalent to
There exists U3 that satisfies the necessary and sufficient condition of (5.11)
After simplification of both sides of (5.12) through elementary transformation of (2.1) and block matrix, we can obtain
and
Combining the above results, it can be seen that (b) and (c) are equivalent. □
Corollary 5.1Assumeandare linearly predictable under model , then the following conclusions are established.
(a) The following results are equivalent:
(i) (with a probability of 1).
(ii)
(b) The following results are equivalent:
(i) (with a probability of 1).
(ii)
Corollary 5.2Assumeandare linearly predictable under model , and let , . Then the following conclusions hold.
(a) The following two conditions are equivalent:
(i) (with a probability of 1).
(ii)
(b) The following two conditions are equivalent:
(i) (with a probability of 1).
(ii)
Next, we use the error uniform correlation models in statistics to illustrate the specific application of equivalent conditions.
Assume
where are vectors of (m+n)×1 order with all components being 1, and Im+nare unit matrices of m+n order. The covariance matrix of model can be rewritten as
That is, there is equal variance in all observations, and the same correlation coefficient as well. Assume , and is linearly predictable under model ,
From the equivalent condition (ii) in Corollary 5.1(a), the left side of the equation is
and the right side of the equation is
Hence
Then, the equivalent condition (ii) in Corollary 5.1(a) is satisfied, namely the SPP is robust under this model.
6 Conclusions
SPP and BLUP are two very basic predictors of unknown parameters in linear models. SPP boasts a simple form and is very convenient to use in statistical inference; BLUP, on the other hand, is the predictor with the smallest covariance among all linear unbiased predictors of unknown parameters under Löwner partial order. It is often used as a comparison standard for the soundness of other predictors for excellent algebraic and statistical properties. The study of the relationship between these two predictors is an important work. This paper mainly uses a new method - the basic block matrix rank formula in matrix algebra theory, i.e. the matrix rank and block matrix method, to provide various necessary and sufficient conditions for the rank and spatial inclusion relationship between them. Finally, the robustness of SPP and BLUP with respect to the covariance matrix is studied separately, and specific applications of equivalent conditions are illustrated with error uniform correlation models in statistics, lending important theoretical references for subsequent statistical inference and research on predictor relationships under other complex statistical models.
Alalouf I S, Styan G P H. Characterizations of estimability in the general linear model. Ann Statist1979; 7(1): 194–200
[2]
Bolfarine H, Rodrigues J. On the simple projection predictor in finite populations. Austral J Statist1988; 30(3): 338–341
[3]
Bolfarine H, Zacks S, Elian S N, Rodrigues J. Optimal prediction of the finite population regression coefficient. Sankhya Ser B1994; 56(1): 1–10
[4]
Drygas H. Estimation and prediction for linear models in general spaces. Math Operationsforsch Statist1975; 6(2): 301–324
[5]
Gan S J, Sun Y Q, Tian Y G. Equivalence of predictors under real and over-parameterized linear models. Comm Statist Theory Methods2017; 46(11): 5368–5383
[6]
Goldberger A S. Best linear unbiased prediction in the generalized linear regression models. J Amer Statist Assoc1962; 57: 369–375
[7]
Henderson C R. Best linear unbiased estimation and prediction under a selection model. Biometrics1975; 31(2): 423–447
[8]
Jiang J M. A derivation of BLUP—best linear unbiased predictor. Statist Probab Lett1997; 32(3): 321–324
[9]
Liu X Q, Rong J Y, Liu X Y. Best linear unbiased prediction for linear combinations in general mixed linear models. J Multivariate Anal2008; 99(8): 1503–1517
[10]
Lu C L, Gan S J, Tian Y G. Some remarks on general linear model with new regressors. Statist Probab Lett2015; 97: 16–24
[11]
Lu C L, Sun Y Q, Tian Y G. A comparison between two competing fixed parameter constrained general linear models with new regressors. Statistics2018; 52(4): 769–781
[12]
Marsaglia G, Styan G P H. Equalities and inequalities for ranks of matrices. Linear Multilinear Algebra1974/1975; 2: 269–292
[13]
Penrose R. A generalized inverse for matrices. Proc Cambridge Philos Soc1955; 51: 406–413
[14]
PuntanenSStyanG P HIsotaloJ. Matrix Tricks for Linear Statistical Models: Our Personal Top Twenty. Heidelberg: Springer, 2011
[15]
Rao C R. Unified theory of linear estimation. Sankhya Ser A1971; 33: 371–394
[16]
Rao C R. Representations of best linear unbiased estimators in the Gauss-Markoff model with a singular dispersion matrix. J Multivariate Anal1973; 3: 276–292
[17]
Rao C R. A lemma on optimization of a matrix function and a review of the unified theory of linear estimation, In: Statistical Data Analysis and Inference (Neuchatel, 1989). Amsterdam: North-Holland1989; 397–417
[18]
Särndal C-E, Wright R L. Cosmetic form of estimators in survey sampling. Scand J Statist1984; 11(3): 146–156
[19]
Tian Y G. A new derivation of BLUPs under random-effects model. Metrika2015; 78(8): 905–918
[20]
Tian Y G. A matrix handling of predictions under a general linear random-effects model with new observations. Electron J Linear Algebra2015; 29: 30–45
[21]
Tian Y G. Solutions of a constrained Hermitian matrix-valued function optimization problem with applications. Oper Matrices2016; 10(4): 967–983
[22]
Tian Y G, Jiang B. A new analysis of the relationships between a general linear model and its mis-specified forms. J Korean Statist Soc2017; 46(2): 182–193
[23]
Yu S H, He C Z. Comparison of general Gauss-Markov models in estimable subspace. Acta Math Appl Sinica1997; 20(4): 580–586
[24]
Yu S H, He C Z. Optimal prediction in finite populations. Appl Math J Chinese Univ Ser A2000; 15(2): 199–205
[25]
Yu S H, Liang X L. The simple projection predictor in finite populations with arbitrary rank. Mathematics in Economics2001; 18(4): 49–52
RIGHTS & PERMISSIONS
Higher Education Press 2024
AI Summary 中Eng×
Note: Please be aware that the following content is generated by artificial intelligence. This website is not responsible for any consequences arising from the use of this content.