ORIGINAL RESEARCH article

Front. Appl. Math. Stat., 16 April 2025

Sec. Statistics and Probability

Volume 11 - 2025 | https://doi.org/10.3389/fams.2025.1379210

Optimal estimators in biadditive models and their families


Manuela Oliveira
Manuela Oliveira1*Eugnio GaroEugénio Garção2Armando AlexandreArmando Alexandre1Joana PaulinoJoana Paulino3Joo MexiaJoão Mexia4
  • 1Department of Mathematics and Center for Research on Mathematics and its Applications (CIMA), Universidade de Évora, Évora, Portugal
  • 2Department of Mechatronics Engineering, Universidade de Évora, Évora, Portugal
  • 3Institute of Contemporary History, Faculty of Social and Human Sciences, New University of Lisbon, Lisboa, Portugal
  • 4Department of Mathematics and Center of Mathematics and its Applications (CMA), Faculty de Ciências e Tencnologia, Universidade Nova de Lisboa, Lisbon, Portugal

Biadditive regression models are linear models with an additive structure for their covariance matrix. We introduce commutative conditions and derive optimal estimators, namely Best Linear Unbiased Estimators (BLUE) and Best Quadratic Unbiased Estimators (BQUE). We develop a simulation study to compare the variance components estimates obtained through the proposed approach with those derived from Analysis of Variance and Markov Chain Monte Carlo methods. This research highlights that commutative orthogonal structures in these models are highly convenient to strengthen inference.

1 Introduction

Linear regression serves as the fundamental starting point for regression methods and remains a valuable and widely used statistical method. In addition, it acts as a solid foundation for exploring newer approaches. Therefore, the significance of a thorough understanding of linear regression before diving into more complex statistical methods cannot be overstated. Mixed-effects models are employed to describe the relationship between a response variable and one or more covariates in grouped data, structured according to factors such as longitudinal observations, repeated measures, hierarchical organization, or block designs [1]. These models extend linear models by incorporating random effects, which introduce an additional error term to account for the correlation between observations within the same group. Mixed models demonstrate broader applicability and greater generality than fixed or random models, making them particularly suitable for analyzing complex data structures with multiple sources of variability. Mixed models can be orthogonal (e.g. [2]) or non-orthogonal (e.g. [3]). Orthogonal models occur when the fixed and random effects are independent of each other, which simplifies the estimation of parameters. Non-orthogonal mixed models arise when there is a correlation between the fixed and random effects.

Moreover, biadditive regression models, which extend the analysis of variance (ANOVA) to models with quadratic terms, have frequent applications in ecology, where the experimental units within a block or stratum are considered a random sample from a population of units and the blocks or strata themselves are viewed as a random sample drawn from a population of blocks or strata [4].

Biadditive regression models are a flexible statistical framework designed to account for both fixed and random effects. The model is given by the expression

Y=Xβ+i=1wXiZi,

where Y is a vector of N random variables Y1, …, YN, Xβ expresses the fixed effects, and i=1wXiZi represents the sum of w independent random terms, each associated with a specific source of variability. The covariance matrix of Y is structured additively as

Σ(Y)=i=1wσi2Mi,

where Mi=XiCiXit, and σi2 and Ci, i = 1, …, w, unknown and known and invertible, respectively. Random terms Z1, …, Zw were assumed to have independent and identically distributed (i.i.d.) components with null mean values and higher-order cumulants (cr)1, …, (cr)w, r = 1, 2, 3, 4. However, these conditions have been refined. Currently, Zi1, …, Ziw are treated as independent, with null mean vectors and covariance matrices σ2C1,,σ2Cw. This revised assumption allows for a more flexible and realistic modeling of variability in the data, accommodating non-identical covariance structures across random effects.

Alexandre et al. [5] conducted a detailed analysis to estimate the covariance components σ12,,σw2 and the coefficient vectors in the biadditive regression models. This estimation process involved modeling the variability and dependence structure of the data through the covariance terms, which represent the scale of variation for each random effect Zi. Additionally, the coefficient vectors β were estimated to quantify the contributions of fixed effects, capturing the systematic relationships between covariates and the response variable.

In this paper, we present biadditive models with two extended frameworks: the orthogonal block structure and the commutative orthogonal block structure. These models introduce a novel perspective, enabling a more detailed exploration of underlying covariance structures. We establish commutativity conditions and derive optimal estimators, focusing specifically on best linear unbiased estimators (BLUE) and best quadratic unbiased estimators (BQUE). Furthermore, we extend the concept of biadditive models to encompass families of biadditive models, offering additional possibilities and broadening the applicability of this approach.

This paper is structured as follows. In Section 2, we present biadditive regression models. In Section 3, estimation and inference procedures for two extensions of the biadditive model; OBS and COBS, with controlled heteroscedasticity, including parameter estimation, variance estimation, and unbiased estimation for the model's parameters, are presented. Section 4 presents families of biadditive regression models with OBS and with COBS and results for estimable functions and derives chi-square tests. In Section 5, we developed a simulation study to compare the estimates of variance components, obtained through the proposed approach with those derived from analysis of variance and Markov Chain Monte Carlo methods. Finally, in Section 6, we conclude the paper with some final remarks.

2 Models and inference

Let us consider a linear mixed model given by

Y=Xβ+i=1wXiZi,    (1)

where Z1, …, Zw, i = 1, …, w are independent random vectors with covariance matrices σ2C1,,σ2Cw. If the vectors Z1, …, Zw have mean vectors μ1, …, μw, we can introduce the centered vectors Z.i=Zi-μi for i = 1, …, w and the extended coefficients vector β.=[βt,μ1t,,μwt]t. This approach simplifies the treatment by assuming that the centered vectors Z.i1,,Z.iw have null mean vectors. Currently, going into inference for Y, given the independence of Z1, …, Zw we consider its covariance matrix

Σ(Y)=i=1wσi2Mi,

where the matrices Mi=XiCiXit for i = 1, …, w provide profound insights into the relationships between variables within our dataset. This analysis reveals crucial information about how each component contributes to the overall variance and covariance observed in the data.

Moreover, we consider the orthogonal basis {α1, …, αn.} for Ω = ℝ(X), the orthogonal complement of the range space, ℝ(X) of matrix X. Currently, the vectors with components

Y.=αtY, =1,,n.,

have null mean vectors and variance given by

σl2(Y.)=i=1whiσi2,

where hi=αitMiα, ℓ = 1, …, n., i = 1, …, w, thus expressing the transformation's impact on the covariance structure.

Putting Y.[2]=(Y.12,,Y.n.2), we obtain

E(Y.[2])=Hσ2,

where H = [hi], ℓ = 1, …, n., i = 1, …, w, and σ2=(σ12,,σw2)t. The least-square estimator (LSE) for the variance components is vector σ~2, given by

σ~2=(HtH)+HtY.[2],    (2)

where the symbol (·)+ indicates the Moore–Penrose inverse matrix [6].

Similarly, as the expected value of Y is Xβ., the LSE for β. will be

β.~=(XtX)+XtY.

We also obtain the estimators for Σ(Y) given by

Σ~(Y)=i=1wσ~i2Mi,

and the generalized least square estimator (GLSE) for β.~,

β.~~=(XtΣ~(Y)+X)+Σ~(Y)XtΣ~(Y)+Y,

as shown in Kariya and Kurata [7].

This method facilitates the estimation of variance components in mixed models where the random-effects factors may follow various distributions, including non-normal ones. This flexibility is achieved by focusing on the structure of the covariance matrix, rather than imposing a specific distributional assumption for random effects.

3 Optimal estimators

Let us now consider two classes of models with orthogonal properties, the OBS and COBS models. Our approach is general case for these models.

3.1 Orthogonal block structure

Models with OBS are linear mixed models whose variance-covariance matrices are linear combinations of known pairwise orthogonal projection matrices (POPM) that add up to the identity matrix and were introduced by Nelder [8, 9] and continue to play an important role in the theory of randomized block designs [10, 11]. In this section, we use the commutative conditions on the matrices Mi to derive optimal estimators for individual biadditive models. We assume that the matrices Mi=XiCiXit, for i = 1, …, w, commute. This implies the existence of an orthogonal matrix P that diagonalizes them, as discussed in Schott [12]. Therefore, we have

{M1,,Mw}A(P),

where A(P) is the family of matrices diagonalized by P. A(P)

is a vector space comprising symmetric matrices that commute and contains their squares, rendering it a commutative Jordan algebra (CJA) [13]. Each CJA has a unique basis known as the principal basis, which is constituted by pairwise orthogonal orthogonal projection matrices [14, 15].

Let {Q1,,Qm̄} be the principal basis of A(P). Then, we have

Mi=j=1mbijQj,i=1,,w,

which leads to

Σ(Y)=i=1wσi2(j=1mbijQj)=j=1mγjQj,

where

γj=i=1wbijσi2,j=1,,m.

Let Pj denote the orthogonal projection matrix in the range space (Ajt), the column space of Ajt, and let pj represent the rank of Pj for j = 1, …, m. If pj < qj, then the estimator

γ~j=Yj.t(Iqj-Pj)Y.jqj-pj,j=1,,m,

is the best quadratic unbiased estimator (BQUE) for γj [5]. This result extends the Hsu theorem to models with OBS. The Hsu theorem [16], provides a framework for deriving optimal quadratic unbiased estimators of variance components in mixed models.

3.2 Commutative orthogonal block structure

In model (1) obtaining the best linear unbiased estimator (BLUE) for β is critical because it ensures that the fixed-effects parameters are estimated efficiently, with minimal variance among all linear unbiased estimators. This is relevant when the model incorporates both fixed and random effects, as the presence of random terms i=1wXiZi introduces additional complexity into the covariance structure of the data.

We currently assume that the matrices Mi, i = 1, …, w, and

Mw+1=XXt,

commute, implying that the model exhibits a commutative orthogonal block structure (COBS) [5]. Consequently, the model also satisfies the conditions for orthogonal block structure (OBS). Under these conditions, there exists an orthogonal matrix P̄ that diagonalizes the matrices M1, …, Mw+1, all of which belong to A(P̄). This property simplifies the estimation process by enabling efficient decomposition of the covariance structure, thus facilitating the derivation of BLUE for β while respecting the hierarchical and orthogonal nature of the block structure.

Let A(P̄) have the principal basis {Q̄1,,Q̄m̄}. As

A(P)A(P̄),

we have QjA(P̄), j = 1, …, m, and thus

Qj=h=1m̄cjlQ̄h,j=1,,m.

Moreover, with

L=h=1m̄hQ̄h,

the orthogonal projection matrix onto ℝ(L) is given by

P(L)=hφ(L)Q̄h,

where φ(L) = {h:ℓh ≠ 0}. Therefore, P(L)A(P̄). Thus,

T=P(Mw+1)A(P̄).

Moreover, we observe that

Σ(Y)=i=1wσi2MiA(P̄),

and thus T and /Σ(Y) commute. This point is crucial as it implies that

β~=(XtX)+XtY,

is BLUE [17].

4 Families of biadditive models

We consider families of models sharing the matrices X, X1, …, Xw, so

Y(h)=Xβ(h)+i=1wXiZi(h), h=1,,d,

where the vectors Zi(h), i = 1, …, w, h = 1, …, d, are independent and have null mean vectors. We also have,

β(h)=(β1(h),,βk(h)), h=1,,d,

and the random vectors are

Zi(h)=(Zi1(h),,Zici(h)), i=1,,w, h=1,,d.

The components Zi1(h), …, Zici(h) are i.i.d with cumulants cri(h), i = 1, …, w, h = 1, …, d, r = 2, 3, 4, the same for all models. In addition, the matrices M = XXt, and

Mi=XiXit, i=1,,w,

are the same for all models. In the homogeneous case, in which the matrices Mi=XiXit, i=1,,w, are null, we have the GLSE given by

β~(h)=(XtΣ~(Y)+X)+XtΣ~(Y)+XY(h), h=1,,d,

where

Σ~(Y)=1di=1dΣ~(Y(h))

and

Y=1dh=1dY(h).

4.1 Orthogonal block structure

The OBS families will consist of models with OBS. Moreover, due to the uniqueness of the matrix X for all models, the vectors {α1,,αn˙} of the orthonormal basis of ℝ(X)and the orthogonal complement of the range space of X, are the same. We concentrate on the moments and cumulants of the random variables within the mixed model, offering a comprehensive analysis of the mathematical expressions and properties that form the foundation of the methodology for estimating variance components.

Therefore, for

Y.i(h)=αitY(h), i=1,,n., h=1,,d,

we have the r-th cumulants of Y.i(h)

Θri=Θr(Y.i(h))=j=1wbij(r)crj, i=1,,n. h=1,,d,

where bij(r)=h=1wαijhr, i=1,,n..

Taking B(r) = [bij(r)], r = 2, 3, as well as

Θr=B(r)cr, r=2,3,

the vectors Θr and cr being the same for the models in the family. For all the models, we also have the estimators

Θr(h)=(Y˙1r(h),,Y˙n˙r(h)), r=2,3, h=1,,d,

which give rise to the LSE estimators

c~r(h)=(B(r)tB(r))+B(r)tΘ~r(h), r=2,3,

from which we obtain

c~r=1dh=1dcr(h), r=2,3,

namely, we have

c~2=1dh=1dσh2.

We estimate the covariance matrices of the models using

Σ~(Y(h))=j=1wσ~j2Mj, h=1,,d,

thus the models in the family have the same estimated covariance matrix.

4.2 Commutative orthogonal block structure

The models within these families have vector coefficient estimators

β~(h)=(XtX)+XtY(h), h=1,,d.

These estimators have identical estimated covariance matrices

Σ~(β(h))=(XtX)+XtΣ~(Y)X(XtX), h=1,,d,

and are BLUE [5].

Additionally, the models have the same pairs of eigenvalues and eigenvectors (ξj, νj), j = 1, …, k for Σ~(β(h)). We then obtain the estimators for the main estimable functions

η~jh=νjtβ~(h), j=1,,k, h=1,,d,

with estimated variances, vjtΣ~(β(h))vj, j=1,,k. Now, for any vector v ∈ ℝk, we have

v=j=1k(vtνj)νj,

which leads to

vtβ~(h)=j=1k(vtνj)η~jh, h=1,,d.

4.3 Hypotheses test

We currently introduce tests for the equality of the parameters in the different models. As η~j1,,η~jd have the same variance, when comparing ηj1, …, ηjd, j = 1, …, k, we use chi-square tests to test the hypotheses

H0j:ηj1==ηjd, j=1,,k.

As we are in the balanced case, where ANOVA and related techniques are robust with respect to non-normality [18], these tests will have test statistics given by

Tj=1ξjh=1d(η~jh-η~j)2,j=1,,k,

where

η~j=1dh=1dη~jh.

Under the null hypothesis H0j, j = 1, …, k, the test statistics Tj roughly follow a chi-square distribution with d−1 degrees of freedom.

Furthermore, the hypothesis

H0(v):vtβ1==vtβd

can be similarly tested. As the vtβl,l=1,,d, have the variance

σ2(v)=vt(Σ(β(h)))v

the

V=j=1d(vtβ~j-vtβ~.)2

will be, when H0(v) hold, the product by

σ~2(v)=vt(Σ~(β(h)))v

of a chi-square with d−1 degrees of freedom.

5 Simulation study

A simulation study was conducted to assess the performance of the proposed estimation method. The R programming language was used to generate the simulation data, following the procedure outlined below. The process was repeated a total of N = 1, 000 times to ensure robust and statistically reliable results. In each iteration, random values for the model parameters were generated according to the specified distributions for the random effects and fixed effects. The corresponding observation vectors were then calculated using the model equation. For each simulated dataset, the variance components were estimated and performance metrics such as bias, standard deviation (SD), and efficiency of the estimators were calculated. This repetition allowed for a comprehensive evaluation of the accuracy and precision of the method across a variety of random configurations. Simulate the observation vectors

Yj=X0β0+X1β1j+X2β2j,j=1,,10,

where X2 and β2j represent an additional design matrix and random effects term, respectively.

Random effects were generated according to the following distributions:

β1j~N(03,σ1j2I3),

where σ1j2 is the variance component for the first random effect,

β2j~Gamma(a,b)-10-jb,

where a = j and b = 10−j.

The true variance components (σj2) for j = 1, …, 10 were estimated using LSE (Equation 2), where

Z = [Zl] contains the mean values of the squared observation vectors Y.l=alY, l = 1, …, g.

K = (BtB)+Bt,

B = [bli], with bli = alMial,

Mi represents the linear transformation matrix for variance components.

We estimate variance components and evaluate bias, standard deviation (SD), and δ. to evaluate the performance of methods, analysis of variance (ANOVA), and Markov chain Monte Carlo (MCMC) (Table 1).

Table 1
www.frontiersin.org

Table 1. Bias, SD, and δ for σj2 with model, ANOVA, and MCMC methods.

For the cases j = 1, …, 10, our estimator consistently exhibited the smallest δ. The probability of this occurring by chance, assuming no superior precision among the methods, would be q=(13)10. This provides strong evidence that our method demonstrates significantly greater precision.

6 Final remarks

In this paper, we consider biadditive models, often used in studies of manuring and other agronomic applications. We incorporated two extensions of these models: orthogonal block structures and commutative orthogonal block structures, which allow for a more detailed analysis of the additive structure of covariance matrices. In addition to individual models, we considered families of biadditive regression models. By incorporating commutative conditions, we derived optimal estimators, including best linear unbiased estimators and best quadratic unbiased estimators. The proposed methodology extends classical results, such as the Hsu theorem, while providing a robust framework for hypothesis testing within families of models. This framework also emphasizes the estimation of covariance components and coefficients, offering researchers valuable tools for investigating variability across models. The chi-square tests presented here establish a solid statistical foundation for evaluating variability and ensuring precision in agronomic studies. Furthermore, we highlighted the relevance of commutative orthogonal structures for factorial models, particularly those based on prime basis factorials. Such models, often used in studies of manuring and other agronomic applications, showcase the versatility of our approach. To evaluate the performance of the proposed methods, we conducted a simulation study. In this study, we simulated observation vectors based on biadditive regression models with predefined covariance structures and random effects. Using simulations conducted for our model, as well as for the ANOVA and MCMC methods, we estimated variance components and computed performance metrics, such as bias, SD, and δ. Simulation results demonstrated the enhanced precision of our estimation approach.

Data availability statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Author contributions

MO: Investigation, Methodology, Writing – original draft, Writing – review & editing. EG: Investigation, Methodology, Writing – review & editing. AA: Writing – original draft, Writing – review & editing. JP: Writing – original draft, Writing – review & editing. JM: Methodology, Supervision, Conceptualization, Validation, Investigation, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. MO was funded by projects UIDB/04674/2020 (CIMA) DOI: 10.54499/UIDB/04674/2020 and H2020-MSCA-RISE- 2020/101007950, with the title Decision Support for the Supply of Ecosystem Services under Global Change (DecisionES), funded by the Marie Curie International Staff Exchange Scheme. JP was funded by National funds through FCT – Fundação para a Ciência e a Tecnologia, I.P., under the projects UID/04209 and LA/P/0132/2020 (DOI: 10.54499/LA/P/0132/2020).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Pinheiro JC, Bates DM. Mixed-effects Models in S and S-PLUS. New York, NY: Springer (2000). doi: 10.1007/978-1-4419-0318-1

Crossref Full Text | Google Scholar

2. Demidenko E. Mixed Models: Theory and Applications With R. New York, NY: Wiley (2013).

Google Scholar

3. Sahai H, Ojeda MM. Analysis of variance for random models. In: Unbalanced data: Theory, methods, applications, and data analysis. Cambridge: Birkhäuser (2005). doi: 10.1007/978-0-8176-8168-5

Crossref Full Text | Google Scholar

4. Dempster AP, Selwyn MR, Patel CM, Roth AJ. Statistical and computational aspects of mixed model analysis. Appl Stat. (1984) 33:203–14. doi: 10.2307/2347446

Crossref Full Text | Google Scholar

5. Alexandre A, Oliveira M, Garção E, Mexia J. Biadditive models: commutativity and optimum estimators. Commun Stat. (2022) 53:2021–2033. doi: 10.1080/03610926.2022.2117560

Crossref Full Text | Google Scholar

6. Zmyslony R. A characterization of best linear unbiased estimators in the general linear model. J Mathem Stat Probab Theory. (1978) 2:365–78. doi: 10.1007/978-1-4615-7397-5_27

Crossref Full Text | Google Scholar

7. Kariya T, Kurata H. Generalized Least Squares. New York, NY: Wiley Series in Probability and Statistics (2004). doi: 10.1002/0470866993

Crossref Full Text | Google Scholar

8. Nelder JA. The analysis of randomized experiments with orthogonal block structure. I Block structure and the null analysis of variance. Proc R Soc Lond Ser A Math Phys Sci. (1965) 283:147–62. doi: 10.1098/rspa.1965.0012

Crossref Full Text | Google Scholar

9. Nelder JA. The analysis of randomized experiments with orthogonal block structure. II Treatment, structure and the general analysis of variance. Proc R Soc Lond Ser A Math Phys Sci. (1965) 283:163–78. doi: 10.1098/rspa.1965.0013

Crossref Full Text | Google Scholar

10. Caliński T, Kageyama S. Block Designs. A Randomization Approach. Volume I: analysis. Berlin: Springer (2000). doi: 10.1007/978-1-4612-1192-1

Crossref Full Text | Google Scholar

11. Caliński T, Kageyama S. A Randomization Approach. Volume II: design. Berlin: Springer (2003). doi: 10.1007/978-1-4419-9246-8

Crossref Full Text | Google Scholar

12. Shot JR. Matrix Analysis for Statistics 3rd ed. New York, NY: Wiley. (2017).

Google Scholar

13. Jordan P, Von Neumam J, Wingner G. An the algebra generation of quantum Mechanics formalism. J Ann Math. (1934) 1:25–64.

Google Scholar

14. Seely J. Quadratic subspaces and completeness. J Ann Math Stat. (1971) 42:710–21. doi: 10.1214/aoms/1177693420

Crossref Full Text | Google Scholar

15. Seely J, Zyskind G. Linear Spaces and summary variance unbiased estimations. J Ann Math Stat. (1971) 42:691–703. doi: 10.1214/aoms/1177693418

Crossref Full Text | Google Scholar

16. Hsu PL. On the best unbiased estimate of variance. Statist Res Mem. (1938) 2:91–104.

Google Scholar

17. Silvey SD. Statistical Inference. London: Chapman & Hall (1975).

Google Scholar

18. Scheffe H. The Analysis of Variance. New York, NY: Wiley (1959).

Google Scholar

Keywords: biadditive regression models, cumulants, heteroscedasticity, optimum estimators, orthogonal block structure, commutative orthogonal block structure

Citation: Oliveira M, Garção E, Alexandre A, Paulino J and Mexia J (2025) Optimal estimators in biadditive models and their families. Front. Appl. Math. Stat. 11:1379210. doi: 10.3389/fams.2025.1379210

Received: 30 January 2024; Accepted: 17 March 2025;
Published: 16 April 2025.

Edited by:

Luciano Antonio De Oliveria, Federal University of Grande Dourados, Brazil

Reviewed by:

Zakariya Yahya Algamal, University of Mosul, Iraq
Joel Jorge Nuvunga, Joaquim Chissano University, Mozambique
Carlos Pereira, Universidade Federal de Lavras, Brazil

Copyright © 2025 Oliveira, Garção, Alexandre, Paulino and Mexia. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Manuela Oliveira, bW1vQHVldm9yYS5wdA==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.