- 1Department of Mathematics and Center for Research on Mathematics and its Applications (CIMA), Universidade de Évora, Évora, Portugal
- 2Department of Mechatronics Engineering, Universidade de Évora, Évora, Portugal
- 3Institute of Contemporary History, Faculty of Social and Human Sciences, New University of Lisbon, Lisboa, Portugal
- 4Department of Mathematics and Center of Mathematics and its Applications (CMA), Faculty de Ciências e Tencnologia, Universidade Nova de Lisboa, Lisbon, Portugal
Biadditive regression models are linear models with an additive structure for their covariance matrix. We introduce commutative conditions and derive optimal estimators, namely Best Linear Unbiased Estimators (BLUE) and Best Quadratic Unbiased Estimators (BQUE). We develop a simulation study to compare the variance components estimates obtained through the proposed approach with those derived from Analysis of Variance and Markov Chain Monte Carlo methods. This research highlights that commutative orthogonal structures in these models are highly convenient to strengthen inference.
1 Introduction
Linear regression serves as the fundamental starting point for regression methods and remains a valuable and widely used statistical method. In addition, it acts as a solid foundation for exploring newer approaches. Therefore, the significance of a thorough understanding of linear regression before diving into more complex statistical methods cannot be overstated. Mixed-effects models are employed to describe the relationship between a response variable and one or more covariates in grouped data, structured according to factors such as longitudinal observations, repeated measures, hierarchical organization, or block designs [1]. These models extend linear models by incorporating random effects, which introduce an additional error term to account for the correlation between observations within the same group. Mixed models demonstrate broader applicability and greater generality than fixed or random models, making them particularly suitable for analyzing complex data structures with multiple sources of variability. Mixed models can be orthogonal (e.g. [2]) or non-orthogonal (e.g. [3]). Orthogonal models occur when the fixed and random effects are independent of each other, which simplifies the estimation of parameters. Non-orthogonal mixed models arise when there is a correlation between the fixed and random effects.
Moreover, biadditive regression models, which extend the analysis of variance (ANOVA) to models with quadratic terms, have frequent applications in ecology, where the experimental units within a block or stratum are considered a random sample from a population of units and the blocks or strata themselves are viewed as a random sample drawn from a population of blocks or strata [4].
Biadditive regression models are a flexible statistical framework designed to account for both fixed and random effects. The model is given by the expression
where Y is a vector of N random variables Y1, …, YN, Xβ expresses the fixed effects, and represents the sum of w independent random terms, each associated with a specific source of variability. The covariance matrix of Y is structured additively as
where , and and Ci, i = 1, …, w, unknown and known and invertible, respectively. Random terms Z1, …, Zw were assumed to have independent and identically distributed (i.i.d.) components with null mean values and higher-order cumulants (cr)1, …, (cr)w, r = 1, 2, 3, 4. However, these conditions have been refined. Currently, Zi1, …, Ziw are treated as independent, with null mean vectors and covariance matrices . This revised assumption allows for a more flexible and realistic modeling of variability in the data, accommodating non-identical covariance structures across random effects.
Alexandre et al. [5] conducted a detailed analysis to estimate the covariance components and the coefficient vectors in the biadditive regression models. This estimation process involved modeling the variability and dependence structure of the data through the covariance terms, which represent the scale of variation for each random effect Zi. Additionally, the coefficient vectors β were estimated to quantify the contributions of fixed effects, capturing the systematic relationships between covariates and the response variable.
In this paper, we present biadditive models with two extended frameworks: the orthogonal block structure and the commutative orthogonal block structure. These models introduce a novel perspective, enabling a more detailed exploration of underlying covariance structures. We establish commutativity conditions and derive optimal estimators, focusing specifically on best linear unbiased estimators (BLUE) and best quadratic unbiased estimators (BQUE). Furthermore, we extend the concept of biadditive models to encompass families of biadditive models, offering additional possibilities and broadening the applicability of this approach.
This paper is structured as follows. In Section 2, we present biadditive regression models. In Section 3, estimation and inference procedures for two extensions of the biadditive model; OBS and COBS, with controlled heteroscedasticity, including parameter estimation, variance estimation, and unbiased estimation for the model's parameters, are presented. Section 4 presents families of biadditive regression models with OBS and with COBS and results for estimable functions and derives chi-square tests. In Section 5, we developed a simulation study to compare the estimates of variance components, obtained through the proposed approach with those derived from analysis of variance and Markov Chain Monte Carlo methods. Finally, in Section 6, we conclude the paper with some final remarks.
2 Models and inference
Let us consider a linear mixed model given by
where Z1, …, Zw, i = 1, …, w are independent random vectors with covariance matrices If the vectors Z1, …, Zw have mean vectors μ1, …, μw, we can introduce the centered vectors for i = 1, …, w and the extended coefficients vector This approach simplifies the treatment by assuming that the centered vectors have null mean vectors. Currently, going into inference for Y, given the independence of Z1, …, Zw we consider its covariance matrix
where the matrices for i = 1, …, w provide profound insights into the relationships between variables within our dataset. This analysis reveals crucial information about how each component contributes to the overall variance and covariance observed in the data.
Moreover, we consider the orthogonal basis {α1, …, } for Ω⊥ = ℝ(X)⊥, the orthogonal complement of the range space, ℝ(X) of matrix X. Currently, the vectors with components
have null mean vectors and variance given by
where ℓ = 1, …, , i = 1, …, w, thus expressing the transformation's impact on the covariance structure.
Putting we obtain
where H = [hℓi], ℓ = 1, …, , i = 1, …, w, and The least-square estimator (LSE) for the variance components is vector given by
where the symbol (·)+ indicates the Moore–Penrose inverse matrix [6].
Similarly, as the expected value of Y is the LSE for will be
We also obtain the estimators for (Y) given by
and the generalized least square estimator (GLSE) for
as shown in Kariya and Kurata [7].
This method facilitates the estimation of variance components in mixed models where the random-effects factors may follow various distributions, including non-normal ones. This flexibility is achieved by focusing on the structure of the covariance matrix, rather than imposing a specific distributional assumption for random effects.
3 Optimal estimators
Let us now consider two classes of models with orthogonal properties, the OBS and COBS models. Our approach is general case for these models.
3.1 Orthogonal block structure
Models with OBS are linear mixed models whose variance-covariance matrices are linear combinations of known pairwise orthogonal projection matrices (POPM) that add up to the identity matrix and were introduced by Nelder [8, 9] and continue to play an important role in the theory of randomized block designs [10, 11]. In this section, we use the commutative conditions on the matrices Mi to derive optimal estimators for individual biadditive models. We assume that the matrices for i = 1, …, w, commute. This implies the existence of an orthogonal matrix P that diagonalizes them, as discussed in Schott [12]. Therefore, we have
where is the family of matrices diagonalized by P.
is a vector space comprising symmetric matrices that commute and contains their squares, rendering it a commutative Jordan algebra (CJA) [13]. Each CJA has a unique basis known as the principal basis, which is constituted by pairwise orthogonal orthogonal projection matrices [14, 15].
Let be the principal basis of Then, we have
which leads to
where
Let Pj denote the orthogonal projection matrix in the range space , the column space of , and let pj represent the rank of Pj for j = 1, …, m. If pj < qj, then the estimator
is the best quadratic unbiased estimator (BQUE) for γj [5]. This result extends the Hsu theorem to models with OBS. The Hsu theorem [16], provides a framework for deriving optimal quadratic unbiased estimators of variance components in mixed models.
3.2 Commutative orthogonal block structure
In model (1) obtaining the best linear unbiased estimator (BLUE) for β is critical because it ensures that the fixed-effects parameters are estimated efficiently, with minimal variance among all linear unbiased estimators. This is relevant when the model incorporates both fixed and random effects, as the presence of random terms introduces additional complexity into the covariance structure of the data.
We currently assume that the matrices Mi, i = 1, …, w, and
commute, implying that the model exhibits a commutative orthogonal block structure (COBS) [5]. Consequently, the model also satisfies the conditions for orthogonal block structure (OBS). Under these conditions, there exists an orthogonal matrix that diagonalizes the matrices M1, …, Mw+1, all of which belong to . This property simplifies the estimation process by enabling efficient decomposition of the covariance structure, thus facilitating the derivation of BLUE for β while respecting the hierarchical and orthogonal nature of the block structure.
Let have the principal basis As
we have j = 1, …, m, and thus
Moreover, with
the orthogonal projection matrix onto ℝ(L) is given by
where φ(L) = {h:ℓh ≠ 0}. Therefore, Thus,
Moreover, we observe that
and thus T and /Σ(Y) commute. This point is crucial as it implies that
is BLUE [17].
4 Families of biadditive models
We consider families of models sharing the matrices X, X1, …, Xw, so
where the vectors Zi(h), i = 1, …, w, h = 1, …, d, are independent and have null mean vectors. We also have,
and the random vectors are
The components Zi1(h), …, Zici(h) are i.i.d with cumulants cri(h), i = 1, …, w, h = 1, …, d, r = 2, 3, 4, the same for all models. In addition, the matrices M = XXt, and
are the same for all models. In the homogeneous case, in which the matrices are null, we have the GLSE given by
where
and
4.1 Orthogonal block structure
The OBS families will consist of models with OBS. Moreover, due to the uniqueness of the matrix X for all models, the vectors of the orthonormal basis of ℝ(X)⊥and the orthogonal complement of the range space of X, are the same. We concentrate on the moments and cumulants of the random variables within the mixed model, offering a comprehensive analysis of the mathematical expressions and properties that form the foundation of the methodology for estimating variance components.
Therefore, for
we have the r-th cumulants of
where
Taking B(r) = [bij(r)], r = 2, 3, as well as
the vectors Θr and cr being the same for the models in the family. For all the models, we also have the estimators
which give rise to the LSE estimators
from which we obtain
namely, we have
We estimate the covariance matrices of the models using
thus the models in the family have the same estimated covariance matrix.
4.2 Commutative orthogonal block structure
The models within these families have vector coefficient estimators
These estimators have identical estimated covariance matrices
and are BLUE [5].
Additionally, the models have the same pairs of eigenvalues and eigenvectors (ξj, νj), j = 1, …, k for We then obtain the estimators for the main estimable functions
with estimated variances, Now, for any vector v ∈ ℝk, we have
which leads to
4.3 Hypotheses test
We currently introduce tests for the equality of the parameters in the different models. As have the same variance, when comparing ηj1, …, ηjd, j = 1, …, k, we use chi-square tests to test the hypotheses
As we are in the balanced case, where ANOVA and related techniques are robust with respect to non-normality [18], these tests will have test statistics given by
where
Under the null hypothesis H0j, j = 1, …, k, the test statistics Tj roughly follow a chi-square distribution with d−1 degrees of freedom.
Furthermore, the hypothesis
can be similarly tested. As the , have the variance
the
will be, when H0(v) hold, the product by
of a chi-square with d−1 degrees of freedom.
5 Simulation study
A simulation study was conducted to assess the performance of the proposed estimation method. The R programming language was used to generate the simulation data, following the procedure outlined below. The process was repeated a total of N = 1, 000 times to ensure robust and statistically reliable results. In each iteration, random values for the model parameters were generated according to the specified distributions for the random effects and fixed effects. The corresponding observation vectors were then calculated using the model equation. For each simulated dataset, the variance components were estimated and performance metrics such as bias, standard deviation (SD), and efficiency of the estimators were calculated. This repetition allowed for a comprehensive evaluation of the accuracy and precision of the method across a variety of random configurations. Simulate the observation vectors
where X2 and β2j represent an additional design matrix and random effects term, respectively.
Random effects were generated according to the following distributions:
where is the variance component for the first random effect,
where a = j and b = 10−j.
The true variance components () for j = 1, …, 10 were estimated using LSE (Equation 2), where
• Z = [Zl] contains the mean values of the squared observation vectors , l = 1, …, g.
• K = (BtB)+Bt,
• B = [bli], with bli = alMial,
• Mi represents the linear transformation matrix for variance components.
We estimate variance components and evaluate bias, standard deviation (SD), and δ. to evaluate the performance of methods, analysis of variance (ANOVA), and Markov chain Monte Carlo (MCMC) (Table 1).
For the cases j = 1, …, 10, our estimator consistently exhibited the smallest δ. The probability of this occurring by chance, assuming no superior precision among the methods, would be . This provides strong evidence that our method demonstrates significantly greater precision.
6 Final remarks
In this paper, we consider biadditive models, often used in studies of manuring and other agronomic applications. We incorporated two extensions of these models: orthogonal block structures and commutative orthogonal block structures, which allow for a more detailed analysis of the additive structure of covariance matrices. In addition to individual models, we considered families of biadditive regression models. By incorporating commutative conditions, we derived optimal estimators, including best linear unbiased estimators and best quadratic unbiased estimators. The proposed methodology extends classical results, such as the Hsu theorem, while providing a robust framework for hypothesis testing within families of models. This framework also emphasizes the estimation of covariance components and coefficients, offering researchers valuable tools for investigating variability across models. The chi-square tests presented here establish a solid statistical foundation for evaluating variability and ensuring precision in agronomic studies. Furthermore, we highlighted the relevance of commutative orthogonal structures for factorial models, particularly those based on prime basis factorials. Such models, often used in studies of manuring and other agronomic applications, showcase the versatility of our approach. To evaluate the performance of the proposed methods, we conducted a simulation study. In this study, we simulated observation vectors based on biadditive regression models with predefined covariance structures and random effects. Using simulations conducted for our model, as well as for the ANOVA and MCMC methods, we estimated variance components and computed performance metrics, such as bias, SD, and δ. Simulation results demonstrated the enhanced precision of our estimation approach.
Data availability statement
The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.
Author contributions
MO: Investigation, Methodology, Writing – original draft, Writing – review & editing. EG: Investigation, Methodology, Writing – review & editing. AA: Writing – original draft, Writing – review & editing. JP: Writing – original draft, Writing – review & editing. JM: Methodology, Supervision, Conceptualization, Validation, Investigation, Writing – original draft, Writing – review & editing.
Funding
The author(s) declare that financial support was received for the research and/or publication of this article. MO was funded by projects UIDB/04674/2020 (CIMA) DOI: 10.54499/UIDB/04674/2020 and H2020-MSCA-RISE- 2020/101007950, with the title Decision Support for the Supply of Ecosystem Services under Global Change (DecisionES), funded by the Marie Curie International Staff Exchange Scheme. JP was funded by National funds through FCT – Fundação para a Ciência e a Tecnologia, I.P., under the projects UID/04209 and LA/P/0132/2020 (DOI: 10.54499/LA/P/0132/2020).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Pinheiro JC, Bates DM. Mixed-effects Models in S and S-PLUS. New York, NY: Springer (2000). doi: 10.1007/978-1-4419-0318-1
3. Sahai H, Ojeda MM. Analysis of variance for random models. In: Unbalanced data: Theory, methods, applications, and data analysis. Cambridge: Birkhäuser (2005). doi: 10.1007/978-0-8176-8168-5
4. Dempster AP, Selwyn MR, Patel CM, Roth AJ. Statistical and computational aspects of mixed model analysis. Appl Stat. (1984) 33:203–14. doi: 10.2307/2347446
5. Alexandre A, Oliveira M, Garção E, Mexia J. Biadditive models: commutativity and optimum estimators. Commun Stat. (2022) 53:2021–2033. doi: 10.1080/03610926.2022.2117560
6. Zmyslony R. A characterization of best linear unbiased estimators in the general linear model. J Mathem Stat Probab Theory. (1978) 2:365–78. doi: 10.1007/978-1-4615-7397-5_27
7. Kariya T, Kurata H. Generalized Least Squares. New York, NY: Wiley Series in Probability and Statistics (2004). doi: 10.1002/0470866993
8. Nelder JA. The analysis of randomized experiments with orthogonal block structure. I Block structure and the null analysis of variance. Proc R Soc Lond Ser A Math Phys Sci. (1965) 283:147–62. doi: 10.1098/rspa.1965.0012
9. Nelder JA. The analysis of randomized experiments with orthogonal block structure. II Treatment, structure and the general analysis of variance. Proc R Soc Lond Ser A Math Phys Sci. (1965) 283:163–78. doi: 10.1098/rspa.1965.0013
10. Caliński T, Kageyama S. Block Designs. A Randomization Approach. Volume I: analysis. Berlin: Springer (2000). doi: 10.1007/978-1-4612-1192-1
11. Caliński T, Kageyama S. A Randomization Approach. Volume II: design. Berlin: Springer (2003). doi: 10.1007/978-1-4419-9246-8
13. Jordan P, Von Neumam J, Wingner G. An the algebra generation of quantum Mechanics formalism. J Ann Math. (1934) 1:25–64.
14. Seely J. Quadratic subspaces and completeness. J Ann Math Stat. (1971) 42:710–21. doi: 10.1214/aoms/1177693420
15. Seely J, Zyskind G. Linear Spaces and summary variance unbiased estimations. J Ann Math Stat. (1971) 42:691–703. doi: 10.1214/aoms/1177693418
Keywords: biadditive regression models, cumulants, heteroscedasticity, optimum estimators, orthogonal block structure, commutative orthogonal block structure
Citation: Oliveira M, Garção E, Alexandre A, Paulino J and Mexia J (2025) Optimal estimators in biadditive models and their families. Front. Appl. Math. Stat. 11:1379210. doi: 10.3389/fams.2025.1379210
Received: 30 January 2024; Accepted: 17 March 2025;
Published: 16 April 2025.
Edited by:
Luciano Antonio De Oliveria, Federal University of Grande Dourados, BrazilReviewed by:
Zakariya Yahya Algamal, University of Mosul, IraqJoel Jorge Nuvunga, Joaquim Chissano University, Mozambique
Carlos Pereira, Universidade Federal de Lavras, Brazil
Copyright © 2025 Oliveira, Garção, Alexandre, Paulino and Mexia. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Manuela Oliveira, bW1vQHVldm9yYS5wdA==