ORIGINAL RESEARCH article

Front. Appl. Math. Stat., 16 April 2025

Sec. Statistics and Probability

Volume 11 - 2025 | https://doi.org/10.3389/fams.2025.1379210

Optimal estimators in biadditive models and their families

  • 1. Department of Mathematics and Center for Research on Mathematics and its Applications (CIMA), Universidade de Évora, Évora, Portugal

  • 2. Department of Mechatronics Engineering, Universidade de Évora, Évora, Portugal

  • 3. Institute of Contemporary History, Faculty of Social and Human Sciences, New University of Lisbon, Lisboa, Portugal

  • 4. Department of Mathematics and Center of Mathematics and its Applications (CMA), Faculty de Ciências e Tencnologia, Universidade Nova de Lisboa, Lisbon, Portugal

Article metrics

View details

1k

Views

319

Downloads

Abstract

Biadditive regression models are linear models with an additive structure for their covariance matrix. We introduce commutative conditions and derive optimal estimators, namely Best Linear Unbiased Estimators (BLUE) and Best Quadratic Unbiased Estimators (BQUE). We develop a simulation study to compare the variance components estimates obtained through the proposed approach with those derived from Analysis of Variance and Markov Chain Monte Carlo methods. This research highlights that commutative orthogonal structures in these models are highly convenient to strengthen inference.

1 Introduction

Linear regression serves as the fundamental starting point for regression methods and remains a valuable and widely used statistical method. In addition, it acts as a solid foundation for exploring newer approaches. Therefore, the significance of a thorough understanding of linear regression before diving into more complex statistical methods cannot be overstated. Mixed-effects models are employed to describe the relationship between a response variable and one or more covariates in grouped data, structured according to factors such as longitudinal observations, repeated measures, hierarchical organization, or block designs [1]. These models extend linear models by incorporating random effects, which introduce an additional error term to account for the correlation between observations within the same group. Mixed models demonstrate broader applicability and greater generality than fixed or random models, making them particularly suitable for analyzing complex data structures with multiple sources of variability. Mixed models can be orthogonal (e.g. [2]) or non-orthogonal (e.g. [3]). Orthogonal models occur when the fixed and random effects are independent of each other, which simplifies the estimation of parameters. Non-orthogonal mixed models arise when there is a correlation between the fixed and random effects.

Moreover, biadditive regression models, which extend the analysis of variance (ANOVA) to models with quadratic terms, have frequent applications in ecology, where the experimental units within a block or stratum are considered a random sample from a population of units and the blocks or strata themselves are viewed as a random sample drawn from a population of blocks or strata [4].

Biadditive regression models are a flexible statistical framework designed to account for both fixed and random effects. The model is given by the expression

where Y is a vector of N random variables Y1, …, YN, Xβ expresses the fixed effects, and represents the sum of w independent random terms, each associated with a specific source of variability. The covariance matrix of Y is structured additively as

where , and and Ci, i = 1, …, w, unknown and known and invertible, respectively. Random terms Z1, …, Zw were assumed to have independent and identically distributed (i.i.d.) components with null mean values and higher-order cumulants (cr)1, …, (cr)w, r = 1, 2, 3, 4. However, these conditions have been refined. Currently, Zi1, …, Ziw are treated as independent, with null mean vectors and covariance matrices . This revised assumption allows for a more flexible and realistic modeling of variability in the data, accommodating non-identical covariance structures across random effects.

Alexandre et al. [5] conducted a detailed analysis to estimate the covariance components and the coefficient vectors in the biadditive regression models. This estimation process involved modeling the variability and dependence structure of the data through the covariance terms, which represent the scale of variation for each random effect Zi. Additionally, the coefficient vectors β were estimated to quantify the contributions of fixed effects, capturing the systematic relationships between covariates and the response variable.

In this paper, we present biadditive models with two extended frameworks: the orthogonal block structure and the commutative orthogonal block structure. These models introduce a novel perspective, enabling a more detailed exploration of underlying covariance structures. We establish commutativity conditions and derive optimal estimators, focusing specifically on best linear unbiased estimators (BLUE) and best quadratic unbiased estimators (BQUE). Furthermore, we extend the concept of biadditive models to encompass families of biadditive models, offering additional possibilities and broadening the applicability of this approach.

This paper is structured as follows. In Section 2, we present biadditive regression models. In Section 3, estimation and inference procedures for two extensions of the biadditive model; OBS and COBS, with controlled heteroscedasticity, including parameter estimation, variance estimation, and unbiased estimation for the model's parameters, are presented. Section 4 presents families of biadditive regression models with OBS and with COBS and results for estimable functions and derives chi-square tests. In Section 5, we developed a simulation study to compare the estimates of variance components, obtained through the proposed approach with those derived from analysis of variance and Markov Chain Monte Carlo methods. Finally, in Section 6, we conclude the paper with some final remarks.

2 Models and inference

Let us consider a linear mixed model given by

where Z1, …, Zw, i = 1, …, w are independent random vectors with covariance matrices If the vectors Z1, …, Zw have mean vectors μ1, …, μw, we can introduce the centered vectors for i = 1, …, w and the extended coefficients vector This approach simplifies the treatment by assuming that the centered vectors have null mean vectors. Currently, going into inference for Y, given the independence of Z1, …, Zw we consider its covariance matrix

where the matrices for i = 1, …, w provide profound insights into the relationships between variables within our dataset. This analysis reveals crucial information about how each component contributes to the overall variance and covariance observed in the data.

Moreover, we consider the orthogonal basis {α1, …, } for Ω = ℝ(X), the orthogonal complement of the range space, ℝ(X) of matrix X. Currently, the vectors with components

have null mean vectors and variance given by

where ℓ = 1, …, , i = 1, …, w, thus expressing the transformation's impact on the covariance structure.

Putting we obtain

where H = [hi], ℓ = 1, …, , i = 1, …, w, and The least-square estimator (LSE) for the variance components is vector given by

where the symbol (·)+ indicates the Moore–Penrose inverse matrix [6].

Similarly, as the expected value of Y is the LSE for will be

We also obtain the estimators for (Y) given by

and the generalized least square estimator (GLSE) for

as shown in Kariya and Kurata [7].

This method facilitates the estimation of variance components in mixed models where the random-effects factors may follow various distributions, including non-normal ones. This flexibility is achieved by focusing on the structure of the covariance matrix, rather than imposing a specific distributional assumption for random effects.

3 Optimal estimators

Let us now consider two classes of models with orthogonal properties, the OBS and COBS models. Our approach is general case for these models.

3.1 Orthogonal block structure

Models with OBS are linear mixed models whose variance-covariance matrices are linear combinations of known pairwise orthogonal projection matrices (POPM) that add up to the identity matrix and were introduced by Nelder [8, 9] and continue to play an important role in the theory of randomized block designs [10, 11]. In this section, we use the commutative conditions on the matrices Mi to derive optimal estimators for individual biadditive models. We assume that the matrices for i = 1, …, w, commute. This implies the existence of an orthogonal matrix P that diagonalizes them, as discussed in Schott [12]. Therefore, we have

where is the family of matrices diagonalized by P.

is a vector space comprising symmetric matrices that commute and contains their squares, rendering it a commutative Jordan algebra (CJA) [13]. Each CJA has a unique basis known as the principal basis, which is constituted by pairwise orthogonal orthogonal projection matrices [14, 15].

Let be the principal basis of Then, we have

which leads to

where

Let Pj denote the orthogonal projection matrix in the range space , the column space of , and let pj represent the rank of Pj for j = 1, …, m. If pj < qj, then the estimator

is the best quadratic unbiased estimator (BQUE) for γj [5]. This result extends the Hsu theorem to models with OBS. The Hsu theorem [16], provides a framework for deriving optimal quadratic unbiased estimators of variance components in mixed models.

3.2 Commutative orthogonal block structure

In model (1) obtaining the best linear unbiased estimator (BLUE) for β is critical because it ensures that the fixed-effects parameters are estimated efficiently, with minimal variance among all linear unbiased estimators. This is relevant when the model incorporates both fixed and random effects, as the presence of random terms introduces additional complexity into the covariance structure of the data.

We currently assume that the matrices Mi, i = 1, …, w, and

commute, implying that the model exhibits a commutative orthogonal block structure (COBS) [5]. Consequently, the model also satisfies the conditions for orthogonal block structure (OBS). Under these conditions, there exists an orthogonal matrix that diagonalizes the matrices M1, …, Mw+1, all of which belong to . This property simplifies the estimation process by enabling efficient decomposition of the covariance structure, thus facilitating the derivation of BLUE for β while respecting the hierarchical and orthogonal nature of the block structure.

Let have the principal basis As

we have j = 1, …, m, and thus

Moreover, with

the orthogonal projection matrix onto ℝ(L) is given by

where φ(L) = {h:ℓh ≠ 0}. Therefore, Thus,

Moreover, we observe that

and thus T and /Σ(Y) commute. This point is crucial as it implies that

is BLUE [17].

4 Families of biadditive models

We consider families of models sharing the matrices X, X1, …, Xw, so

where the vectors Zi(h), i = 1, …, w, h = 1, …, d, are independent and have null mean vectors. We also have,

and the random vectors are

The components Zi1(h), …, Zici(h) are i.i.d with cumulants cri(h), i = 1, …, w, h = 1, …, d, r = 2, 3, 4, the same for all models. In addition, the matrices M = XXt, and

are the same for all models. In the homogeneous case, in which the matrices are null, we have the GLSE given by

where

and

4.1 Orthogonal block structure

The OBS families will consist of models with OBS. Moreover, due to the uniqueness of the matrix X for all models, the vectors of the orthonormal basis of ℝ(X)and the orthogonal complement of the range space of X, are the same. We concentrate on the moments and cumulants of the random variables within the mixed model, offering a comprehensive analysis of the mathematical expressions and properties that form the foundation of the methodology for estimating variance components.

Therefore, for

we have the r-th cumulants of

where

Taking B(r) = [bij(r)], r = 2, 3, as well as

the vectors Θr and cr being the same for the models in the family. For all the models, we also have the estimators

which give rise to the LSE estimators

from which we obtain

namely, we have

We estimate the covariance matrices of the models using

thus the models in the family have the same estimated covariance matrix.

4.2 Commutative orthogonal block structure

The models within these families have vector coefficient estimators

These estimators have identical estimated covariance matrices

and are BLUE [5].

Additionally, the models have the same pairs of eigenvalues and eigenvectors (ξj, νj), j = 1, …, k for We then obtain the estimators for the main estimable functions

with estimated variances, Now, for any vector v ∈ ℝk, we have

which leads to

4.3 Hypotheses test

We currently introduce tests for the equality of the parameters in the different models. As have the same variance, when comparing ηj1, …, ηjd, j = 1, …, k, we use chi-square tests to test the hypotheses

As we are in the balanced case, where ANOVA and related techniques are robust with respect to non-normality [18], these tests will have test statistics given by

where

Under the null hypothesis H0j, j = 1, …, k, the test statistics Tj roughly follow a chi-square distribution with d−1 degrees of freedom.

Furthermore, the hypothesis

can be similarly tested. As the , have the variance

the

will be, when H0(v) hold, the product by

of a chi-square with d−1 degrees of freedom.

5 Simulation study

A simulation study was conducted to assess the performance of the proposed estimation method. The R programming language was used to generate the simulation data, following the procedure outlined below. The process was repeated a total of N = 1, 000 times to ensure robust and statistically reliable results. In each iteration, random values for the model parameters were generated according to the specified distributions for the random effects and fixed effects. The corresponding observation vectors were then calculated using the model equation. For each simulated dataset, the variance components were estimated and performance metrics such as bias, standard deviation (SD), and efficiency of the estimators were calculated. This repetition allowed for a comprehensive evaluation of the accuracy and precision of the method across a variety of random configurations. Simulate the observation vectors

where X2 and β2j represent an additional design matrix and random effects term, respectively.

Random effects were generated according to the following distributions:

where is the variance component for the first random effect,

where a = j and b = 10−j.

The true variance components () for j = 1, …, 10 were estimated using LSE (Equation 2), where

  • Z = [Zl] contains the mean values of the squared observation vectors , l = 1, …, g.

  • K = (BtB)+Bt,

  • B = [bli], with bli = alMial,

  • Mi represents the linear transformation matrix for variance components.

We estimate variance components and evaluate bias, standard deviation (SD), and δ. to evaluate the performance of methods, analysis of variance (ANOVA), and Markov chain Monte Carlo (MCMC) (Table 1).

Table 1

jModelANOVAMCMC
BiasSDδBiasSDδBiasSDδ
10.932.144.180.512.454.600.532.765.020.55
21.683.527.910.444.048.700.464.569.490.48
31.112.114.630.462.445.090.482.765.560.50
41.824.008.850.454.589.740.475.1710.630.49
51.914.168.670.484.779.540.505.3710.400.52
60.571.232.720.451.412.990.471.593.270.49
71.292.475.760.432.856.330.453.236.910.47
81.843.678.000.464.228.790.484.779.590.50
91.332.866.400.453.287.040.473.697.680.48
101.182.385.260.452.735.790.473.096.310.49

Bias, SD, and δ for with model, ANOVA, and MCMC methods.

For the cases j = 1, …, 10, our estimator consistently exhibited the smallest δ. The probability of this occurring by chance, assuming no superior precision among the methods, would be . This provides strong evidence that our method demonstrates significantly greater precision.

6 Final remarks

In this paper, we consider biadditive models, often used in studies of manuring and other agronomic applications. We incorporated two extensions of these models: orthogonal block structures and commutative orthogonal block structures, which allow for a more detailed analysis of the additive structure of covariance matrices. In addition to individual models, we considered families of biadditive regression models. By incorporating commutative conditions, we derived optimal estimators, including best linear unbiased estimators and best quadratic unbiased estimators. The proposed methodology extends classical results, such as the Hsu theorem, while providing a robust framework for hypothesis testing within families of models. This framework also emphasizes the estimation of covariance components and coefficients, offering researchers valuable tools for investigating variability across models. The chi-square tests presented here establish a solid statistical foundation for evaluating variability and ensuring precision in agronomic studies. Furthermore, we highlighted the relevance of commutative orthogonal structures for factorial models, particularly those based on prime basis factorials. Such models, often used in studies of manuring and other agronomic applications, showcase the versatility of our approach. To evaluate the performance of the proposed methods, we conducted a simulation study. In this study, we simulated observation vectors based on biadditive regression models with predefined covariance structures and random effects. Using simulations conducted for our model, as well as for the ANOVA and MCMC methods, we estimated variance components and computed performance metrics, such as bias, SD, and δ. Simulation results demonstrated the enhanced precision of our estimation approach.

Statements

Data availability statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Author contributions

MO: Investigation, Methodology, Writing – original draft, Writing – review & editing. EG: Investigation, Methodology, Writing – review & editing. AA: Writing – original draft, Writing – review & editing. JP: Writing – original draft, Writing – review & editing. JM: Methodology, Supervision, Conceptualization, Validation, Investigation, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. MO was funded by projects UIDB/04674/2020 (CIMA) DOI: 10.54499/UIDB/04674/2020 and H2020-MSCA-RISE- 2020/101007950, with the title Decision Support for the Supply of Ecosystem Services under Global Change (DecisionES), funded by the Marie Curie International Staff Exchange Scheme. JP was funded by National funds through FCT – Fundação para a Ciência e a Tecnologia, I.P., under the projects UID/04209 and LA/P/0132/2020 (DOI: 10.54499/LA/P/0132/2020).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  • 1.

    PinheiroJCBatesDM. Mixed-effects Models in S and S-PLUS. New York, NY: Springer (2000). 10.1007/978-1-4419-0318-1

  • 2.

    DemidenkoE. Mixed Models: Theory and Applications With R. New York, NY: Wiley (2013).

  • 3.

    SahaiHOjedaMM. Analysis of variance for random models. In: Unbalanced data: Theory, methods, applications, and data analysis. Cambridge: Birkhäuser (2005). 10.1007/978-0-8176-8168-5

  • 4.

    DempsterAPSelwynMRPatelCMRothAJ. Statistical and computational aspects of mixed model analysis. Appl Stat. (1984) 33:20314. 10.2307/2347446

  • 5.

    AlexandreAOliveiraMGarçãoEMexiaJ. Biadditive models: commutativity and optimum estimators. Commun Stat. (2022) 53:20212033. 10.1080/03610926.2022.2117560

  • 6.

    ZmyslonyR. A characterization of best linear unbiased estimators in the general linear model. J Mathem Stat Probab Theory. (1978) 2:36578. 10.1007/978-1-4615-7397-5_27

  • 7.

    KariyaTKurataH. Generalized Least Squares. New York, NY: Wiley Series in Probability and Statistics (2004). 10.1002/0470866993

  • 8.

    NelderJA. The analysis of randomized experiments with orthogonal block structure. I Block structure and the null analysis of variance. Proc R Soc Lond Ser A Math Phys Sci. (1965) 283:14762. 10.1098/rspa.1965.0012

  • 9.

    NelderJA. The analysis of randomized experiments with orthogonal block structure. II Treatment, structure and the general analysis of variance. Proc R Soc Lond Ser A Math Phys Sci. (1965) 283:16378. 10.1098/rspa.1965.0013

  • 10.

    CalińskiTKageyamaS. Block Designs. A Randomization Approach. Volume I: analysis. Berlin: Springer (2000). 10.1007/978-1-4612-1192-1

  • 11.

    CalińskiTKageyamaS. A Randomization Approach. Volume II: design. Berlin: Springer (2003). 10.1007/978-1-4419-9246-8

  • 12.

    ShotJR. Matrix Analysis for Statistics 3rd ed. New York, NY: Wiley. (2017).

  • 13.

    JordanPVon NeumamJWingnerG. An the algebra generation of quantum Mechanics formalism. J Ann Math. (1934) 1:2564.

  • 14.

    SeelyJ. Quadratic subspaces and completeness. J Ann Math Stat. (1971) 42:71021. 10.1214/aoms/1177693420

  • 15.

    SeelyJZyskindG. Linear Spaces and summary variance unbiased estimations. J Ann Math Stat. (1971) 42:691703. 10.1214/aoms/1177693418

  • 16.

    HsuPL. On the best unbiased estimate of variance. Statist Res Mem. (1938) 2:91104.

  • 17.

    SilveySD. Statistical Inference. London: Chapman & Hall (1975).

  • 18.

    ScheffeH. The Analysis of Variance. New York, NY: Wiley (1959).

Summary

Keywords

biadditive regression models, cumulants, heteroscedasticity, optimum estimators, orthogonal block structure, commutative orthogonal block structure

Citation

Oliveira M, Garção E, Alexandre A, Paulino J and Mexia J (2025) Optimal estimators in biadditive models and their families. Front. Appl. Math. Stat. 11:1379210. doi: 10.3389/fams.2025.1379210

Received

30 January 2024

Accepted

17 March 2025

Published

16 April 2025

Volume

11 - 2025

Edited by

Luciano Antonio De Oliveria, Federal University of Grande Dourados, Brazil

Reviewed by

Zakariya Yahya Algamal, University of Mosul, Iraq

Joel Jorge Nuvunga, Joaquim Chissano University, Mozambique

Carlos Pereira, Universidade Federal de Lavras, Brazil

Updates

Copyright

*Correspondence: Manuela Oliveira

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics