ORIGINAL RESEARCH article

Front. Appl. Math. Stat., 12 February 2025

Sec. Statistics and Probability

Volume 11 - 2025 | https://doi.org/10.3389/fams.2025.1487331

Computation and interpretation of mean absolute deviations by cumulative distribution functions

  • Department of Computer Science, Metropolitan College, Boston University, Boston, MA, United States

Article metrics

View details

2

Citations

1,9k

Views

1,3k

Downloads

Abstract

In recent years, there has been an increased interest in using the mean absolute deviation (MAD) around the mean and median (the L1 norm) as an alternative to standard deviation σ (the L2 norm). Till now, the MAD has been computed for some distributions. For other distributions, expressions for mean absolute deviations (MADs) are not available nor reported. Typically, MADs are derived using the probability density functions (PDFs). By contrast, we derive simple expressions in terms of the integrals of the cumulative distribution functions (CDFs). We show that MADs have simple geometric interpretations as areas under the appropriately folded CDF. As a result, MADs can be computed directly from CDFs by computing appropriate integrals or sums for both continuous and discrete distributions, respectively. For many distributions, these CDFs have a simpler form than PDFs. Moreover, the CDFs are often expressed in terms of special functions, and indefinite integrals and sums for these functions are well known. We compute MADs for many well-known continuous and discrete distributions. For some of these distributions, the expressions for MADs have not been reported. We hope this study will be useful for researchers and practitioners interested in MADs.

1 Introduction

Consider a real-valued random variable X on a sample space Ω ⊆ R with density f(x) and cumulative distribution function F(x) [1]. If X is a discrete random variable, then Ω is some countable sample space and f(x) is the probability mass function (discrete density function). Let μ denote the mean E(X) and M denote the median of X. For any real-value a, define the MAD of X from a as Bloomfield and Steiger [2]

This is defined in Lebesque–Stieltjes integration [3] and applies to continuous and discrete distributions. It is well known [2, 46] that for any distribution with finite variance, we have H(X, M) ≤ H(X, μ) ≤ σ. On the other hand, MADs require only the existence of the first-order moment.

MADs have been used for a long time, as early as the eighteenth century by Boscovitch and Laplace [7, 8]. For a historical survey, see Gorard [9] and Pham-Gia and Hung [10]. There is currently a renewed interest in using the L1 norm for robust statistical modeling and inference (e.g. [1117], to name just a few). A MAD-based alternative to kurtosis based on H(X, M) was recently introduced in Pinsky and Klawansky [18].

Since the pioneering work of Fisher [19], standard deviation has been used as the primary metric for deviation. As noted in Gorard [9] and Pham-Gia and Hung [10], the convenience of differentiation and optimization, and the summation of variances for independent variables have contributed to the widespread use of the standard deviation in estimation and hypothesis testing.

On the other hand, using MAD offers a direct measure of deviation and is more resilient to outliers. In computing standard deviation, we square the differences between values and the central point such as mean or median. As a result, standard deviation emphasizes larger deviations. The computation of MAD offers a direct and easily interpretable measure of deviation.

As discussed in Gorard [9], difficulties associated with the absolute value operation are one of the main reasons for the lower use of the MAD. We address this by deriving the closed-form expressions for MADs directly from the cumulative distribution function of the underlying distribution. Moreover, the computation of MAD requires only the existence of the first-order moment, whereas the use of standard deviation requires the existence of the second moment as well. As an example, consider Pareto distribution with scale α [20] analyzed in Section 5. This distribution has a finite mean for α > 1 but the infinite variance for α < 2. As a result, we cannot use standard deviation to analyze cases 1 < α < 2 but can use MADs.

This study will primarily focus on computing the MAD about the mean. Such deviations have been derived for some distributions, primarily using density [21]. The most complete list (without derivations) is presented in Weisstein [22]. We significantly extend this list using a unified approach to computing these deviations as integrals and sums from cumulative distribution functions (CDFs) instead of probability density functions (PDFs). For many distributions, these CDFs have a simpler form than PDFs. Moreover, the CDFs are often expressed in terms of special functions, and indefinite integrals and sums for these functions are well known. When closed-form expressions for the median are available, we will compute MAD (about) median as well.

To our knowledge, the obtained results for some distributions, such as Pareto Type II, and logarithmic distributions, are presented for the first time. Our main contribution is a unified and simple computational approach to compute these deviations in closed form without absolute value operation from the cumulative distribution functions. This removes one of the main obstacles in using MADs [9]. We hope this article will be a useful reference for researchers and practitioners using MAD.

This article is organized as follows. In Section 2, we establish the expressions for H(X, μ) in terms of the integral of the CDF for the case of continuous distributions and in terms of sums of the CDFs for discrete distributions. In Section 3, we provide a simple geometric interpretation of our results in terms of appropriately folded CDF. In Section 4, we review definitions and some integrals for the special functions often appearing in CDFs for many distributions considered. In Section 5, we derive expressions for MAD for many continuous distributions. In Section 6, we derive expressions for many widely used discrete distributions.

2 Computing MADs by CDF

2.1 MAD formula for continuous distributions

We first start with continuous distributions. Let us define the left sample sub-space of Ω by ΩL(a) = {x ∈ Ω | xa} and the right sample sub-space of Ω by ΩR(a) = {x ∈ Ω | x > a}. Additionally, we will require the existence of the first-order moment,

To put it formally, let us consider the functional [23]

where a ∈ (−t, t) is finite and both integrals exist for any t. Then the first moment is defined as

It is well known that the integral in Equation 3 could be infinite when the support of X is unbounded. We will insist that the above integrals in the right-hand side of the above Equation 2 are finite and therefore μ < ∞. This can be shown to be equivalent:

and therefore, for distributions with possible negative values, we have

This condition will ensure that all terms in subsequent equations are finite.

We will find it convenient to introduce the following auxiliary integral:

This is a partial first moment of X computed over all xz. We can express H(X, a) in terms of the cumulative distribution function F(·) and auxiliary integral I(·) as follows:

Consider the function h(x) = xF(x). Then dh(x) = F(x) dx + x dF(x) and from the integration by parts formula, we can express I(a) as

Substituting this into Equation 8 we obtain

A geometric interpretation of H(X, a) for the case a ≤ μ is illustrated in Figure 1.

Figure 1

The MAD is twice the sum of two areas: the first area is the area under the cumulative distribution function for xa and the second area is (μ − a) · 0.5.

For MAD around the mean μ and the median M, we obtain

2.2 MAD formula for discrete distributions

We now derive an analogous formula for the discrete distribution. Let Ω = {x1, x2, …}. Without loss of generality, assume x1 < x2 < ⋯ . By analogy with the continuous case, define the auxiliary sum I(x) as

Then MAD of X from μ is

To compute I(a), we use the following summation by parts formula (analogous to integration by parts) [24]. If U = {u1, u2, …, un} and V = {v1, …, vn} are two sequences and Δ is the forward difference operator, then

To apply the formula, we let m = 1, vk = xk and uk = F(xk). Then we have

If we choose for a given a, then F(a) = F(xn(a)) and we have

in complete analogy for I(a) for the continuous case in Equation 9.

Substituting this expression for I(a) from Equation 16 into Equation 13, we obtain

The sum on the right of Equation 17 is the area under the cumulative distribution function for xia and can be interpreted as the integral of the CDF computed for xa in direct analogy with the continuous case.

For MAD about mean μ and median M, the above equation gives us

The advantage of these formulas is that for many distributions, xi are consecutive integers, giving us Δxi = 1. In such cases, the above formulas for H(X, μ) and H(X, M) simplify to

with a direct analogy to the continuous case.

We will apply the above results to compute MADs of several well-known discrete distributions in Section 6.

3 Interpretation of MADs via “folded” cumulative distribution functions

We can interpret H(x, a) as follows. Consider the function G(x) defined as follows:

We can interpret G(x) as the "folded" cumulative distribution function curve. Note that unless F(a) = 1/2 (i.e., a is the median M), the curve G(x) is discontinuous at x = a. The case a = μ is illustrated in Figure 2.

Figure 2

For non-negative valued random variables X, we can interpret MAD H(X, a) in terms of G(x) as follows: let A(G) denote the area under this curve. For continuous case, we have

Comparing Equation 21 with Equation 8, we can interpret AL = aF(a) − I(a) as the area under left part of G(x, a) for xa and AR = (μ − I(a) − a(1 − F(a)) as the area under the right part of G(x, a) for x > a.

For the case a = μ, the areas under the curve for xa and x > a are equal. In this case, H(X, μ) = 2AL(G) is just twice the area under the "left" part of the folded CDF curve for xa. This is illustrated in Figure 2. From Equation 9, the auxiliary integral I(μ) is the area above cumulative distribution function F(x) for x ≤ μ and F(x) ≤ F(μ) (shown in yellow color).

For the case a = M, the folded cumulative distribution function G(x) is continuous as shown in Figure 3. The area under this curve is the MAD from the median [25]. It is interesting to compare the MAD H(X, M) around the median with the quartile deviation (QD) defined as half of the interquartile range semi-interquartile range QD = IQR/2 = (Q3Q1)/2 where Q1 and Q3 are the first and third quartile of X. The MAD around the median is the area under the folded CDF G(x) where the quartile deviation is the area spanned between the quartiles on the x-axis and between 0 and 1/2 on the y-axis.

Figure 3

For the discrete case with non-negative valued X, we get a similar result. As before we assume that 0 ≤ x1 < x2 < ⋯  and let . Recall that F(xk) = f(x1) + ⋯ + f(xk). Therefore, we obtain

in exact analogy with the continuous case. Therefore, the MAD H(X, a) can also be interpreted as the area under the folded CDF curve G(x, a).

4 Some special functions and integrals

For many distributions, their cumulative distribution functions involve some special functions (for real arguments) such as Gamma and Beta functions [26, 27]. Indefinite integrals for many of these functions are readily available [2830]. Some of these functions and their integrals are presented here for easy reference.

4.1 Some special functions and integrals

  • Error function ([30], Section 7.2 in Olver et al. [31]):

  • Density of the standard normal [31, 32]:

  • Cumulative distribution of the standard normal [31, 32]:

  • Lower and upper incomplete Gamma functions ([30], Section 8.6 in Olver et al. [31]):

  • Gamma function and regularized Gamma function ([30], Section 5.2 in Olver et al. [31]):

  • Beta function (Section 5.12 in Olver et al. [31] and Gupta and Nadarajah [33]):

  • Incomplete Beta function and regularized incomplete Beta functions ([33], Section 8.17 in Olver et al. [31]):

4.2 Some indefinite integrals

The following expressions will be used in many of the computations (C is a constant).

  • Integration of standard normal CDF (p. 396 in Owen [32]):

  • Integration of the product of standard normal density and x (integral #11, p. 393 in Owen [32]):

  • Integration of the product of error function and exponential (integral #1, Section 4.2 in Ng and Geller [30]):

  • Integration of incomplete gamma function (using integration by parts and Equation 24):

    In particular, for b = 1 using the recurrence relation (chapter 8.6 in Olver et al. [31]),

    we obtain

  • Integration of the incomplete beta function [33]

  • Integration of reciprocal functions with exponentials (integral 2.313 p. 92 in Gradshteyn and Ryzhik [29])

5 MAD computations for continuous distributions

We now illustrate the above formula for some well-known continuous distributions. We will focus on MAD around the mean μ. We will consider the following continuous distributions (in alphabetical order): Beta, chi-squared, exponential, Gamma, Gumbel type I, half-normal, Laplace, logistic, Maxwell, normal, Pareto type I and II, Rayleigh, Student-t, triangular, uniform, and Weibull. For some of these distributions, the expressions for MAD have been known. In contrast, to our knowledge, the expressions for MAD have not been published for some distributions like Pareto II and Weibull.

  • Beta distribution with parameters α, β > 0: The corresponding density f(x) and the cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34]:

    where Ix(·) denotes the regularized incomplete beta function in Equation 27. The mean and standard deviation of this distribution are

    Therefore, using Equation 33 we compute H(X, μ) as follows:

    There is no closed-form formula for the median M. The MAD around the median M is

  • Exponential distribution with rate λ > 0 : The corresponding density f(x) and cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34]:

    Its mean μ = 1/λ and standard deviation σ = 1/λ. We compute H(X, μ) as

    In terms of σ, for this distribution H(X, μ) = (2/e)σ ≈ 0.74σ.

    The median is M = (log2)/λ. The MAD about median is then

  • F (Fisher–Snedecor) distribution with α, β > 0 degrees of freedom: The corresponding density f(x) and cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34],

    The mean μ and standard deviation σ are defined only for β > 2 and β > 5 respectively and are given by

    To compute H(X, μ) we note that x = βz/(α(1 − z)) and therefore, dz = (β/α(1 − z)2)dz. Let μ* = αμ/(αμ + β) = α/(α + β − 2β). Therefore, for the MAD, we obtain

    To compute the integral, we note that B′(z; α/2, β/2) = zα/2 − 1(1 − z)β/2 − 1. If we consider the function h(z) = B(z; α/2, β/2)/(1 − z) then

    This immediately gives us the expression for the integral

    We can now compute the MAD for the F distribution:

  • Gamma distribution with shape parameter α > 0 and rate β > 0: The corresponding density f(x) and cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34]:

    where Γ(·) denotes the Gamma function (Equation 25 and ν(·) denotes the lower incomplete Gamma function (Equation 24). The mean of this distribution is μ = α/β and its standard deviation .

    Let us define z = βx. Then we have

    There is no closed-form expression for the median M. The MAD (around median) is

  • χ2 Distribution with k degrees of freedom: The corresponding density f(x) and the cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34]:

    where Γ(·) denotes the Gamma function (Equation 25 and ν(·) denotes the lower incomplete Gamma function (Equation 24. The mean of this distribution is μ = k and its standard deviation . This is a special case of Gamma distribution with shape α = k/2 and rate β = 1/2. This immediately gives us

  • Gumbel type I (extreme value distribution) with location μ and scale β > 0: The corresponding density and f(x) and cumulative distribution function F(x) are given by [20]):

    Its mean is μ* = (μ + βγ) where γ is the Euler-Mascheroni constant. Its standard deviation . Let t = ez. Then dt = (t/β)dx or dx = βt−1dt. Then for MAD (about mean) we obtain

    where ν(·) denotes the incomplete lower gamma function in Equation 24.

    The median for this distribution is M = μ − β log(log2). Therefore, for the MAD about the median we have

  • Half-normal distribution with scale parameter σ > 0: The corresponding density f(x) and cumulative distribution function F(x are given by Johnson and Kotz [20]):

    The mean is and standard deviation . Let . Then and we have

    The median is . Define c = erf−1(1/2). Then the MAD around the median M we have

  • Laplace distribution with location μ and scale b: The density f(x) and cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34]:

    For this distribution, μ = M = b and . For the MAD, we obtain

    In terms of σ, we have .

  • Logistic distribution with location μ and scale s > 0: The corresponding density function f(x) and cumulative distribution function F(x are given by Johnson and Kotz [20]:

    The mean and the median are μ and the standard deviation is . Let z = (x − μ)/s. Since for this distribution M = μ, we have H(X, μ) = H(X, M). Using Equation 34 we obtain

    In terms of σ we have

  • Log-normal distribution with parameters μ ∈ (∞, +∞) and σ2 (σ > 0): The density f(x) and cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34]:

    We will use the notation μ* and σ* to denote the mean and standard deviation of X to distinguish it from the mean μ and standard deviation σ of the underlying normal distribution. The mean of the log-normal distribution is μ* = eμ+σ2/2 and its standard deviation .

    To compute H(X, μ), define z = (log x − μ)/σ. Then, dx = σx dz = σeμ + σzdz and using Equation 30 with and b = σ, we obtain

    The median M = eμ and therefore for the MAD around median, we obtain

  • Maxwell (or Maxwell-Boltzmann) distribution with scale parameter b > 0: The density function f(x) and the cumulative distribution function F(x) are given by Johnson and Kotz [20]

    We can re-write these as

    The mean is and standard deviation . We let z = x/b. Then dx = b dz and using Equations 28, 29, with a = 1/b, we obtain

    There is no simple closed-form for the median M.

  • Normal distribution N(μ, σ): The density f(x) and cumulative distribution function F(x) are [20, 34]:

    where Φ(·) denotes the cumulative distribution of the standard normal (Equation 23). Using Equation 28 with a = −μ/σ and b = 1/σ we obtain

    In terms of σ, we have . Since this is a symmetric distribution H(X, M) = H(X, μ).

  • Pareto type I distribution with shape α > 0 and scale β > 0: The density f(x) and the cumulative distribution function F(x) are given by Johnson and Kotz [20]

    the mean and standard deviation for this distribution are

    This distribution has infinite mean μ for α ≤ 1 and infinite variance for α ≤ 2. We assume that α > 1. Then for MAD, we obtain

    The median is . Therefore, for the MAD about the median, we obtain

  • Generalized Pareto (Type II) distribution with location μ ∈ R, scale σ > 0, and shape ξ ∈ R: We exclude the case ξ = μ = 0, that corresponds to the exponential distribution. The density f(x) and the cumulative distribution function F(x) are given by Johnson and Kotz [20]:

    where the support of X is x ≥ μ for ξ ≥ 0 and μ ≤ x ≤ σ/ξ for ξ < 0. The mean μ* and standard deviation σ* of this distribution are defined only for ξ < 1 and ξ < 1/2 respectively and are

    To compute H(X, μ*), we need to consider two cases

    • (a) ξ < 1 and ξ ≠ 0. We have

      Note that the case ξ = −1 gives H(X, μ*) = σ/4 and corresponds to continuous uniform distribution U(0, σ). The case ξ > 0 and location μ = σ/ξ corresponds to the Pareto distribution with scale β = σ/ξ and shape α = 1/ξ.

    • (b) Case ξ = 0. For this case, we obtain

      Note that for ξ = 0, the case μ = 0 corresponds to an exponential distribution with parameter 1/σ.

    For μ = 0 and ξ > 0, let α = 1/ξ and λ = ξσ. Then we obtain the Lomax distribution with shape α > 0 and scale λ > 0. Its mean μ* = λ/(α − 1) is defined only for α > 1 (corresponding to ξ < 1). Therefore, for α > 1 its MAD for the Lomax distribution is

  • Rayleigh distribution with scale parameter σ > 0: The density function f(x) and the cumulative distribution function F(x) are given by Johnson and Kotz [20]:

    The mean is and standard deviation is . Let z = x/2σ. Then dx = 2σ dz and we have

    For this distribution, the median . Therefore,

  • Student-t distribution with k > 1 degrees of freedom: Its probability density function f(x) and its cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34]:

    The mean μ is μ = 0 for k > 1, otherwise it is undefined. The standard deviation is defined only for k > 2 and σ = k/(k − 2). From Equation 8 we have

    Since this distribution is symmetric, M = μ and H(X, M) = H(X, μ).

  • Triangular distribution with lower limit a, upper limit b and mode c ∈ [0, 1]: The density function f(x) and the cumulative distribution function F(x) is given by [20, 34, 35]:

    For this distribution, μ = (a + b + c)/3 and . It is easy to show that for 0 ≤ c ≤ (a + b)/2, we have μ ≥ c and for (a + b)/2 ≤ cb, we have μ ≤ c. Therefore,

    The median M of this distribution is

    Therefore, the MAD about the median is then

  • Uniform continuous distribution in [a, b]: The density function f(x) and the cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34]:

    Mean is μ = (a + b)/2 and standard deviation . The MAD is

  • Weibull distribution with scale α > 0 and shape k > 0: The density f(x) and the cumulative distribution function F(x) are given by Johnson and Kotz [20].

    Its mean is μ = αΓ(1 + 1/k) and its standard deviation is σ = α(Γ(1 + 2/k) − Γ2(1 + 1/k)), If we let z = x/α then dx = α dz and using Equation 24 we obtain

    For k = 1 we have Γ(1 + 1/k) = 1, μ = 2α, and Γ(1, 1) = 1/e. Therefore, H(X, μ) = 2α/e. This case with k = 1 corresponds to exponential distribution with λ = 1/α.

    The median . Therefore for MAD about median, we obtain

6 MAD computations for discrete distributions

In this section, we compute MADs for the following discrete distributions (in alphabetical order): binomial, geometric, logarithmic, negative binomial, Poisson, uniform, Zipf, and zeta distributions. For some of these distributions, the expressions for MAD have been known, whereas for some distributions like logarithmic, the formulae for MAD have not been published, to our knowledge.

  • Binomial distribution with parameters nN and success probability 0 < p < 1: Let q = 1 − p. Its density f(x) and umulative distribution function F(x) are [20, 34]:

    where Iq(·) is the regularized incomplete beta function defined in equation 27.

    The mean μ = Np and standard deviation is . Let n = ⌊Np⌋. Then we have

    The last step in the above equation can be established by induction on N.

  • Geometric distribution with success probability 0 < p < 1: The density f(x) and cumulative distribution function are [20, 34]:

    The mean μ = (1−p)/p and standard deviation is . Let n = ⌊μ⌋ = ⌊1/p⌋ − 1. Therefore, we obtain

  • Logarithmic distribution with parameter 0 < p < 1: The density f(x) and cumulative distribution function are [20]:

    where B(·) is the incomplete beta function. Its mean μ = −p/((1 − p)log(1 − p)).

    Let n = ⌊μ⌋. Therefore, for the MAD, we obtain

    To evaluate the integral in the above equation, we note that log′(1 − t) = −1/(1 − t) and (1/(1−t))′ = −1/(1 − t)2. Therefore, for this integral, we have

    Substituting this into the above equation for H(X, μ), we obtain

  • Negative binomial distribution with number of successes parameter r > 0 and success probability 0 < p < 1: The density f(x) and cumulative distribution function are [20, 34]:

    where Ip(·) denotes the regularized incomplete beta function.

    The mean is μ = r(1 − p)/p and standard deviation is . Let n = ⌊μ⌋. Then from Equation 26 we have . On the other hand, we have

    From this, we obtain

    Using Equation 26 for B·), we obtain for the MAD

  • Poisson distribution with rate λ > 0: The density f(x) and cumulative distribution function F(x) are [20, 34]:

    where Q(·) is the regularized Gamma function in Equation 25.

    The mean is μ = λ and standard deviation is . Let n = ⌊λ⌋. For this distribution, we have

    On the other hand, from Equation 24, we have F′(n) = Q′(n + 1, λ) = −e − λλn/n!. Therefore, I(n) = λF(n) − e − λλn+1/n! and we obtain for the MAD

  • Uniform (discrete) distribution on interval [1, N]: The density f(x) and cumulative distribution function F(x) are [20, 34]:

    Its mean μ = N/2 and its standard deviation is . Let n = ⌊μ⌋. Then, for the MAD, we obtain

    There are two cases to consider:

    • (a) N is odd. Then N = (2m + 1) for some m and μ = (m + 1). Then n = (m + 1) and H(X, μ) = (m + 1)(m + 2)/N = (N − 1)(N + 1)/4N.

    • (b) N is even. Then N = 2m for some m and μ = (m + 1/2). Then n = m and H(X, μ) = m(m + 1)/N = N/4

    We summarize both cases as follows:

    For odd N we can express MAD in terms of σ as H(X, μ) = 12σ2/N

  • Zipf and zeta distributions We start with generalized Zipf distribution [20, 34]. Assume X is distributed according to zeta distribution with parameters s > 1 and integer N. Its density f(x) and cumulative distribution functions are

    where hk, s is the generalized harmonic number defined as . The mean and standard deviation of this distribution are

    Let n = ⌊μ⌋. Then, using the following equation for the sum of harmonic numbers

    we obtain the MAD of the Zipf's distribution with N elements

    For s > 1, we can extend to N = ∞. In this case, we have the Zeta distribution, with its density f(x) and cumulative distribution function given by

    where ζ(s) is the Riemann zeta function . The mean μ is defined for s > 2 and standard deviation σ is defined for s > 3 and are

    Let n = ⌊ζ(s − 1)/ζ(s)⌋. Then, using the above equation for the sum of harmonic numbers we obtain for the MAD of Zeta distribution:

7 Concluding remarks

This article presented a unified approach to compute MADs from cumulative distribution functions. The computation involves computing integrals of these functions often expressed in closed form in terms of special functions such as beta, gamma, or error functions. For many distributions, the integrals of these functions are well known and the resulting expressions for MADs can be easily obtained. For some of these distributions, the obtained results have not been reported elsewhere. We hope that this study will be useful for researchers and practitioners interested in using mean absolute deviations.

Statements

Data availability statement

The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding author/s.

Author contributions

EP: Writing – original draft, Writing – review & editing.

Funding

The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.

Acknowledgments

The author would like to thank Metropolitan College of Boston University for their support.

Conflict of interest

The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fams.2025.1487331/full#supplementary-material

References

  • 1.

    PapoulisA. Probability, Random variable, and Stochastic Processes, 2nd Edn. New York, NY: McGraw-Hill (1984).

  • 2.

    BloomfieldPSteigerWL. Least Absolute Deviations: Theory, Applications and Algorithms. Boston, MA: Birkhauser. (1983). 10.1007/978-1-4684-8574-5

  • 3.

    RudinW. Principles of Mathematical Analysis. New York, NY: McGraw-Hill (1976).

  • 4.

    ElsayedEA. On uses of mean absolute deviation: shape exploring and distribution function estimation. arXiv. (2022) [Preprint]. 10.48550/arXiv.2206.09196

  • 5.

    SchwertmanNCGilksAJCameronJ. A simple noncalculus proof that the median minimizes the sum of the absolute deviations. Am Stat. (1990) 44:3841. 10.2307/2684955

  • 6.

    ShadSM. On the minimum property of the first absolute moment. Am Stat. (1969) 23:27. 10.2307/2682577

  • 7.

    FarebrotherRW. The historical development of the L_1 and L_∞ estimation methods. In:DodgeY, editor. Statistical data Analysis Based on the L_1-norm and Related Topics. Amsterdam: North-Holland (1987), p. 3763.

  • 8.

    PortnoySKoenkerR. The Gaussian hare and the Laplacian tortoise: computability of square-error versus abolute-error estimators. Stat Sci. (1997) 2:279300. 10.1214/ss/1030037960

  • 9.

    GorardS. Revisiting a 90-year-old debate: the advantages of the mean deviation. Br J Educ Stud. (2005) 53:41730. 10.1111/j.1467-8527.2005.00304.x

  • 10.

    Pham-GiaTHungTL. The mean and median absolute deviations. J Math Comput Modell. (2001) 34:92136. 10.1016/S0895-7177(01)00109-1

  • 11.

    DodgeY. Statistical Data Analysis Based on the L_1 Norm and Related Topics. Amsterdam: North-Holland (1987).

  • 12.

    ElsayedKMT. Mean absolute deviation: analysis and applications. Int J Bus Stat Anal. (2015) 2:6374. 10.12785/ijbsa/020201

  • 13.

    GorardS. An absolute deviation approach to assessing correlation. Br J Educ Soc Behav Sci. (2015) 53:7381. 10.9734/BJESBS/2015/11381

  • 14.

    GorardS. Introducing the mean deviation “effect" size. Int J Res Method Educ. (2015) 38:10514. 10.1080/1743727X.2014.920810

  • 15.

    HabibE. Mean absolute deviation about median as a tool of exploratory data analysis. Int J Res Rev Appl Sci. (2012) 11:51723.

  • 16.

    LeysCKleinOBernardPLicataL. Detecting outliers: do not use standard deviation around the mean, use absolute deviation around the median. J Exp Soc Psychol. (2013) 49:7646. 10.1016/j.jesp.2013.03.013

  • 17.

    YagerRRAlajilanN. A note on mean absolute deviation. Inf Sci. (2014) 279:63241. 10.1016/j.ins.2014.04.016

  • 18.

    PinskyEKlawanskyS. MAD (about median) vs. quantile-based alternatives for classical standard deviation, skew, and kurtosis. Front Appl Math Stat. (2023) 9:1206537. 10.3389/fams.2023.1206537

  • 19.

    FisherRA. A mathematical examination of the methods of determining the accuracy of observation by the mean error, and by the mean square error. Mon Not R Astron Soc. (1920) 80:75870. 10.1093/mnras/80.8.758

  • 20.

    JohnsonNLKotzS. Distributions in Statistics. New York, NY: J Wiley (1970).

  • 21.

    KenneyJFKeepingES. Mathematics of Statistics. Princeton, NJ: Van Nostrand (1962).

  • 22.

    WeissteinE. Mean Deviation (Wolfram Mathworld). (2011). Available at: https://mathworld.wolfram.com/MeanDeviation.htm (accessed February 1, 2025).

  • 23.

    TrapaniL. Testing for (in)finite moments. J Econom. (2016) 191:5768. 10.1016/j.jeconom.2015.08.006

  • 24.

    ChuW. Abel's lemma on summation by parts and basic hypergeometric series. Adv Appl Math. (2007) 39:490514. 10.1016/j.aam.2007.02.001

  • 25.

    XueJHTitteringtonDM. The p-folded cumulative distribution function and the mean absolute deviation from the p-quantile. Stat Probab Lett. (2011) 81:117982. 10.1016/j.spl.2011.03.014

  • 26.

    AndrewsGEAskeyRRoyR. Special Functions. Cambridge, MA: Cambridge University Press. (2000). 10.1017/CBO9781107325937

  • 27.

    OlverFW. Introduction to Asymptotics and Special Functions, 1st Edn. New York, NY: Academic Press. (1974). 10.1016/B978-0-12-525856-2.50005-X

  • 28.

    ConwayJT. Indefinite integrals of products of special functions. Integer Transforms Spec Funct. (2017) 28:16680. 10.1080/10652469.2016.1259619

  • 29.

    GradshteynISRyzhikIM. Table of Integrals, Series, and products. New York, NY: Academic Press. (1980).

  • 30.

    NgEGellerM. A table of integrals of the error function. J Res Natl Bureau Stand B Math Sc. (1969) 738:120. 10.6028/jres.073B.001

  • 31.

    OlverFWLozierDWBoisvertRFClarkCW. NIST Handbook of Mathematical Functions. New York, NY: Cambridge University Press (2010).

  • 32.

    OwenDB. A table of normal integrals. Commun Stat Simul Comput. (1980) 9:389419. 10.1080/03610918008812164

  • 33.

    GuptaAKNadarajahS. Handbook of Beta Distributions and its Applications. New York, NY: CRC Press, Taylor & Francis (2018).

  • 34.

    FellerJJ. Probability Theory and Applications. New York, NY: J Wiley (1956).

  • 35.

    WangYPinskyE. Geometry of deviation measures for triangular distributions. Front Appl Math Stat. (2023) 9:1274787. 10.3389/fams.2023.1274787

Summary

Keywords

mean absolute deviations, probability distributions, cumulative distribution functions, central absolute moments, folded CDFs

Citation

Pinsky E (2025) Computation and interpretation of mean absolute deviations by cumulative distribution functions. Front. Appl. Math. Stat. 11:1487331. doi: 10.3389/fams.2025.1487331

Received

27 August 2024

Accepted

20 January 2025

Published

12 February 2025

Volume

11 - 2025

Edited by

George Michailidis, University of Florida, United States

Reviewed by

Levent Özbek, Ankara University, Türkiye

Maria Luz Gmiz, University of Granada, Spain

Updates

Copyright

*Correspondence: Eugene Pinsky

†ORCID: Eugene Pinsky orcid.org/0000-0002-3836-1851

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics