Skip to main content

ORIGINAL RESEARCH article

Front. Appl. Math. Stat. , 12 February 2025

Sec. Statistics and Probability

Volume 11 - 2025 | https://doi.org/10.3389/fams.2025.1487331

Computation and interpretation of mean absolute deviations by cumulative distribution functions

  • Department of Computer Science, Metropolitan College, Boston University, Boston, MA, United States

In recent years, there has been an increased interest in using the mean absolute deviation (MAD) around the mean and median (the L1 norm) as an alternative to standard deviation σ (the L2 norm). Till now, the MAD has been computed for some distributions. For other distributions, expressions for mean absolute deviations (MADs) are not available nor reported. Typically, MADs are derived using the probability density functions (PDFs). By contrast, we derive simple expressions in terms of the integrals of the cumulative distribution functions (CDFs). We show that MADs have simple geometric interpretations as areas under the appropriately folded CDF. As a result, MADs can be computed directly from CDFs by computing appropriate integrals or sums for both continuous and discrete distributions, respectively. For many distributions, these CDFs have a simpler form than PDFs. Moreover, the CDFs are often expressed in terms of special functions, and indefinite integrals and sums for these functions are well known. We compute MADs for many well-known continuous and discrete distributions. For some of these distributions, the expressions for MADs have not been reported. We hope this study will be useful for researchers and practitioners interested in MADs.

1 Introduction

Consider a real-valued random variable X on a sample space Ω ⊆ R with density f(x) and cumulative distribution function F(x) [1]. If X is a discrete random variable, then Ω is some countable sample space and f(x) is the probability mass function (discrete density function). Let μ denote the mean E(X) and M denote the median of X. For any real-value a, define the MAD of X from a as Bloomfield and Steiger [2]

H(X,a)=Ω|x-a|dF(x)    (1)

This is defined in Lebesque–Stieltjes integration [3] and applies to continuous and discrete distributions. It is well known [2, 46] that for any distribution with finite variance, we have H(X, M) ≤ H(X, μ) ≤ σ. On the other hand, MADs require only the existence of the first-order moment.

MADs have been used for a long time, as early as the eighteenth century by Boscovitch and Laplace [7, 8]. For a historical survey, see Gorard [9] and Pham-Gia and Hung [10]. There is currently a renewed interest in using the L1 norm for robust statistical modeling and inference (e.g. [1117], to name just a few). A MAD-based alternative to kurtosis based on H(X, M) was recently introduced in Pinsky and Klawansky [18].

Since the pioneering work of Fisher [19], standard deviation has been used as the primary metric for deviation. As noted in Gorard [9] and Pham-Gia and Hung [10], the convenience of differentiation and optimization, and the summation of variances for independent variables have contributed to the widespread use of the standard deviation in estimation and hypothesis testing.

On the other hand, using MAD offers a direct measure of deviation and is more resilient to outliers. In computing standard deviation, we square the differences between values and the central point such as mean or median. As a result, standard deviation emphasizes larger deviations. The computation of MAD offers a direct and easily interpretable measure of deviation.

As discussed in Gorard [9], difficulties associated with the absolute value operation are one of the main reasons for the lower use of the MAD. We address this by deriving the closed-form expressions for MADs directly from the cumulative distribution function of the underlying distribution. Moreover, the computation of MAD requires only the existence of the first-order moment, whereas the use of standard deviation requires the existence of the second moment as well. As an example, consider Pareto distribution with scale α [20] analyzed in Section 5. This distribution has a finite mean for α > 1 but the infinite variance for α < 2. As a result, we cannot use standard deviation to analyze cases 1 < α < 2 but can use MADs.

This study will primarily focus on computing the MAD about the mean. Such deviations have been derived for some distributions, primarily using density [21]. The most complete list (without derivations) is presented in Weisstein [22]. We significantly extend this list using a unified approach to computing these deviations as integrals and sums from cumulative distribution functions (CDFs) instead of probability density functions (PDFs). For many distributions, these CDFs have a simpler form than PDFs. Moreover, the CDFs are often expressed in terms of special functions, and indefinite integrals and sums for these functions are well known. When closed-form expressions for the median are available, we will compute MAD (about) median as well.

To our knowledge, the obtained results for some distributions, such as Pareto Type II, and logarithmic distributions, are presented for the first time. Our main contribution is a unified and simple computational approach to compute these deviations in closed form without absolute value operation from the cumulative distribution functions. This removes one of the main obstacles in using MADs [9]. We hope this article will be a useful reference for researchers and practitioners using MAD.

This article is organized as follows. In Section 2, we establish the expressions for H(X, μ) in terms of the integral of the CDF for the case of continuous distributions and in terms of sums of the CDFs for discrete distributions. In Section 3, we provide a simple geometric interpretation of our results in terms of appropriately folded CDF. In Section 4, we review definitions and some integrals for the special functions often appearing in CDFs for many distributions considered. In Section 5, we derive expressions for MAD for many continuous distributions. In Section 6, we derive expressions for many widely used discrete distributions.

2 Computing MADs by CDF

2.1 MAD formula for continuous distributions

We first start with continuous distributions. Let us define the left sample sub-space of Ω by ΩL(a) = {x ∈ Ω | xa} and the right sample sub-space of Ω by ΩR(a) = {x ∈ Ω | x > a}. Additionally, we will require the existence of the first-order moment,

μ=xΩx dF(x)=xΩLx dF(x)+xΩRx dF(x)<    (2)

To put it formally, let us consider the functional [23]

L(X,t)=-ta|x|dF(x)+at|x|dF(x)    (3)

where a ∈ (−t, t) is finite and both integrals exist for any t. Then the first moment is defined as

μ=limtL(X,t)    (4)

It is well known that the integral in Equation 3 could be infinite when the support of X is unbounded. We will insist that the above integrals in the right-hand side of the above Equation 2 are finite and therefore μ < ∞. This can be shown to be equivalent:

limt  tP(|X|t)=limt  t(1P(|X|t)=limt  t(1F(t)+F(t))=0    (5)

and therefore, for distributions with possible negative values, we have

limx-xF(x)=0    (6)

This condition will ensure that all terms in subsequent equations are finite.

We will find it convenient to introduce the following auxiliary integral:

I(z)=xzx dF(x)    (7)

This is a partial first moment of X computed over all xz. We can express H(X, a) in terms of the cumulative distribution function F(·) and auxiliary integral I(·) as follows:

H(X,a)=ΩL(a)(ax)dF(x)+ΩR(a)(xa)dF(x)                   =(aΩL(a)dF(x)ΩL(a)xdF(x))                   +(ΩR(a)dF(x)aΩR(a)dF(x))                   =(aF(a)I(a))+(μI(a)a(1F(a))                   =a(2F(a)1)+μ2I(a)    (8)

Consider the function h(x) = xF(x). Then dh(x) = F(x) dx + x dF(x) and from the integration by parts formula, we can express I(a) as

I(a)=xax dF(x)=aF(a)-xaF(x) dx    (9)

Substituting this into Equation 8 we obtain

H(X,a)=(μ-a)+2xaF(x) dx    (10)

A geometric interpretation of H(X, a) for the case a ≤ μ is illustrated in Figure 1.

Figure 1
www.frontiersin.org

Figure 1. Geometric interpretation of H(X, a).

The MAD is twice the sum of two areas: the first area is the area under the cumulative distribution function for xa and the second area is (μ − a) · 0.5.

For MAD around the mean μ and the median M, we obtain

H(X,μ)=2xμF(x) dx  andH(X,M)=(μ-M)+2xMF(x) dx    (11)

2.2 MAD formula for discrete distributions

We now derive an analogous formula for the discrete distribution. Let Ω = {x1, x2, …}. Without loss of generality, assume x1 < x2 < ⋯ . By analogy with the continuous case, define the auxiliary sum I(x) as

I(z)=xizxif(xi)    (12)

Then MAD of X from μ is

H(X,a)=xia(axi)f(xi)+xi>a(xia)f(xi)                  =(axiaf(xi)xiaxif(xi))                  +(xi>axif(xi)axi>af(xi))                  =(aF(a)I(a))+((μI(a))a(1F(a))                  =a(2F(a)1)+μ2I(a)    (13)

To compute I(a), we use the following summation by parts formula (analogous to integration by parts) [24]. If U = {u1, u2, …, un} and V = {v1, …, vn} are two sequences and Δ is the forward difference operator, then

k=mnukΔvk=(unvn+1-umvm)-k=m+1nvkΔuk    (14)

To apply the formula, we let m = 1, vk = xk and uk = F(xk). Then we have

k=1nF(xk)Δxk=(xn+1F(xn)-x1F(x1))
-k=2nxkf(xk)=xn+1F(xn)-I(xn)    (15)

If we choose n(a)=argmaxxia{xi} for a given a, then F(a) = F(xn(a)) and we have

I(a)=F(a)·xn(a)+1-k=1nΔxkF(xk)    (16)

in complete analogy for I(a) for the continuous case in Equation 9.

Substituting this expression for I(a) from Equation 16 into Equation 13, we obtain

H(X,a)=(μ-a)+2inΔxiF(xi)    (17)

The sum on the right of Equation 17 is the area under the cumulative distribution function for xia and can be interpreted as the integral of the CDF computed for xa in direct analogy with the continuous case.

For MAD about mean μ and median M, the above equation gives us

H(X,μ)=2xiμΔxiF(xi)   and H(X,M))=(μM)+2xiMΔxiF(xi)    (18)

The advantage of these formulas is that for many distributions, xi are consecutive integers, giving us Δxi = 1. In such cases, the above formulas for H(X, μ) and H(X, M) simplify to

H(X,μ)=2xiμF(xi) andH(X,M)=(μ-M)+2iMF(xi)    (19)

with a direct analogy to the continuous case.

We will apply the above results to compute MADs of several well-known discrete distributions in Section 6.

3 Interpretation of MADs via “folded” cumulative distribution functions

We can interpret H(x, a) as follows. Consider the function G(x) defined as follows:

G(x,a)={F(x)xΩL(a)1-F(x)xΩR(a)0otherwise    (20)

We can interpret G(x) as the "folded" cumulative distribution function curve. Note that unless F(a) = 1/2 (i.e., a is the median M), the curve G(x) is discontinuous at x = a. The case a = μ is illustrated in Figure 2.

Figure 2
www.frontiersin.org

Figure 2. Illustration of folded cumulative distribution function (a = μ) and X ≥ 0.

For non-negative valued random variables X, we can interpret MAD H(X, a) in terms of G(x) as follows: let A(G) denote the area under this curve. For continuous case, we have

A(G)=ΩG(x,a) dx=ΩL(a)F(x) dxAL+ΩR(a)(1-F(x)) dxAR          =ΩL(a)[yxdF(y)]dx+ΩR(a)[y>xdF(y)]dx          =ΩL(a)[yadx]dF(y)+ΩR(a)[aydx]dF(y)          =ΩL(a)(a-y)dF(y)+ΩR(a)(y-a)dF(y)          =Ω|a-y|dF(y)=H(X,a)

Comparing Equation 21 with Equation 8, we can interpret AL = aF(a) − I(a) as the area under left part of G(x, a) for xa and AR = (μ − I(a) − a(1 − F(a)) as the area under the right part of G(x, a) for x > a.

For the case a = μ, the areas under the curve for xa and x > a are equal. In this case, H(X, μ) = 2AL(G) is just twice the area under the "left" part of the folded CDF curve for xa. This is illustrated in Figure 2. From Equation 9, the auxiliary integral I(μ) is the area above cumulative distribution function F(x) for x ≤ μ and F(x) ≤ F(μ) (shown in yellow color).

For the case a = M, the folded cumulative distribution function G(x) is continuous as shown in Figure 3. The area under this curve is the MAD from the median [25]. It is interesting to compare the MAD H(X, M) around the median with the quartile deviation (QD) defined as half of the interquartile range semi-interquartile range QD = IQR/2 = (Q3Q1)/2 where Q1 and Q3 are the first and third quartile of X. The MAD around the median is the area under the folded CDF G(x) where the quartile deviation is the area spanned between the quartiles on the x-axis and between 0 and 1/2 on the y-axis.

Figure 3
www.frontiersin.org

Figure 3. Mean absolute vs. quantile deviation (a = M and ≥ > 0.

For the discrete case with non-negative valued X, we get a similar result. As before we assume that 0 ≤ x1 < x2 < ⋯  and let n(a)=argmaxxia{xi}. Recall that F(xk) = f(x1) + ⋯ + f(xk). Therefore, we obtain

A(G)=iΔxi-1G(xi,a)=in(a)Δxi-1F(xi)          +i>n(a)Δxi-1(1-F(xi))          =in(a)((xi-a)+(a-xi-1))F(xi)          +i>n(a)((a-xi-1)+(xi-a))(1-F(xi))          =ia(n)(a-xi)ΔF(xi-1)+i>a(n)(xi-a)ΔF(xi-1)                =ia(n)(a-xi)f(xi)+i>a(n)(xi-a)f(xi)                =i|a-xi|f(xi)=H(X,a)

in exact analogy with the continuous case. Therefore, the MAD H(X, a) can also be interpreted as the area under the folded CDF curve G(x, a).

4 Some special functions and integrals

For many distributions, their cumulative distribution functions involve some special functions (for real arguments) such as Gamma and Beta functions [26, 27]. Indefinite integrals for many of these functions are readily available [2830]. Some of these functions and their integrals are presented here for easy reference.

4.1 Some special functions and integrals

1. Error function ([30], Section 7.2 in Olver et al. [31]):

erf(z)=2π0ze-t2dt    (21)

2. Density of the standard normal [31, 32]:

ϕ(x)=12πe-x2/2    (22)

3. Cumulative distribution of the standard normal [31, 32]:

Φ(x)=12π-xe-t2/2dt=12+12erf(x2)    (23)

4. Lower and upper incomplete Gamma functions ([30], Section 8.6 in Olver et al. [31]):

ν(s,x)=0xts-1e-tdt  and  Γ(s,x)=xts-1e-tdt    (24)

5. Gamma function and regularized Gamma function ([30], Section 5.2 in Olver et al. [31]):

Γ(s)=0ts-1e-tdt=ν(s,x)+Γ(s,x) andQ(s,x)=Γ(s,x)Γ(s)    (25)

6. Beta function (Section 5.12 in Olver et al. [31] and Gupta and Nadarajah [33]):

B(z1,z2)=01tz1-1(1-t)z2-1dt=Γ(z1)Γ(z2)Γ(z1+z2)    (26)

7. Incomplete Beta function and regularized incomplete Beta functions ([33], Section 8.17 in Olver et al. [31]):

B(x;a,b)=0xta-1(1-t)b-1dt andIx(a,b)=B(x;a,b)B(a,b)    (27)

4.2 Some indefinite integrals

The following expressions will be used in many of the computations (C is a constant).

1. Integration of standard normal CDF (p. 396 in Owen [32]):

Φ(a+bx) dx=(a+bx)Φ(a+bx)+ϕ(a+bx)b+C    (28)

2. Integration of the product of standard normal density and x (integral #11, p. 393 in Owen [32]):

xϕ(ax) dx=-ϕ(ax)a+C    (29)

3. Integration of the product of error function and exponential (integral #1, Section 4.2 in Ng and Geller [30]):

ebx erf(az)dz=ebzb erf(az)-1bexp(b24a2)erf(az-b2a)+C    (30)

4. Integration of incomplete gamma function (using integration by parts and Equation 24):

xb-1ν(α,z)dz=1b(zbν(α,z)-ν(α+b,z))+C    (31)

In particular, for b = 1 using the recurrence relation (chapter 8.6 in Olver et al. [31]),

ν(α+1,z)=zν(s,z)-zαe-z

we obtain

ν(α,z)dz=zαe-z+C    (32)

5. Integration of the incomplete beta function [33]

B(x;a,b) dx=xB(x;a,b)-B(x;a+1,b)+C    (33)

6. Integration of reciprocal functions with exponentials (integral 2.313 p. 92 in Gradshteyn and Ryzhik [29])

dxa+bemx=mx-log(1+bemx)am+C    (34)

5 MAD computations for continuous distributions

We now illustrate the above formula for some well-known continuous distributions. We will focus on MAD around the mean μ. We will consider the following continuous distributions (in alphabetical order): Beta, chi-squared, exponential, Gamma, Gumbel type I, half-normal, Laplace, logistic, Maxwell, normal, Pareto type I and II, Rayleigh, Student-t, triangular, uniform, and Weibull. For some of these distributions, the expressions for MAD have been known. In contrast, to our knowledge, the expressions for MAD have not been published for some distributions like Pareto II and Weibull.

1. Beta distribution with parameters α, β > 0: The corresponding density f(x) and the cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34]:

f(x)=xα-1(1-x)β-1B(α,β) and F(x)=Ix(α,β),x(0,1)

where Ix(·) denotes the regularized incomplete beta function in Equation 27. The mean and standard deviation of this distribution are

μ=αα+β and σ=1α+βαβα+β+1

Therefore, using Equation 33 we compute H(X, μ) as follows:

H(X,μ)=20μIx(α,β) dx               =2xB(x;α,β)-2B(x;α+1,β)B(α,β)|0μ               =2μB(μ;α,β)-2B(μ;α+1,β)B(α,β)    (35)

There is no closed-form formula for the median M. The MAD around the median M is

H(X,M)=(μ-M)+20MIx(α,β) dx=(αα+β-M)                   +2M·B(M;α,β)-2B(M;α+1,β)B(α,β)    (36)

2. Exponential distribution with rate λ > 0 : The corresponding density f(x) and cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34]:

f(x)=λe-λx  and  F(x)=1-e-λx,   x0

Its mean μ = 1/λ and standard deviation σ = 1/λ. We compute H(X, μ) as

H(X,μ)=201/λ(1-e-λx) dx=2(x+e-λxλ)|01/λ=2λe=(2e)σ    (37)

In terms of σ, for this distribution H(X, μ) = (2/e)σ ≈ 0.74σ.

The median is M = (log2)/λ. The MAD about median is then

H(X,M)=(μ-M)+20log2/λ(1-e-λx) dx               =(1λ-log2λ)+2(x+e-λxλ)|0log2/λ=log2λ    (38)

3. F (Fisher–Snedecor) distribution with α, β > 0 degrees of freedom: The corresponding density f(x) and cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34],

f(x α,β)=(αx)αββ(αx+β)α+βxB(α2,β2) and F(x)=Iz(α2,β2),where{z=αx/(αx+β)x>0,if α=1x0if α>0

The mean μ and standard deviation σ are defined only for β > 2 and β > 5 respectively and are given by

μ=ββ-2, for β>2 andσ=β(β-2)2(α+β-2)α(β-4), for β>4

To compute H(X, μ) we note that x = βz/(α(1 − z)) and therefore, dz = (β/α(1 − z)2)dz. Let μ* = αμ/(αμ + β) = α/(α + β − 2β). Therefore, for the MAD, we obtain

H(X,μ)=20μF(x) dx=2βα0μ*Iz(α/2,β/2)(1-z)2dz               =2βαB(α/2,β/2)[0μ*B(z;α/2,β/2)(1-z)2dz]

To compute the integral, we note that B′(z; α/2, β/2) = zα/2 − 1(1 − z)β/2 − 1. If we consider the function h(z) = B(z; α/2, β/2)/(1 − z) then

h(z)=B(z;α/2,β/2)(1-z)+B(z;α/2,β/2)(1-z)2          =zα/2(1-z)β/2-1+B(z;α/2,β/2)(1-z)2

This immediately gives us the expression for the integral

B(z;α/2,β/2)(1-z)2dz=B(z;α/2,β/2)(1-z)                                    -B(z;α/2,β/2-1)

We can now compute the MAD for the F distribution:

H(X,μ)=2βαB(α/2,β/2)[B(z;α/2,β/2)(1-z)-B(z;α/2,β/2-1)]|0μ*=2βαB(α/2,β/2)[B(μ*;α/2,β/2)(1-μ*)-B(μ*;α/2,β/2-1)]    (39)

4. Gamma distribution with shape parameter α > 0 and rate β > 0: The corresponding density f(x) and cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34]:

f(x)=βαxα-1e-βxΓ(α)  and  F(x)=ν(α,βx)Γ(α),x0

where Γ(·) denotes the Gamma function (Equation 25 and ν(·) denotes the lower incomplete Gamma function (Equation 24). The mean of this distribution is μ = α/β and its standard deviation σ=α/β.

Let us define z = βx. Then we have

H(X,μ)=2Γ(α)0μν(α,βx) dx=2βΓ(α)0αν(α,z)dz               =2βΓ(α)[zαe-z]|0α=2βααe-αΓ(α)    (40)

There is no closed-form expression for the median M. The MAD (around median) is

H(X,M)=(μ-M)+2xMF(x) dx=(αβ-M)               +2βΓ(α)0βMν(α,βx)d(βx)               =(αβ-M)+2βΓ(α)[zαe-z]|0βM=(αβ-M)               +2β(βM)αΓ(α)e-βM    (41)

5. χ2 Distribution with k degrees of freedom: The corresponding density f(x) and the cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34]:

f(x)=xk/2-1e-x/22k/2Γ(k/2)  and  F(x)=ν(k/2,x/2)Γ(k/2),x>0

where Γ(·) denotes the Gamma function (Equation 25 and ν(·) denotes the lower incomplete Gamma function (Equation 24. The mean of this distribution is μ = k and its standard deviation σ=2k. This is a special case of Gamma distribution with shape α = k/2 and rate β = 1/2. This immediately gives us

H(X,μ)=(2e)-k/2kk/2Γ(k/2) andH(X,M)=(k-M)+(M/2)k/2e-M/2Γ(k/2)    (42)

6. Gumbel type I (extreme value distribution) with location μ and scale β > 0: The corresponding density and f(x) and cumulative distribution function F(x) are given by [20]):

f(x)=1βe-(z+e-z) and F(x)=e-e-z,xR,z=(x-μ)/β

Its mean is μ* = (μ + βγ) where γ is the Euler-Mascheroni constant. Its standard deviation σ=βπ/6. Let t = ez. Then dt = (t/β)dx or dx = βt−1dt. Then for MAD (about mean) we obtain

H(X,μ*)=2xμ*F(x) dx=2β0e-γe-tt-1dt=2βν(0,e-γ)    (43)

where ν(·) denotes the incomplete lower gamma function in Equation 24.

The median for this distribution is M = μ − β log(log2). Therefore, for the MAD about the median we have

H(X,M)=(μ*-M)+2xMF(x) dx=β(γ+log(log2))               +2β0log2e-zz-1dz               =β[γ+log(log2)+2ν(0,log2)]    (44)

7. Half-normal distribution with scale parameter σ > 0: The corresponding density f(x) and cumulative distribution function F(x are given by Johnson and Kotz [20]):

f(x)=2σπe-x2/(2σ2), and F(x)=erf(xσ2).  x0

The mean is μ=σ2/π and standard deviation σ*=σ(1-2/π). Let z=x/(σ2). Then dx=σ2dz and we have

H(X,μ)=2σ201/πerf(z)dz               =2σ2[z erf(z)+1πe-z2]|01/π               =2σ2π[erf(1π)+e-1/π-1]    (45)

The median is M=σ2erf-1(1/2). Define c = erf−1(1/2). Then the MAD around the median M we have

H(X,M)              =(M-μ)+2σ20cerf(z)dz               =σ2(1π-c)+2σ2[z erf(z)+1πe-z2]|0c                                         =σ2π(2e-c2-1)    (46)

8. Laplace distribution with location μ and scale b: The density f(x) and cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34]:

f(x)=12be-|x-μ|/b andF(x)={12e(x-μ)/b,if xμ1-12e-(x-μ)/b,if xμ

For this distribution, μ = M = b and σ=b2. For the MAD, we obtain

H(X,μ)=H(X,M)=μe(x-μ)/bdx=be(x-μ)/b|-μ=b    (47)

In terms of σ, we have H(X,μ)=σ2/20.71σ.

9. Logistic distribution with location μ and scale s > 0: The corresponding density function f(x) and cumulative distribution function F(x are given by Johnson and Kotz [20]:

f(x)=-e-(x-μ)/ss(1+e-(x-μ)/s)2, and F(x)=11+e-(x-μ)/s,xR

The mean and the median are μ and the standard deviation is σ=sπ/3. Let z = (x − μ)/s. Since for this distribution M = μ, we have H(X, μ) = H(X, M). Using Equation 34 we obtain

H(X,μ)=H(X,M)=2μdx1+e(xμ)/s=20zdz1+ez                                              =2s[z+log(1+ez)]|0                                              =2s[z+log(ez(ez+1))]|0                                              =2slog(ez+1)|0=2slog2    (48)

In terms of σ we have H(X,μ)=(23log2/π)σ0.76σ

10. Log-normal distribution with parameters μ ∈ (∞, +∞) and σ2 (σ > 0): The density f(x) and cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34]:

f(x)=1xσ2πe-(logx-μ)2/2σ2=xσϕ(logx-μσ),  andF(x)=Φ(logx-μσ),x>0

We will use the notation μ* and σ* to denote the mean and standard deviation of X to distinguish it from the mean μ and standard deviation σ of the underlying normal distribution. The mean of the log-normal distribution is μ* = eμ+σ2/2 and its standard deviation σ*=eμ+σ2/2eσ2-1.

To compute H(X, μ), define z = (log x − μ)/σ. Then, dx = σx dz = σeμ + σzdz and using Equation 30 with a=1/2 and b = σ, we obtain

H(X,μ*)=20μ*Φ(logx-μσ) dx=2σeμ-σ/2eσzΦ(z)dz               =2σeμ-σ/2eσz[12+12 erf(z2)]dz               =2eμ+σ2/2 erf(σ22)    (49)

The median M = eμ and therefore for the MAD around median, we obtain

H(X,M)=(μ*-M)+20MΦ(logx-μσ) dx               =eμ(eσ2/2-1)+2σeμ-0eσzΦ(z)dz               =eμ(eσ2/2-1)+2σeμ-0eσz[12+12 erf(z2)]dz               =eμ+σ2/2+σeμ-0eσz erf(z2)dz               =eμ+σ2/2 erf(σ2)    (50)

11. Maxwell (or Maxwell-Boltzmann) distribution with scale parameter b > 0: The density function f(x) and the cumulative distribution function F(x) are given by Johnson and Kotz [20]

f(x)=1b32πx2e-x2/(2b)2  andF(x)=2Φ(xb)-1b2πxe-x2/(2b)2   x0

We can re-write these as

f(x)=2(xb)2ϕ(xb) andF(x)=2Φ(xb)-2(xb)ϕ(xb)x0

The mean is μ=2b2/π and standard deviation σ=b(3π-8)/π. We let z = x/b. Then dx = b dz and using Equations 28, 29, with a = 1/b, we obtain

H(X,μ)=4b022(Φ(z)-zϕ(z))dz               =4b(zΦ(z)+2ϕ(z))|022/π               =4b2π(erf(2π)+e-π/4)    (51)

There is no simple closed-form for the median M.

12. Normal distribution N(μ, σ): The density f(x) and cumulative distribution function F(x) are [20, 34]:

f(x)=1σϕ(x-μσ) and F(x)=Φ(x-μσ),xR

where Φ(·) denotes the cumulative distribution of the standard normal (Equation 23). Using Equation 28 with a = −μ/σ and b = 1/σ we obtain

H(X,μ)           =2xμΦ(xμσ)             dx=2σ(xΦ(x)+f(x))|0=σ2π    (52)

In terms of σ, we have H(X,μ)=σ2/π0.80σ. Since this is a symmetric distribution H(X, M) = H(X, μ).

13. Pareto type I distribution with shape α > 0 and scale β > 0: The density f(x) and the cumulative distribution function F(x) are given by Johnson and Kotz [20]

f(x)={αβαxα+1xβ0x<β    and    F(x)={1-(βx)αxβ0  x<β

the mean and standard deviation for this distribution are

μ=αβ(α-1) if α>1 andσ=β(α-1)α(α-2) if α>2

This distribution has infinite mean μ for α ≤ 1 and infinite variance for α ≤ 2. We assume that α > 1. Then for MAD, we obtain

H(X,μ)=2βμ(1-(βx)α) dx=2[x+βα(α-1)x-α+1]|βμ               =2[αβ(α-1)+βα(α-1)(αβα-1)-α+1]               -2[β+βα(α-1)β-α+1]               =2βα(1-1α)α-2    (53)

The median is M=β2α. Therefore, for the MAD about the median, we obtain

H(X,M)=(μ-M)+2βM(1-(βx)α) dx               =(αβ(α-1)-β2α)+2[x+βα(α-1)x-α+1]|βM               =(αβ(α-1)+β2α)+2[βα(α-1)x-α+1]|ββ2α               =αβ(2α-1)(α-1)    (54)

14. Generalized Pareto (Type II) distribution with location μ ∈ R, scale σ > 0, and shape ξ ∈ R: We exclude the case ξ = μ = 0, that corresponds to the exponential distribution. The density f(x) and the cumulative distribution function F(x) are given by Johnson and Kotz [20]:

f(x)={1σ(1+ξ(x-μ)σ)-1/ξ-1ξ0exp(-(x-μ)σ)ξ=0 andF(x)={1-(1+ξ(x-μ)σ)-1/ξξ01-exp(-(x-μ)σ)ξ=0

where the support of X is x ≥ μ for ξ ≥ 0 and μ ≤ x ≤ σ/ξ for ξ < 0. The mean μ* and standard deviation σ* of this distribution are defined only for ξ < 1 and ξ < 1/2 respectively and are

μ*=μ+σ(1-ξ),ξ<1 andσ*=σ(1-ξ)1-2ξ,ξ<1/2

To compute H(X, μ*), we need to consider two cases

(a) ξ < 1 and ξ ≠ 0. We have

H(X,μ*)=2μμ*F(x) dx               =[x+σ1-ξ(1+ξ(x-μ)σ)1-1/ξ]|μμ*               =2[μ+σ1-ξ+σ(11-ξ)2-1/ξ]                     -2[μ+σ1-ξ]=2σ(11-ξ)2-1/ξ    (55)

Note that the case ξ = −1 gives H(X, μ*) = σ/4 and corresponds to continuous uniform distribution U(0, σ). The case ξ > 0 and location μ = σ/ξ corresponds to the Pareto distribution with scale β = σ/ξ and shape α = 1/ξ.

(b) Case ξ = 0. For this case, we obtain

H(X,μ*)=2x=μμ*F(x) dx               =2[x+σexp(-(x-μ)σ)]|μμ*               =2[μ+σ1-ξ-σexp(1-ξ)]-2[μ+σe]               =2σ[11-ξ-1exp(1-ξ)+1e]

Note that for ξ = 0, the case μ = 0 corresponds to an exponential distribution with parameter 1/σ.

For μ = 0 and ξ > 0, let α = 1/ξ and λ = ξσ. Then we obtain the Lomax distribution with shape α > 0 and scale λ > 0. Its mean μ* = λ/(α − 1) is defined only for α > 1 (corresponding to ξ < 1). Therefore, for α > 1 its MAD for the Lomax distribution is

H(X,μ*)=2λα(αα-1)2-α

15. Rayleigh distribution with scale parameter σ > 0: The density function f(x) and the cumulative distribution function F(x) are given by Johnson and Kotz [20]:

f(x)=-xσ2e-x2/(2σ)2, and F(x)=1-e-x2/(2σ)2,x0

The mean is μ=σπ/2 and standard deviation is σ*=σ(4-π)/2. Let z = x/2σ. Then dx = 2σ dz and we have

H(X,μ)=2σπ2-4σ0π/22e-z2dz               =2σπ(22-erf(π22))    (56)

For this distribution, the median M=σ2log2. Therefore,

H(X,M)=(μM)+2xM(1ex2/(2σ)2)dx                     =σ(π/22 log 2)+2σ2 log 2                    4σ0(log 2)/2ez2dz                    =σ[π/2+2 log 2)2π erf(log 22)]    (57)

16. Student-t distribution with k > 1 degrees of freedom: Its probability density function f(x) and its cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34]:

f(x)=1kB(k/2,1/2)(1+x2k)-(k+1)/2 andF(x)=1-B(1/(1+x2/k),k/2,1/2)2B(k/2,1/2),xR

The mean μ is μ = 0 for k > 1, otherwise it is undefined. The standard deviation is defined only for k > 2 and σ = k/(k − 2). From Equation 8 we have

H(X,μ)=-2-0xf(x) dx               =-2kB(1/2,k/2)-0x(1+x2k)-(k+1)/2dx               =-2kB(1/2,k/2)[-k2(k-1)(1+x2k)-(k-1)/2]|-0               =k(k-1)B(k/2,1/2)    (58)

Since this distribution is symmetric, M = μ and H(X, M) = H(X, μ).

17. Triangular distribution with lower limit a, upper limit b and mode c ∈ [0, 1]: The density function f(x) and the cumulative distribution function F(x) is given by [20, 34, 35]:

f(x)={2(xa)(ba)(ca)2(bx)(ba)(bcF(x)={(xa)2(ba)(ca)for 0xc,1(bx)2(ba)(bc)for c<x1

For this distribution, μ = (a + b + c)/3 and σ=(b-a)2-(c-a)(b-c)/32. It is easy to show that for 0 ≤ c ≤ (a + b)/2, we have μ ≥ c and for (a + b)/2 ≤ cb, we have μ ≤ c. Therefore,

H(X,μ)               =2xμF(x) dx               ={23(b-a)(b-c)(2b-(a+c)3)3for 0c(a+b)/223(b-a)(c-a)(b+c-2a3)3for (a+b)/2c

The median M of this distribution is

M={b-22(b-a)(b-c)c(a+b)/2a+22(b-a)(c-a)c(a+b)/2    (59)

Therefore, the MAD about the median is then

H(X,M)=(μ-M)+xMF(x) dx               ={(b-a)+(b-c)-2(b-a)(b-c)3ac(a+b)/2(b-a)+(c-a)-2(b-a)(c-a)3(a+b)/2cb    (60)

18. Uniform continuous distribution in [a, b]: The density function f(x) and the cumulative distribution function F(x) are given by Johnson and Kotz [20] and Feller [34]:

f(x)=1(b-a) and F(x)=(x-a)(b-a)x[a,b]

Mean is μ = (a + b)/2 and standard deviation σ=(b-a)/23. The MAD is

H(X,μ)=20μ(x-a)(b-a)dx=(x-a)2(b-a)|0(a+b)/2               =(b-a)4=32σ    (61)

19. Weibull distribution with scale α > 0 and shape k > 0: The density f(x) and the cumulative distribution function F(x) are given by Johnson and Kotz [20].

f(x)=kα(xα)k-1e-(x/α)k and F(x)=1-e-(x/α)k,x0

Its mean is μ = αΓ(1 + 1/k) and its standard deviation is σ = α(Γ(1 + 2/k) − Γ2(1 + 1/k)), If we let z = x/α then dx = α dz and using Equation 24 we obtain

H(X,μ)=20Γ(1+1k)(1-e-zb)dz               =2αkΓ(1k,zk)|0Γ(1+1k)+2μ               =2αkΓ(1k,Γk(1+1k))-2αkΓ(1k)               +2αΓ(1+1k)               =2αkΓ(1k,Γk(1+1k))    (62)

For k = 1 we have Γ(1 + 1/k) = 1, μ = 2α, and Γ(1, 1) = 1/e. Therefore, H(X, μ) = 2α/e. This case with k = 1 corresponds to exponential distribution with λ = 1/α.

The median M=αlog2k. Therefore for MAD about median, we obtain

H(X,M)=(μM)+20M(1ezb)dz                     =(μ+M)+2αkΓ(1k,zk)|0M                     =α(Γ(1+1k)+log2k+2k[Γ(1k,αklog2)                     Γ(1k)])                     =α[log2k+2kΓ(1k,αklog2)1kΓ(1k)]

6 MAD computations for discrete distributions

In this section, we compute MADs for the following discrete distributions (in alphabetical order): binomial, geometric, logarithmic, negative binomial, Poisson, uniform, Zipf, and zeta distributions. For some of these distributions, the expressions for MAD have been known, whereas for some distributions like logarithmic, the formulae for MAD have not been published, to our knowledge.

1. Binomial distribution with parameters nN and success probability 0 < p < 1: Let q = 1 − p. Its density f(x) and umulative distribution function F(x) are [20, 34]:

f(k)=(Nk)pk(1-p)N-k  and F(k)=Iq(N-k,1+k)k={0,1,,N}

where Iq(·) is the regularized incomplete beta function defined in equation 27.

The mean μ = Np and standard deviation is σ=Np(1-p). Let n = ⌊Np⌋. Then we have

H(X,μ)=2nF(n)-2I(n)               =2[Npk=0n(Nk)pk(1-p)N-k]               -2[k=0nk(Nk)pk(1-p)N-k]               =2k=0n(Np-k)(Nk)pk(1-p)N-k               =(N+1)(Nk+1)pn+1(1-p)N-n    (63)

The last step in the above equation can be established by induction on N.

2. Geometric distribution with success probability 0 < p < 1: The density f(x) and cumulative distribution function are [20, 34]:

f(k)=p(1-p)k  and F(k)=1-(1-p)k+1,k=0,1,2,

The mean μ = (1−p)/p and standard deviation is σ=1-p/p. Let n = ⌊μ⌋ = ⌊1/p⌋ − 1. Therefore, we obtain

H(X,μ)=2knF(n)=2kn(1-(1-p)k+1)               =2(n+1)-k=0n(1-p)k+1               =(n+1)-(1-p)(1-(1-p)n+1)p               =(1-p)1/p1p=2(n+1)(1-p)n+1    (64)

3. Logarithmic distribution with parameter 0 < p < 1: The density f(x) and cumulative distribution function are [20]:

f(k)=-pkklog(1-p) and F(k)=1+B(p;k+1,0)log(1-p),k=1,2,

where B(·) is the incomplete beta function. Its mean μ = −p/((1 − p)log(1 − p)).

Let n = ⌊μ⌋. Therefore, for the MAD, we obtain

H(X,μ)=2k=1nF(k)=k=1n(1+B(p,k+1,0)log(1-p))               =2n+2log(1-p)k=1nB(p,k+1,0)               =2n+2log(1-p)k=1n[0ptk(1-t)dt]               =2n+2log(1-p)0p[k=1ntk(1-t)dt]               =2n+2log(1-p)0pt(1-tn)(1-t)2dt               =2n+2log(1-p)[0pt(1-t)2dt]               -2B(p;n+2,-1)log(1-p)

To evaluate the integral in the above equation, we note that log′(1 − t) = −1/(1 − t) and (1/(1−t))′ = −1/(1 − t)2. Therefore, for this integral, we have

0pt(1-t)2dt=[t(1-t)+log(1-t)]|0p                              =p(1-p)+log(1-p)                              =-μlog(1-p)+log(1-p)

Substituting this into the above equation for H(X, μ), we obtain

H(X,μ)=2-2B(p;n+2,-1)log(1-p)    (65)

4. Negative binomial distribution with number of successes parameter r > 0 and success probability 0 < p < 1: The density f(x) and cumulative distribution function are [20, 34]:

f(k)=(k+r-1k)(1-p)kpr andF(k)=Ip(r,k+1),k=0,1,2,

where Ip(·) denotes the regularized incomplete beta function.

The mean is μ = r(1 − p)/p and standard deviation is σ=r(1-p)/p. Let n = ⌊μ⌋. Then from Equation 26 we have F(n)=Ip(r,n+1)=(1-p)npr-1/B(r,n+1). On the other hand, we have

F(n)=k=0nf(k)=k=0n(k+r-1k)            (-k(1-p)k-1pr+r(1-p)kpr-1)            =k=0n[-kf(k)(1-p)+rf(k)p]=-I(n)(1-p)+rF(n)p

From this, we obtain

I(n)=r(1-p)pF(n)-(1-p)n+1pr-1B(r,n+1)=nF(n)-(1-p)n+1pr-1B(r,n+1)

Using Equation 26 for B·), we obtain for the MAD

H(X,μ)=2nF(n)-2I(n)=2(1-p)n+1pr-1B(r,n+1)                                 =2r(n+rr)(1-p)n+1pr-1    (66)

5. Poisson distribution with rate λ > 0: The density f(x) and cumulative distribution function F(x) are [20, 34]:

f(k)=λke-λk! and F(k)=Q(k+1,λ)λ>0,k=1,2,

where Q(·) is the regularized Gamma function in Equation 25.

The mean is μ = λ and standard deviation is σ=λ. Let n = ⌊λ⌋. For this distribution, we have

F(n)=k=1nF(k)=k=1n[-λke-λk!+kλ·λke-λk!]=-F(n)+I(n)λ

On the other hand, from Equation 24, we have F′(n) = Q′(n + 1, λ) = −e − λλn/n!. Therefore, I(n) = λF(n) − e − λλn+1/n! and we obtain for the MAD

H(X,μ)=2μF(μ)-2I(μ)=2nF(n)-2[λF(n)-e-λλn+1n!]               =2e-λλn+1n!    (67)

6. Uniform (discrete) distribution on interval [1, N]: The density f(x) and cumulative distribution function F(x) are [20, 34]:

f(k)=1N  and F(k)=kN,k={1,,N}

Its mean μ = N/2 and its standard deviation is σ=N2-1/23. Let n = ⌊μ⌋. Then, for the MAD, we obtain

H(X,μ)=2i=1nF(k)=(n+1)nN

There are two cases to consider:

(a) N is odd. Then N = (2m + 1) for some m and μ = (m + 1). Then n = (m + 1) and H(X, μ) = (m + 1)(m + 2)/N = (N − 1)(N + 1)/4N.

(b) N is even. Then N = 2m for some m and μ = (m + 1/2). Then n = m and H(X, μ) = m(m + 1)/N = N/4

We summarize both cases as follows:

H(X,μ)={(N-1)(N+1)4NN is oddN4N is even    (68)

For odd N we can express MAD in terms of σ as H(X, μ) = 12σ2/N

7. Zipf and zeta distributions We start with generalized Zipf distribution [20, 34]. Assume X is distributed according to zeta distribution with parameters s > 1 and integer N. Its density f(x) and cumulative distribution functions are

f(k)=k-shN,s and F(k)=hk,shN,s,k={1,2,,N}

where hk, s is the generalized harmonic number defined as hk,s=iki-s. The mean and standard deviation of this distribution are

μ=hN,s-1hN,s  and  σ=hN,s-2hN,s-hN,s-12hN,s

Let n = ⌊μ⌋. Then, using the following equation for the sum of harmonic numbers

k=1nhk,s=(n+1)hn,s-hn,s-1fors>1

we obtain the MAD of the Zipf's distribution with N elements

H(X,μ)=2k=1nhk,shN,s=2[(n+1)hn,s-hn,s-1hN,s]    (69)

For s > 1, we can extend to N = ∞. In this case, we have the Zeta distribution, with its density f(x) and cumulative distribution function given by

f(k)=k-sζ(s) and F(k)=hk,sζ(s),k=1,2,

where ζ(s) is the Riemann zeta function ζ(s)=i>0i-s. The mean μ is defined for s > 2 and standard deviation σ is defined for s > 3 and are

μ=ζ(s-1)ζ(s),s>2  and  σ=ζ(s-2)ζ(s)-ζ2(s)ζ(s),s>3

Let n = ⌊ζ(s − 1)/ζ(s)⌋. Then, using the above equation for the sum of harmonic numbers we obtain for the MAD of Zeta distribution:

H(X,μ)=2k=1nhk,sξ(s)=2[(n+1)hn,s-hn,s-1ξ(s)]    (70)

7 Concluding remarks

This article presented a unified approach to compute MADs from cumulative distribution functions. The computation involves computing integrals of these functions often expressed in closed form in terms of special functions such as beta, gamma, or error functions. For many distributions, the integrals of these functions are well known and the resulting expressions for MADs can be easily obtained. For some of these distributions, the obtained results have not been reported elsewhere. We hope that this study will be useful for researchers and practitioners interested in using mean absolute deviations.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding author/s.

Author contributions

EP: Writing – original draft, Writing – review & editing.

Funding

The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.

Acknowledgments

The author would like to thank Metropolitan College of Boston University for their support.

Conflict of interest

The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fams.2025.1487331/full#supplementary-material

References

1. Papoulis A. Probability, Random variable, and Stochastic Processes, 2nd Edn. New York, NY: McGraw-Hill (1984).

Google Scholar

2. Bloomfield P, Steiger WL. Least Absolute Deviations: Theory, Applications and Algorithms. Boston, MA: Birkhauser. (1983). doi: 10.1007/978-1-4684-8574-5

Crossref Full Text | Google Scholar

3. Rudin W. Principles of Mathematical Analysis. New York, NY: McGraw-Hill (1976).

Google Scholar

4. Elsayed EA. On uses of mean absolute deviation: shape exploring and distribution function estimation. arXiv. (2022) [Preprint]. doi: 10.48550/arXiv.2206.09196

Crossref Full Text | Google Scholar

5. Schwertman NC, Gilks AJ, Cameron J. A simple noncalculus proof that the median minimizes the sum of the absolute deviations. Am Stat. (1990) 44:38–41. doi: 10.2307/2684955

Crossref Full Text | Google Scholar

6. Shad SM. On the minimum property of the first absolute moment. Am Stat. (1969) 23:27. doi: 10.2307/2682577

Crossref Full Text | Google Scholar

7. Farebrother RW. The historical development of the L_1 and L_∞ estimation methods. In:Dodge Y, , editor. Statistical data Analysis Based on the L_1-norm and Related Topics. Amsterdam: North-Holland (1987), p. 37–63.

Google Scholar

8. Portnoy S, Koenker R. The Gaussian hare and the Laplacian tortoise: computability of square-error versus abolute-error estimators. Stat Sci. (1997) 2:279–300. doi: 10.1214/ss/1030037960

Crossref Full Text | Google Scholar

9. Gorard S. Revisiting a 90-year-old debate: the advantages of the mean deviation. Br J Educ Stud. (2005) 53:417–30. doi: 10.1111/j.1467-8527.2005.00304.x

Crossref Full Text | Google Scholar

10. Pham-Gia T, Hung TL. The mean and median absolute deviations. J Math Comput Modell. (2001) 34:921–36. doi: 10.1016/S0895-7177(01)00109-1

Crossref Full Text | Google Scholar

11. Dodge Y. Statistical Data Analysis Based on the L_1 Norm and Related Topics. Amsterdam: North-Holland (1987).

Google Scholar

12. Elsayed KMT. Mean absolute deviation: analysis and applications. Int J Bus Stat Anal. (2015) 2:63–74. doi: 10.12785/ijbsa/020201

Crossref Full Text | Google Scholar

13. Gorard S. An absolute deviation approach to assessing correlation. Br J Educ Soc Behav Sci. (2015) 53:73–81. doi: 10.9734/BJESBS/2015/11381

Crossref Full Text | Google Scholar

14. Gorard S. Introducing the mean deviation “effect" size. Int J Res Method Educ. (2015) 38:105–14. doi: 10.1080/1743727X.2014.920810

Crossref Full Text | Google Scholar

15. Habib E. Mean absolute deviation about median as a tool of exploratory data analysis. Int J Res Rev Appl Sci. (2012) 11:517–23.

Google Scholar

16. Leys C, Klein O, Bernard P, Licata L. Detecting outliers: do not use standard deviation around the mean, use absolute deviation around the median. J Exp Soc Psychol. (2013) 49:764–6. doi: 10.1016/j.jesp.2013.03.013

Crossref Full Text | Google Scholar

17. Yager RR, Alajilan N. A note on mean absolute deviation. Inf Sci. (2014) 279:632–41. doi: 10.1016/j.ins.2014.04.016

Crossref Full Text | Google Scholar

18. Pinsky E, Klawansky S. MAD (about median) vs. quantile-based alternatives for classical standard deviation, skew, and kurtosis. Front Appl Math Stat. (2023) 9:1206537. doi: 10.3389/fams.2023.1206537

Crossref Full Text | Google Scholar

19. Fisher RA. A mathematical examination of the methods of determining the accuracy of observation by the mean error, and by the mean square error. Mon Not R Astron Soc. (1920) 80:758–70. doi: 10.1093/mnras/80.8.758

Crossref Full Text | Google Scholar

20. Johnson NL, Kotz S. Distributions in Statistics. New York, NY: J Wiley (1970).

Google Scholar

21. Kenney JF, Keeping ES. Mathematics of Statistics. Princeton, NJ: Van Nostrand (1962).

Google Scholar

22. Weisstein E. Mean Deviation (Wolfram Mathworld). (2011). Available at: https://mathworld.wolfram.com/MeanDeviation.htm (accessed February 1, 2025).

Google Scholar

23. Trapani L. Testing for (in)finite moments. J Econom. (2016) 191:57–68. doi: 10.1016/j.jeconom.2015.08.006

Crossref Full Text | Google Scholar

24. Chu W. Abel's lemma on summation by parts and basic hypergeometric series. Adv Appl Math. (2007) 39:490–514. doi: 10.1016/j.aam.2007.02.001

Crossref Full Text | Google Scholar

25. Xue JH, Titterington DM. The p-folded cumulative distribution function and the mean absolute deviation from the p-quantile. Stat Probab Lett. (2011) 81:1179–82. doi: 10.1016/j.spl.2011.03.014

Crossref Full Text | Google Scholar

26. Andrews GE, Askey R, Roy R. Special Functions. Cambridge, MA: Cambridge University Press. (2000). doi: 10.1017/CBO9781107325937

Crossref Full Text | Google Scholar

27. Olver FW. Introduction to Asymptotics and Special Functions, 1st Edn. New York, NY: Academic Press. (1974). doi: 10.1016/B978-0-12-525856-2.50005-X

Crossref Full Text | Google Scholar

28. Conway JT. Indefinite integrals of products of special functions. Integer Transforms Spec Funct. (2017) 28:166–80. doi: 10.1080/10652469.2016.1259619

Crossref Full Text | Google Scholar

29. Gradshteyn IS, Ryzhik IM. Table of Integrals, Series, and products. New York, NY: Academic Press. (1980).

Google Scholar

30. Ng E, Geller M. A table of integrals of the error function. J Res Natl Bureau Stand B Math Sc. (1969) 738:1–20. doi: 10.6028/jres.073B.001

Crossref Full Text | Google Scholar

31. Olver FW, Lozier DW, Boisvert RF, Clark CW. NIST Handbook of Mathematical Functions. New York, NY: Cambridge University Press (2010).

Google Scholar

32. Owen DB. A table of normal integrals. Commun Stat Simul Comput. (1980) 9:389–419. doi: 10.1080/03610918008812164

Crossref Full Text | Google Scholar

33. Gupta AK, Nadarajah S. Handbook of Beta Distributions and its Applications. New York, NY: CRC Press, Taylor & Francis (2018).

Google Scholar

34. Feller JJ. Probability Theory and Applications. New York, NY: J Wiley (1956).

Google Scholar

35. Wang Y, Pinsky E. Geometry of deviation measures for triangular distributions. Front Appl Math Stat. (2023) 9:1274787. doi: 10.3389/fams.2023.1274787

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: mean absolute deviations, probability distributions, cumulative distribution functions, central absolute moments, folded CDFs

Citation: Pinsky E (2025) Computation and interpretation of mean absolute deviations by cumulative distribution functions. Front. Appl. Math. Stat. 11:1487331. doi: 10.3389/fams.2025.1487331

Received: 27 August 2024; Accepted: 20 January 2025;
Published: 12 February 2025.

Edited by:

George Michailidis, University of Florida, United States

Reviewed by:

Levent Özbek, Ankara University, Türkiye
Maria Luz Gmiz, University of Granada, Spain

Copyright © 2025 Pinsky. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Eugene Pinsky, ZXBpbnNreUBidS5lZHU=

ORCID: Eugene Pinsky orcid.org/0000-0002-3836-1851

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Research integrity at Frontiers

Man ultramarathon runner in the mountains he trains at sunset

94% of researchers rate our articles as excellent or good

Learn more about the work of our research integrity team to safeguard the quality of each article we publish.


Find out more