Professional Documents
Culture Documents
7, 2022-11
Hugo Hernandez
ForsChem Research, 050030 Medellin, Colombia
hugo.hernandez@forschem.org
doi: 10.13140/RG.2.2.16305.86881
Abstract
Keywords
Cite as: Hernandez, H. (2022). Probability Transforms: Expected Value of Functions. ForsChem Research
Reports, 7, 2022-11, 1 - 27. Publication Date: 01/08/2022.
Probability Transforms:
Expected Value of Functions
Hugo Hernandez
ForsChem Research
hugo.hernandez@forschem.org
1. Basic Definitions
Let us consider a randomistic§ variable [1] which can take an arbitrary value following a
particular probability distribution function ( ), such that:
( ) ( )
(1.1)
( )
(1.2)
∑ ( )
(1.3)
Let us now define a new variable , which is a mathematical function of :
( )
(1.4)
The possible values taken by variable will depend on the possible values of . The probability
of taking a particular value ( ) depends on the probability function ( ). If ( ) is an
**
injective function , then:
( ) ( ( ))
(1.5)
where represents the inverse function of .
( ) ∑ ( ( ))
(1.6)
where represents each of the possible values of such that ( ) .
Then the expected value (or probability weighted-average) of the function is (in general):
§
The term randomistic results from combining the terms random (unknown or unpredictable) and
deterministic (known or predictable). A randomistic variable is a general concept comprising both
random and deterministic variables.
**
An injective function is characterized because each value of the function can only be obtained by a
single value of the independent variable.
( ) ∑ ( ) ∑ ∑ ( ( )) ∑ ( ) ( )
(1.7)
( ) ( ) ( )
(1.8)
( ) ( ) ( ) ( )
(1.9)
where represents the backward finite difference operator.
( )
( ) ∑ ( ) ( ) ∑ ( )( )
(1.10)
( )
The term represents the finite probability density of . If (as is the case of
continuous variables), such finite probability density can be represented by the probability
density function ( ):
( )
( )
(1.11)
and Eq. (1.10) becomes:
( ) ∫ ( ) ( )
(1.12)
where .
As we have seen in Eq. (1.7), (1.10) and (1.12), the expected value of (which is a function of )
is determined by summing (or integrating) the products between the function evaluated at
each value and the corresponding probability of . In other words, the expected value
operator transforms the function ( ) into a constant value. The particular value obtained will
depend on both the nature of the function and the probability distribution of .
In Section 2 the concept of probability transform will be explored in more detail, presenting
various definitions such as complete probability transforms, incomplete probability transforms,
probability transform density, and inverse probability transform. In Section 3 the most
representative probability distribution models are considered. Section 4 introduces the
standard probability transform in order to simplify the calculations. Finally, in Section 5,
probability transforms of common mathematical functions are presented for the
representative probability models considered.
2. Probability Transforms
The formulas presented in this section assume that the variables are continuous, or at least,
that they follow the continuous approximation [2]. Equivalent expressions can also be obtained
for discrete variables by using the equivalence: ( ) ( ) and sums instead of integrals.
The complete probability transform of an arbitrary function ( ) is the expected value of the
function:
{ ( )} ∫ ( ) ( )
(2.1)
where ( ) is the probability density function of , such that:
( )
(2.2)
∫ ( )
(2.3)
Notice that curly brackets { } are used in Eq. (2.1) to emphasize the notion of a transform.
(2.4)
( )( ) ( )( )
{ ( )} ∑ ∫ ( ) ∑ { }
(2.5)
The term { } is also known as the -th raw moment of .
{ ( )} ∫ ( ) ( ) ( )
(2.6)
{ ( )} ∫ ( ) ( )
(2.7)
Notice that:
{ ( )} { ( )}
(2.8)
and
{ ( )} { ( )} { ( )}
(2.9)
Also:
{ } ( )
(2.10)
Considering the infinite polynomial expansion given in Eq. (2.4), the incomplete probability
transform of ( ) becomes:
( )( )
{ ( )} ( )( )∫ ( )
∑ ∑ { }
(2.11)
In the case of parametric probability density functions, the incomplete probability transform is
not only function of the parameter vector but also a function of :
{ ( )} ∫ ( ) ( ) ( )
(2.12)
{ ( )}
̃ { ( )} ∫ ( ) ( ) ( ) ( )
(2.13)
̃ { } ( )
(2.14)
Notice also that:
{ ( )} ∫ ̃ { ( )}
(2.15)
{ ( )} ∫ ̃ { ( )}
(2.16)
Considering the infinite polynomial expansion given in Eq. (2.4), the probability transform
density becomes:
( )( ) ( )( )
̃ { ( )} ∑ ( ) ∑ ̃ { }
(2.17)
The term ̃ { } represents the probability transform density of the -th raw moment of .
In the case of parametric probability density functions, the probability transform density is also
parametric:
̃ { ( )} ( ) ( ) ( )
(2.18)
There are different types of inverse probability transforms. The inverse complete probability
transform of an arbitrary function ( ) is (from Eq. 2.6):
{ ( )} {∫ ( ) ( ) } ( )
(2.19)
The inverse probability transform density ̃ of an arbitrary function ( ) is (from Eq. 2.18):
( )
̃ { ( )} ( )
( )
(2.21)
Thus, we may conclude that:
( ) ( )
( ) { ( )}
( ) ( )
(2.22)
In this Section, four different examples of probability distributions of randomistic variables are
considered. The simplest case considered is the pure deterministic variable, which will be
discussed first. Then, the most common models of probability distribution will be presented:
Uniform, Exponential, and Normal distribution models. In all cases, parametric versions of the
models are used, thus representing whole families of randomistic variables.
Let us begin the examples of probability transforms with the deterministic transform. In this
case, the randomistic variable can be expressed in terms of the standard deterministic variable
( ) [3] as follows:
(3.1)
where is a constant parameter, and
( ) ( ) {
(3.2)
where is Dirac’s delta function.
Using the change of variable theorem [4] and Eq. (3.1) and (3.2), it can be shown that [3]:
( ) ( )
(3.3)
( )
̃ { ( )} ( ) ( ) {
(3.4)
{ ( )} ∫ ̃ { ( )} {
( )
(3.5)
{ ( )} { ( )} ( )
(3.6)
Thus, the complete deterministic transform of a function is simply the value of the function
evaluated at the constant parameter . Also, for a given function of the constant parameter,
the inverse transform is straightforward:
{ ( )} ( )
(3.8)
In second place, let us consider the transforms obtained when the randomistic variable follows
a uniform probability distribution:
( )
(3.9)
where and are constant parameters representing the lower and upper limits of the
distribution respectively, and is a type III (bounded) standard uniform variable [5], with
the following probability density function:
( ) {
(3.10)
Using the change of variable theorem [4] the probability density function of becomes:
( ) {
(3.11)
( )
̃ { ( )} {
(3.12)
∫ ( )
{ ( )} ∫ ̃ { ( )}
∫ ( )
{
(3.13)
{ ( )} { ( )} ∫ ( )
(3.14)
and particularly for the -th power:
{ }
( )( )
(3.15)
(3.16)
( ) {
(3.17)
Using the change of variable theorem [4] the probability density function of becomes:
( ) { ( )
(3.18)
̃ { ( )} { ( )
( )
(3.19)
{ ( )} ∫ ̃ { ( )} { ( )
∫ ( )
(3.20)
{ ( )} { ( )} ( ) ( )
∫
(3.21)
and particularly for the -th power:
( ) ( )
{ } ∫
(3.22)
where represents the incomplete gamma function.
Notice that the integral in Eq. (3.21) can be related to the definition of the Laplace transform
[6]:
{ ( )} ∫ ( )
(3.23)
by means of the following expression:
{ ( ) }
{ ( )}
(3.24)
As a final example, let us consider the normal randomistic variable given by:
(3.25)
where is a type I standard normal variable with probability density function:
( )
( )
( )
√ √
(3.26)
where represents the normal bell function [7].
Using the change of variable theorem [4] the probability density function of becomes:
( ) ( )
( )
√ √
(3.27)
( ) ( )
̃ { ( )}
√
(3.28)
{ ( )} ∫ ̃ { ( )} ∫ ( ) ( )
√
(3.29)
{ ( )} { ( )} ∫ ( ) ( )
√
(3.30)
and particularly for the -th power:
{ } ∫ ( )
√
(3.31)
Unfortunately, some integrals found in the previous examples do not have a direct analytical
solution, increasing the complexity of probability transforms. In part, such complexity is
introduced by the distribution model parameters. So, the goal of this section is representing
probability transforms in terms of transforms of standard random variables, or standard
probability transforms, for simplicity.
First of all, let us notice the additive property of the probability transforms defined in Section 2.
Assuming that the function ( ) results by the linear combination of two or more functions we
obtain:
( ) ∑ ( )
(4.1)
Using Eq.(2.13) and (4.1), the probability transform density of the function becomes:
̃ { ( )} ( ) ( ) ∑ ( ) ( ) ∑ ̃ { ( )}
(4.2)
Proceeding similarly, the incomplete and complete probability transforms become:
{ ( )} ∑ { ( )}
(4.3)
{ ( )} ∑ { ( )}
(4.4)
Let us now assume that any randomistic variable can be represented in general as a linear
function of the standard deterministic and random variables as follows [3]:
(4.5)
By using the change of variable theorem [4], the probability transform density of can be
expressed as:
̃ { } ( ) ( ) ( )| | ( ) ( ) ̃ { }
| | | |
(4.6)
In general, if Eq. (4.5) remains valid, and the function ( ) can be expressed in terms of the
standard random variable as follows:
( ) ( )
(4.7)
( )
̃ { ( )} ̃ { }
| |
(4.8)
Now, since any arbitrary function of expressed as an infinite polynomial (McLaurin series)
simply becomes (from Eq. 2.4):
( )( ) ( )( ) ( )( )
( ) ∑ ∑ ( ) ∑ ∑( )
(4.9)
Then, the probability transform density of this function can be expressed as follows:
( )(
̃ { ( )} ∑ )∑ ̃ { }
( ) | |
(4.10)
Integrating Eq. (4.8) in order to obtain the incomplete transform results in:
( ) ̃ { ( )}
{ ( )} ∫ ̃ { ( )} ∫ ̃ { } ∫ | | { ( )}
| | | |
(4.11)
Therefore:
{ ( )} ( )( )∑
∑ { }
( )
(4.12)
{ ( )} ( )( )∑
∑ { }
( )
(4.13)
indicating that the probability transforms of a function can ultimately be expressed in terms of
the probability transform of naturals powers of standard random variables.
Extending the definitions given in Section 2, the standard probability transform density for
powers of standard random variables is:
̃ { } ( )
(4.14)
while the incomplete and complete probability transforms are:
{ } ∫ ( )
(4.15)
{ } ∫ ( )
(4.16)
The terms given in Eq. (4.15) are the raw moments of the standard random variable which can
be more easily generalized than parametric distribution models.
For example, for the standard uniform, exponential and normal random variables we have [8]:
( )
(4.17)
( )
(4.18)
( ) {
∏| |
(4.19)
The term ∏ | | can be alternatively expressed as follows:
∏| |
( )
(4.20)
Therefore, the expected value of an arbitrary function ( ) becomes for the uniform
distribution ( , ):
( )(
( ) ( )(
( )
{ ( )} ∑ )∑ ∑ )∑
( ) ( ) ( ) ( )
(4.21)
{ ( )} ( )( )∑ ( )( )∑
∑ ∑
( ) ( )
(4.22)
⌊ ⌋ ⌊ ⌋
( )( ( )(
{ ( )} ∑ )∑ ∏| | ∑ )∑
( ) ( ) ( )
(4.23)
Eq. (4.13) represents the most general expression for the complete probability transform of
biparametric randomistic variables. In the case of uniparametric variables we have two cases:
For a purely deterministic variable ( ) we obtain:
( )( )
{ ( )} ∑ ( )
(4.24)
For a purely random variable ( ), the complete transform becomes:
( )( )
{ ( )} ∑ { }
(4.25)
And particularly, for the probability models considered we obtain:
( )( )
{ ( ) } ∫ ( ) ∑
( )
(4.26)
( )( )
{ ( ) } ∫ ( ) ∑
(4.27)
( ) ( )( )
{ ( ) } ∫ ( ) ∑
√
(4.28)
In this Section, some representative functions are considered to illustrate the effect of
probability transformation using typical distribution models. For transform density (Figure 1 to
Figure 3) and incomplete transforms (Figure 4 to Figure 6), only the graphical representation of
standard random variables are considered for simplicity. Similarly, in the case of complete
transforms, uniparametric pure random variables are considered in Figure 7 to Figure 9.
Complete transforms for biparametric randomistic variables were only considered for the first
powers of the variable (Table 1), showing the increasing analytical complexity of the
transformation as non-linearity increases.
The following uniparametric complete transforms assume that the deterministic part of the
randomistic variable is exactly zero. That is, that the randomistic variable is a purely random
variable. The parameter of the random term is usually considered to be only positive. While
negative values are mathematically possible, they are conceptually erroneous, and will not be
considered.
In the case of biparametric distributions, the analytical complexity increases. For that reason,
only the first power functions are considered. Let us recall that a general non-linear function
can ultimately be represented by a polynomial function, and therefore its transform can be
expressed as a linear combination of individual power transforms.
( )
( )
( )
( )
( )
6. Final Remark
Apart from the well-known applications of the Laplace transform, probability transforms in
general are valuable tools for understanding uncertainty/noise propagation in a wide variety of
systems, including both static and dynamic systems.
This report provides data, information and conclusions obtained by the author(s) as a result of original
scientific research, based on the best scientific knowledge available to the author(s). The main purpose
of this publication is the open sharing of scientific knowledge. Any mistake, omission, error or inaccuracy
published, if any, is completely unintentional.
This research did not receive any specific grant from funding agencies in the public, commercial, or not-
for-profit sectors.
This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC
4.0). Anyone is free to share (copy and redistribute the material in any medium or format) or adapt
(remix, transform, and build upon the material) this work under the following terms:
Attribution: Appropriate credit must be given, providing a link to the license, and indicating if
changes were made. This can be done in any reasonable manner, but not in any way that
suggests endorsement by the licensor.
NonCommercial: This material may not be used for commercial purposes.
References
[1] Hernandez, H. (2018). The Realm of Randomistic Variables. ForsChem Research Reports, 3, 2018-10,
1-19. doi: 10.13140/RG.2.2.29034.16326.
[2] Hernandez, H. (2020). On the Discreteness of Measured Variables and the Continuous
Approximation. ForsChem Research Reports, 5, 2020-20, 1-18. doi: 10.13140/RG.2.2.27740.00646.
[3] Hernandez, H. (2022). Standard Deterministic, Standard Random, and Randomistic Variables.
ForsChem Research Reports, 7, 2022-06, 1 - 18. doi: 10.13140/RG.2.2.36316.87688.
[4] Hernandez, H. (2017). Multivariate Probability Theory: Determination of Probability Density
Functions. ForsChem Research Reports, 2, 2017-13, 1-13. doi: 10.13140/RG.2.2.28214.60481.
[5] Hernandez, H. (2018). Multidimensional Randomness, Standard Random Variables and Variance
Algebra. ForsChem Research Reports, 3, 2018-02, 1-35. doi: 10.13140/RG.2.2.11902.48966.
[6] Schiff, J. L. (1999). The Laplace Transform: Theory and Applications. Springer-Verlag, New York. doi:
10.1007/978-0-387-22757-3.
[7] Hernandez, H. (2022). Normal Distribution and Transcendental Functions: Mathematical and
Historical Relations. ForsChem Research Reports, 7, 2022-04, 1 - 47. doi: 10.13140/RG.2.2.21535.23203.
[8] Hernandez, H. (2018). Expected Value, Variance and Covariance of Natural Powers of
Representative Standard Random Variables. ForsChem Research Reports, 3, 2018-08, 1-19. doi:
10.13140/RG.2.2.15187.07205.