You are on page 1of 12

THE GAUSSIAN INTEGRAL

KEITH CONRAD

Let Z ∞ Z ∞ Z ∞
− 21 x2 −x2 2
I= e dx, and K =
dx, J = e e−πx dx.
−∞ 0 −∞
√ √
These numbers are positive, and J = I/(2 2) and K = I/ 2π.
√ √
Theorem. With notation as above, I = 2π, or equivalently J = π/2, or equivalently K = 1.

We will give multiple proofs of this result. (Other lists of proofs are in [3] and [8].) The theorem
1 2 2 2
is
Z ∞subtle because there is no simple antiderivative for e− 2 x (or e−x or e−πx ). For comparison,
1 2 1 2
xe− 2 x dx can be computed using the antiderivative −e− 2 x : this integral is 1.
0

1. First Proof: Polar coordinates


The most widely known proof uses multivariable calculus: express J 2 as a double integral and
then pass to polar coordinates:
Z ∞ Z ∞ Z ∞Z ∞
2 2 2 2
J2 = e−x dx e−y dy = e−(x +y ) dx dy.
0 0 0 0
This is a double integral over the first quadrant, which we will compute by using polar coordinates.
The first quadrant is {(r, θ) : r ≥ 0 and 0 ≤ θ ≤ π/2}. Writing x2 + y 2 as r2 and dx dy as r dr dθ,
Z π/2 Z ∞
2
2
J = e−r r dr dθ
0 0
Z ∞ Z π/2
−r2
= re dr · dθ
0 0
1 −r2 ∞ π

= − e ·2
2 0
1 π
= ·
2 2
π
= .
4

Taking square roots, J = π/2. This method is due to Poisson [8, p. 3].

2. Second Proof: Another change of variables


Our next proof uses another change of variables to compute J 2 , but this will only rely on single-
variable calculus. As before, we have
Z ∞Z ∞ Z ∞ Z ∞ 
2 −(x2 +y 2 ) −(x2 +y 2 )
J = e dx dy = e dx dy,
0 0 0 0
1
2 KEITH CONRAD

but instead of using polar coordinates we make a change of variables x = yt on the inside integral,
with dx = y dt, so
Z ∞ Z ∞  Z ∞ Z ∞ 
2 −y 2 (t2 +1) −y 2 (t2 +1)
J = e y dt dy = ye dy dt.
0 0 0 0
Z ∞
2 1
Since ye−ay dy = for a > 0, we have
0 2a
Z ∞
2 dt 1 π π
J = 2
= · = ,
0 2(t + 1) 2 2 4

so J = π/2. This approach is due to Laplace [6, pp. 94–96] and historically precedes the more
familiar technique in the first proof above. We will see in the eighth proof that this was not
Laplace’s first method.

3. Third Proof: Differentiating under the integral sign


For t > 0, set
Z t 2
−x2
A(t) = e dx .
0
The integral we want to calculate is A(∞) = 2
J and
then take a square root.
Differentiating A(t) with respect to t and using the Fundamental Theorem of Calculus,
Z t Z t
0 −x2 −t2 −t2 2
A (t) = 2 e dx · e = 2e e−x dx.
0 0
Let x = ty, so
Z 1 Z 1
2 2 y2 2 )t2
A0 (t) = 2e−t te−t dy = 2te−(1+y dy.
0 0
The function under the integral sign is easily antidifferentiated with respect to t:
2 2 2 2
∂ e−(1+y )t d 1 e−(1+y )t
Z 1 Z
0
A (t) = − dy = − dy.
0 ∂t 1 + y 2 dt 0 1 + y2
Letting
2 2
e−t (1+x ) 1
Z
B(t) = dx,
0 1 + x2
we have A0 (t) = −B 0 (t) for all t > 0, so there is a constant C such that
(3.1) A(t) = −B(t) + C
Z 0 2
−x2
for all t > 0. To find C, we let t → 0+
in (3.1). The left side tends to e dx = 0 while
Z 1 0

the right side tends to − dx/(1 + x2 ) + C = −π/4 + C. Thus C = π/4, so (3.1) becomes
0
2 2 2
t
e−t (1+x ) 1
Z Z
−x2 π
e dx = − dx.
0 0 1 + x2
4

Letting t → ∞ in this equation, we obtain J 2 = π/4, so J = π/2.
A comparison of this proof with the first proof is in [19].
THE GAUSSIAN INTEGRAL 3

4. Fourth Proof: Another differentiation under the integral sign


Here is a second approach to finding J by differentiation under the integral sign. I heard about
it from Michael Rozman [13], who modified an idea on math.stackexchange [21], and in a slightly
less elegant form it appeared much earlier in [17].
For t ∈ R, set
Z ∞ −t2 (1+x2 )
e
F (t) = dx.
0 1 + x2
R∞
Then F (0) = 0 dx/(1 + x2 ) = π/2 and F (∞) = 0. Differentiating under the integral sign,
Z ∞ Z ∞
0 −t2 (1+x2 ) −t2 2
F (t) = −2te dx = −2te e−(tx) dx.
0 0

Make the substitution y = tx, with dy = t dx, so


Z ∞
0 −t2 2 2
F (t) = −2e e−y dy = −2Je−t .
0

For b > 0, integrate both sides from 0 to b and use the Fundamental Theorem of Calculus:
Z b Z b Z b
0 −t2 2
F (t) dt = −2J e dt =⇒ F (b) − F (0) = −2J e−t dt.
0 0 0

Letting b → ∞ in the last equation,



π 2 2 π π
0 − = −2J =⇒ J = =⇒ J = .
2 4 2

5. Fifth Proof: A volume integral


Our next proof is due to T. P. Jameson [4] and it was rediscovered by A. L. Delgado [2]. Revolve
1 2 1 2 2
the curve z = e− 2 x in the xz-plane around the z-axis to produce the “bell surface” z = e− 2 (x +y ) .
See below, where the z-axis is vertical and passes through the top point, the x-axis lies just under
the surface through the point 0 in front, and the y-axis lies just under the surface through the
point 0 on the left. We will compute the volume V below the surface and above the xy-plane in
two ways. Z 1
First we compute V by horizontal slices, which are discs: V = A(z) dz where A(z) is the area
0
of the disc formed by slicing the surface at height z. Writing the radius of the disc at height z as
1 2
r(z), A(z) = πr(z)2 . To compute r(z), the surface cuts the xz-plane at a pair of points (x, e− 2 x )
1 2
where the height is z, so e− 2 x = z. Thus x2 = −2 ln z. Since x is the distance of these points from
the z-axis, r(z)2 = x2 = −2 ln z, so A(z) = πr(z)2 = −2πln z. Therefore
Z 1 1

V = −2π ln z dz = −2π (z ln z − z) = −2π(−1 − lim z ln z).
0 0 z→0+

By L’Hospital’s rule, limz→0+ z ln z = 0, so V = 2π. (A calculation of V by shells is in [10].)


Next we compute the volume by vertical slices in planes x = constant. Vertical slices are scaled
bell curves: look at the black contour lines in the picture. The equation of the bell curve along the
1 2 2
top of the vertical slice with x-coordinate x is z = e− 2 (x +y ) , where y varies and x is fixed. Then
4 KEITH CONRAD

Z ∞
V = A(x) dx, where A(x) is the area of the x-slice:
−∞
Z ∞ 1 1 2
Z ∞ 1 2 1 2
2 +y 2 )
A(x) = e− 2 (x dy = e− 2 x e− 2 y dy = e− 2 x I.
−∞ −∞
Z ∞ Z ∞ 1 2
Z ∞ 1 2
Thus V = A(x) dx = e− 2 x I dx = I
e− 2 x dx = I 2 .
−∞ −∞ −∞ √
Comparing the two formulas for V , we have 2π = I 2 , so I = 2π.

6. Sixth Proof: The Γ-function


Z ∞
For any integer n ≥ 0, we have n! = tn e−t dt. For x > 0 we define
0
Z ∞
dt
Γ(x) = tx e−t ,
0 t
so Γ(n) = (n − 1)! when n ≥ 1. Using integration by parts, Γ(x + 1) = xΓ(x). One of the basic
properties of the Γ-function [14, pp. 193–194] is
Z 1
Γ(x)Γ(y)
(6.1) = tx−1 (1 − t)y−1 dt.
Γ(x + y) 0
THE GAUSSIAN INTEGRAL 5

Set x = y = 1/2:
 2 Z 1
1 dt
Γ = p .
2 0 t(1 − t)
Note
  Z ∞
√ −t dt Z ∞ e−t Z ∞ −x2 Z ∞
1 e 2
Γ = te = √ dt = 2x dx = 2 e−x dx = 2J,
2 0 t 0 t 0 x 0
2
R1 p 2
so 4J = 0 dt/ t(1 − t). With the substitution t = sin θ,
Z π/2
2 2 sin θ cos θ dθ π
4J = = 2 = π,
0 sin θ cos θ 2
√ √ √
so J = π/2. Equivalently,
Z ∞ Γ(1/2) = π. Any method that proves Γ(1/2) = π is also a method
2
that calculates e−x dx.
0

7. Seventh Proof: Asymptotic estimates



We will show J = π/2 by a technique whose steps are based on [15, p. 371].
For x ≥ 0, power series expansions show 1 + x ≤ ex ≤ 1/(1 − x). Reciprocating and replacing x
with x2 , we get
2 1
(7.1) 1 − x2 ≤ e−x ≤ .
1 + x2
for all x ∈ R.
For any positive integer n, raise the terms in (7.1) to the nth power and integrate from 0 to 1:
Z 1 Z 1 Z 1
2 dx
(1 − x2 )n dx ≤ e−nx dx ≤ .
0 0 0 (1 + x2 )n

Under the changes of variables x = sin θ on the left, x = y/ n in the middle, and x = tan θ on the
right,
Z π/2 Z √n Z π/4
2n+1 1 −y 2
(7.2) (cos θ) dθ ≤ √ e dy ≤ (cos θ)2n−2 dθ.
0 n 0 0
R π/2
Set Ik = 0 (cos θ)k dθ, so I0 = π/2, I1 = 1, and (7.2) implies
Z √n
√ 2 √
(7.3) nI2n+1 ≤ e−y dy ≤ nI2n−2 .
0

We will show that as k → ∞, kIk2 → π/2. Then


√ √
√ n √
r
1 π π
nI2n+1 = √ 2n + 1I2n+1 → √ =
2n + 1 2 2 2
and √ √
√ n √
r
1 π π
nI2n−2 = √ 2n − 2I2n−2 → √ = ,
2n − 2 2 2 2

Z n √ √
2
so by (7.3) e−y dy → π/2. Thus J = π/2.
0
6 KEITH CONRAD

To show kIk2 → π/2, first we compute several values of Ik explicitly by a recursion. Using
integration by parts,

Z π/2 Z π/2
k
Ik = (cos θ) dθ = (cos θ)k−1 cos θ dθ = (k − 1)(Ik−2 − Ik ),
0 0

so

k−1
(7.4) Ik = Ik−2 .
k

Using (7.4) and the initial values I0 = π/2 and I1 = 1, the first few values of Ik are computed and
listed in Table 1.

k Ik k Ik
0 π/2 1 1
2 (1/2)(π/2) 3 2/3
4 (3/8)(π/2) 5 8/15
6 (15/48)(π/2) 7 48/105
Table 1.

From Table 1 we see that

1 π
(7.5) I2n I2n+1 =
2n + 1 2

for 0 ≤ n ≤ 3, and this can be proved for all n by induction using (7.4). Since 0 ≤ cos θ ≤ 1 for
k
θ ∈ [0, π/2], we have Ik ≤ Ik−1 ≤ Ik−2 = k−1 Ik by (7.4), so Ik−1 ∼ Ik as k → ∞. Therefore (7.5)
implies

2 1 π 2 π
I2n ∼ =⇒ (2n)I2n →
2n 2 2

as n → ∞. Then
2 2 π
(2n + 1)I2n+1 ∼ (2n)I2n →
2

as n → ∞, so kIk2 → π/2 as k → ∞. This completes our proof that J = π/2.

Remark 7.1. This proof is closely related to the fifth proof using the Γ-function. Indeed, by (6.1)

Γ( k+1 1
2 )Γ( 2 )
1
Z
= t(k+1)/2+1 (1 − t)1/2−1 dt,
Γ( k+1
2 + 1
2) 0
THE GAUSSIAN INTEGRAL 7

and with the change of variables t = (cos θ)2 for 0 ≤ θ ≤ π/2, the integral on the right is equal to
R π/2
2 0 (cos θ)k dθ = 2Ik , so (7.5) is the same as
π
= I2n I2n+1
2(2n + 1)
Γ( 2n+1 1 2n+2 1
2 )Γ( 2 ) Γ( 2 )Γ( 2 )
=
2Γ( 2n+2
2 ) 2Γ( 2n+3
2 )
Γ( 2n+1 1 2
2 )Γ( 2 )
=
4Γ( 2n+1
2 + 1)
Γ( 2n+1 1 2
2 )Γ( 2 )
=
4 2n+1 2n+1
2 Γ( 2 )
Γ( 12 )2
= ,
2(2n + 1)
√ √
or equivalently Γ(1/2)2 = π. We saw in the fifth proof that Γ(1/2) = π if and only if J = π/2.

8. Eighth Proof: The original proof



The original proof that J = π/2 is due to Laplace [7] in 1774. (An English translation of
Laplace’s article is mentioned in the bibliographic citation for [7], with preliminary comments on
that article in [16].) He wanted to compute
Z 1
dx
(8.1) √ .
0 − log x
√ R∞ 2 √
Setting y = − log x, this integral is 2 0 e−y dy = 2J, so we expect (8.1) to be π.
Laplace’s starting point for evaluating (8.1) was a formula of Euler:
Z 1 Z 1 s+r
xr dx x dx 1 π
(8.2) √ √ =
0 1−x 2s
0 1−x 2s s(r + 1) 2
for positive r and s. (Laplace himself said this formula held “whatever be” r or s, but if s < 0 then
the number under the square root is negative.) Accepting (8.2), let r → 0 in it to get
Z 1 Z 1
dx xs dx 1π
(8.3) √ √ = .
0 1 − x2s 0 1 − x2s s2
Now let s → 0 in (8.3). Then 1 − x2s ∼ −2s log x by L’Hopital’s rule, so (8.3) becomes
Z 1 2
dx
√ = π.
0 − log x

Thus (8.1) is π.
Euler’s formula (8.2) looks mysterious, but we have met it before. In the formula let xs = cos θ
with 0 ≤ θ ≤ π/2. Then x = (cos θ)1/s , and after some calculations (8.2) turns into
Z π/2 Z π/2
(r+1)/s−1 1 π
(8.4) (cos θ) dθ (cos θ)(r+1)/s dθ = .
0 0 (r + 1)/s 2
R π/2
We used the integral Ik = 0 (cos θ)k dθ before when k is a nonnegative integer. This notation
1 π
makes sense when k is any positive real number, and then (8.4) assumes the form Iα Iα+1 = α+1 2 for
α = (r +1)/s − 1, which is (7.5) with a possibly nonintegral index. Letting r = 0 and s = 1/(2n + 1)
8 KEITH CONRAD

in (8.4) recovers (7.5). Letting s → 0 in (8.3) corresponds to letting n → ∞ in (7.5), so the 6th
proof is in essence a more detailed version of Laplace’s 1774 argument.

9. Ninth Proof: Contour Integration


Z ∞
2 /2
We will calculate e−x dx using contour integrals and the residue theorem. However, we
−∞
2
can’t just integrate e−z /2 ,
as this function has no poles. For a long time nobody knew how to
handle this integral using contour integration. For instance, in 1914 WatsonZ [18, p. 79] wrote∞
2
“Cauchy’s theorem cannot be employed to evaluate all definite integrals; thus e−x dx has not
0
been evaluated except by other methods.” In the 1940s several contour integral solutions were
published using awkward contours such as parallelograms [9], [11, Sect. 5] (see [1, Exer. 9, p. 113]
for a recent appearance). Our approach will follow Kneser [5, p. 121] (see also [12, pp. 413–414] or
[20]), using a rectangular contour and the function
2
e−z /2
√ .
1 − e− π(1+i)z
This function comes out of nowhere, so our first task is to motivate the introduction of this function.
We seek a meromorphic function f (z) to integrate around the rectangular contour γR in the
figure below, with vertices at −R, R, R + ib, and −R + ib, where b will be fixed and we let R → ∞.

Suppose f (z) → 0 along the right and left sides of γR uniformly as R → ∞. Then by applying
the residue theorem and letting R → ∞, we would obtain (if the integrals converge)
Z ∞ Z −∞ X
f (x) dx + f (x + ib) dx = 2πi Resz=a f (z),
−∞ ∞ a

where the sum is over poles of f (z) with imaginary part between 0 and b. This is equivalent to
Z ∞ X
(f (x) − f (x + ib)) dx = 2πi Resz=a f (z).
−∞ a

Therefore we want f (z) to satisfy


2 /2
(9.1) f (z) − f (z + ib) = e−z ,
where f (z) and b need to be determined.
2
Let’s try f (z) = e−z /2 /d(z), for an unknown denominator d(z) whose zeros are poles of f (z).
We want f (z) to satisfy
2 /2
(9.2) f (z) − f (z + τ ) = e−z
THE GAUSSIAN INTEGRAL 9

for some τ (which will not be purely imaginary, so (9.1) doesn’t quite work, but (9.1) is only
2
motivation). Substituting e−z /2 /d(z) for f (z) in (9.2) gives us
2
!
−z 2 /2 1 e−τ z−τ /2 2
(9.3) e − = e−z /2 .
d(z) d(z + τ )

Suppose d(z + τ ) = d(z). Then (9.3) implies


2 /2
d(z) = 1 − e−τ z−τ ,
2
and with this definition of d(z), f (z) satisfies (9.2) if and only if eτ = 1, or equivalently τ 2 ∈ 2πiZ.
√ 2
The simplest nonzero solution is τ = π(1 + i). From now on this is the value of τ , so e−τ /2 =
e−iπ = −1 and then
2 2
e−z /2 e−z /2
f (z) = = ,
d(z) 1 + e−τ z
which is Kneser’s function mentioned earlier. This function satisfies (9.2) and we henceforth ignore
the motivation (9.1). Poles of f (z) are at odd integral multiples of τ /2.
We will integrate this f (z) around the rectangular contour γR below, whose height is Im(τ ).

The poles of f (z) nearest the origin


√ are plotted in the figure; they lie along the line y = x. The
only pole of f (z) inside γR (for R > π/2) is at τ /2, so by the residue theorem
2 2
Z
e−τ /8 2πie3τ /8 √
f (z) dz = 2πiResz=τ /2 f (z) = 2πi −τ 2 /2 = √ = 2π.
γR (−τ )e − π(1 + i)

As R → ∞, the value of |f (z)| tends to 0 uniformly along the left and right sides of γR , so

√ Z ∞ Z −∞+i π
2π = f (x) dx + √
f (z) dz
−∞ ∞+i π
Z ∞ Z ∞

= f (x) dx − f (x + i π) dx.
−∞ −∞
10 KEITH CONRAD


In the second integral, write i π as τ − π and use (real) translation invariance of dx to obtain
√ Z ∞ Z ∞
2π = f (x) dx − f (x + τ ) dx
−∞ −∞
Z ∞
= (f (x) − f (x + τ )) dx
−∞
Z ∞
2
= e−x /2 dx by (9.2).
−∞

10. Tenth Proof: Stirling’s Formula


Z ∞
1 2 √
Besides the integral formula e− 2 x dx = 2π that we have been discussing, another place
√ −∞
in mathematics where 2π appears is in Stirling’s formula:
nn √
n! ∼ n 2πn as n → ∞.
e
n n √
In 1730 De Moivre proved n! ∼ C(n /e ) n for some √ positive number C without being able to
determine C. Stirling soon thereafter showed C = 2π and wound up having the whole formula √
named after him. We will show that√ determining that the constant√C in Stirling’s formula is 2π
is equivalent to showing that J = π/2 (or, equivalently, that I = 2π).
Applying (7.4) repeatedly,
2n − 1
I2n = I2n−2
2n
(2n − 1)(2n − 3)
= I2n−4
(2n)(2n − 2)
..
.
(2n − 1)(2n − 3)(2n − 5) · · · (5)(3)(1)
= I0 .
(2n)(2n − 2)(2n − 4) · · · (6)(4)(2)
Inserting (2n − 2)(2n − 4)(2n − 6) · · · (6)(4)(2) in the top and bottom,
(2n − 1)(2n − 2)(2n − 3)(2n − 4)(2n − 5) · · · (6)(5)(4)(3)(2)(1) π (2n − 1)! π
I2n = 2
= n−1 2
.
(2n)((2n − 2)(2n − 4) · · · (6)(4)(2)) 2 2n(2 (n − 1)!) 2

Applying De Moivre’s asymptotic formula n! ∼ C(n/e)n n, ,
√ 1

C((2n − 1)/e)2n−1 2n − 1 π (2n − 1)2n 2n−1 2n − 1 π
I2n ∼ n−1 n−1
√ 2
= 2(n−1) 2n 1
2n(2 C((n − 1)/e) n − 1) 2 2n · 2 Ce(n − 1) (n−1)2 (n − 1) 2
as n → ∞. For any a ∈ R, (1 + a/n)n → ea as n → ∞, so (n + a)n ∼ ea nn . Substituting this into
the above formula with a = −1 and n replaced by 2n,
e−1 (2n)2n √12n π π
(10.1) I2n ∼ = √ .
2n · 22(n−1) Ce(e−1 nn )2 n12 n
2 C 2n
√ √
Since Ik−1 ∼ Ik , the outer terms in (7.3) are both asymptotic to nI2n ∼ π/(C 2) by (10.1).
Therefore √
Z n
2 π
e−y dy → √
0 C 2
THE GAUSSIAN INTEGRAL 11
√ √ √
as n → ∞, so J = π/(C 2). Therefore C = 2π if and only if J = π/2.

11. Eleventh Proof: Fourier transforms


For a continuous function f : R → C that is rapidly decreasing at ±∞, its Fourier transform is
the function Ff : R → C defined by
Z ∞
(Ff )(y) = f (x)e−ixy dx.
−∞
R∞
For example, (Ff )(0) = −∞ f (x) dx.
Here are three properties of the Fourier transform.
• If f is differentiable, then after using differentiation under the integral sign on the Fourier
transform of f we obtain
Z ∞
(Ff )0 (y) = −ixf (x)e−ixy dx = −i(F(xf (x)))(y).
−∞

• Using integration by parts on the Fourier transform of f , with u = f (x) and dv = e−ixy dx,
we obtain
F(f 0 )(y) = iy(Ff )(y).
• If we apply the Fourier transform twice then we recover the original function up to interior
and exterior scaling:

(11.1) (F 2 f )(x) = 2πf (−x).



Let’s show the appearance of 2π in (11.1) is equivalent to the evaluation of I as 2π.
2
Fixing a > 0, set f (x) = e−ax , so

f 0 (x) = −2axf (x).

Applying the Fourier transform to both sides of this equation implies iy(Ff )(y) = −2a −i1
(Ff )0 (y),
0 1 0 1
which simplifies to (Ff ) (y) = − 2a y(Ff )(y). The general solution of g (y) = − 2a yg(y) is g(y) =
2
Ce−y /(4a) , so
2 /(4a)
(Ff )(y) = Ce−y
2 /2
for some constant C. Letting a = 12 , so f (x) = e−x , we obtain
2
(Ff )(y) = Ce−y /2 = Cf (y).
R∞ 2
Setting y = 0, the left side is (Ff )(0) = −∞ e−x /2 dx = I, so I = Cf (0) = C.
Applying the Fourier transform to both sides of the equation (Ff )(y)
√ = Cf (y), we get 2πf (−x) =
2 2
C(Ff√)(x) = C f (x). At x = 0 this becomes 2π = C , so I = C = ± 2π. Since I > 0, the number
I is 2π. If we didn’t know the constant on the right side of (11.1) were 2π, whatever its value is
would 2
√ wind up being C , so saying 2π appears on the right side of (11.1) is equivalent to saying
I = 2π.
12 KEITH CONRAD

References
[1] C. A. Berenstein and R. Gay, Complex
Z ∞ Variables, Springer-Verlag, New York, 1991.
2
[2] A. L. Delgado, “A Calculation of e−x dx,” The College Math. J. 34 (2003), 321–323.
0
[3] H. Iwasawa, “Gaussian Integral Puzzle,” Math. Intelligencer 31 (2009), 38–41.
[4] T. P. Jameson, “The Probability Integral by Volume of Revolution,” Mathematical Gazette 78 (1994), 339–340.
[5] H. Kneser, Funktionentheorie, Vandenhoeck and Ruprecht, 1958.
[6] P. S. Laplace, Théorie Analytique des Probabilités, Courcier, 1812.
[7] P. S. Laplace, “Mémoire sur la probabilité des causes par les évènemens,” Oeuvres Complétes 8, 27–65. (English
trans. by S. Stigler as “Memoir on the Probability of Causes of Events,” Statistical Science 1 (1986), 364–378.)
[8] P. M. Lee, http://www.york.ac.uk/depts/maths/histstat/normal_history.pdf.
[9] L. Mirsky, The Probability Integral, Math. Gazette 33 (1949), 279. Online at http://www.jstor.org/stable/
3611303.
[10] C. P. Nicholas and R. C. Yates, “The Probability Integral,” Amer. Math. Monthly 57 (1950), 412–413.
[11] G. Polya, “Remarks on Computing the Probability Integral in One and Two Dimensions,” pp. 63–78 in Berkeley
Symp. on Math. Statist. and Prob., Univ. California Press, 1949.
[12] R. Remmert, Theory of Complex Functions, Springer-Verlag, 1991.
[13] M. Rozman, “Evaluate Gaussian integral using differentiation under the integral sign,” Course notes for Physics
2400 (UConn), Spring 2016.
[14] W. Rudin, Principles of Mathematical Analysis, 3rd ed., McGraw-Hill, 1976.
[15] M. Spivak, Calculus, W. A. Benjamin, 1967.
[16] S. Stigler, “Laplace’s 1774 Memoir on Inverse Probability,” Statistical Science 1 (1986), 359–363.
[17] J. van Yzeren, “Moivre’s and Fresnel’s Integrals by Simple Integration,” Amer. Math. Monthly 86 (1979),
690–693.
[18] G. N. Watson, Complex Integration and Cauchy’s Theorem, Cambridge Univ. Press, Cambridge, 1914.
[19] http://gowers.wordpress.com/2007/10/04/when-are-two-proofs-essentially-the-same/#comment-239.
[20] http://math.stackexchange.com/questions/34767/int-infty-infty-e-x2-dx-with-complex-analysis.
[21] http://math.stackexchange.com/questions/390850/integrating-int-infty-0-e-x2-dx-using-feynmans-
parametrization-trick

You might also like