Gaussian Integral
Gaussian Integral
Gaussian Integral
KEITH CONRAD
Let Z ∞ Z ∞ Z ∞
− 21 x2 −x2 2
I= e e dx, J =
dx, and K = e−πx dx.
−∞ 0 −∞
√ √
These positive numbers are related: J = I/(2 2) and K = I/ 2π.
√ √
Theorem. With notation as above, I = 2π, or equivalently J = π/2, or equivalently K = 1.
We will give multiple proofs of this. (Other lists of proofs are in [4] and [9].) It is subtle since
1 2 R∞ 1 2
e− 2 x has no simple antiderivative. For comparison, 0 xe− 2 x dx can be computed with the
1 2
antiderivative −e− 2 x and equals 1.
1For a visualization of this calculation as a volume, in terms of R ∞ e−x2 dx instead of J, see https://www.
−∞
youtube.com/watch?v=cy8r7WSuT1I. We’ll do a volume calculation for I 2 in Section 5.
1
2 KEITH CONRAD
Instead of using polar coordinates, set x = yt in the inner integral (y is fixed). Then dx = y dt and
Z ∞ Z ∞ Z ∞ Z ∞
2 −y 2 (t2 +1) −y 2 (t2 +1)
(2.1) J = e y dt dy = ye dy dt,
0 0 0 0
where the interchange of integrals is justified by Fubini’s theorem for improper Riemann integrals.
(The appendix
Z ∞ gives an approach using Fubini’s theorem for Riemann integrals on rectangles.)
2 1
Since ye−ay dy = for a > 0, we have
0 2a
Z ∞
2 dt 1 π π
J = 2
= · = ,
0 2(t + 1) 2 2 4
√
so J = π/2. This proof is due to Laplace [7, pp. 94–96] and historically precedes the widely used
technique of the previous proof. We will see in Section 9 what Laplace’s first proof was.
the right side tends to − dx/(1 + x2 ) + C = −π/4 + C. Thus C = π/4, so (3.1) becomes
0
2 2 2
t
e−t (1+x ) 1
Z Z
−x2 π
e dx = − dx.
0 0 1 + x2
4
√
Letting t → ∞ in this equation, we obtain J 2 = π/4, so J = π/2.
A comparison of this proof with the first proof is in [20].
THE GAUSSIAN INTEGRAL 3
For b > 0, integrate both sides from 0 to b and use the Fundamental Theorem of Calculus:
Z b Z b Z b
0 −t2 2
F (t) dt = −2J e dt =⇒ F (b) − F (0) = −2J e−t dt.
0 0 0
Z ∞
V = A(x) dx, where A(x) is the area of the x-slice:
−∞
Z ∞ 1 1 2
Z ∞ 1 2 1 2
2 +y 2 )
A(x) = e− 2 (x dy = e− 2 x e− 2 y dy = e− 2 x I.
−∞ −∞
Z ∞ Z ∞ 1 2
Z ∞ 1 2
Thus V = A(x) dx = e− 2 x I dx = I
e− 2 x dx = I 2 .
−∞ −∞ −∞ √
Comparing the two formulas for V , we have 2π = I 2 , so I = 2π.
Set x = y = 1/2:
2 Z 1
1 dt
Γ = p .
2 0 t(1 − t)
Note
Z ∞
√ −t dt Z ∞ e−t Z ∞ −x2 Z ∞
1 e 2
Γ = te = √ dt = 2x dx = 2 e−x dx = 2J,
2 0 t 0 t 0 x 0
2
R1 p 2
so 4J = 0 dt/ t(1 − t). With the substitution t = sin θ,
Z π/2
2 2 sin θ cos θ dθ π
4J = = 2 = π,
0 sin θ cos θ 2
√ √ √
so J = π/2. Equivalently,
Z ∞ Γ(1/2) = π. Any method that proves Γ(1/2) = π is also a method
2
that calculates e−x dx.
0
To show kIk2 → π/2, first we compute several values of Ik explicitly by a recursion. Using
integration by parts,
Z π/2 Z π/2
k
Ik = (cos θ) dθ = (cos θ)k−1 cos θ dθ = (k − 1)(Ik−2 − Ik ),
0 0
so
k−1
(7.4) Ik = Ik−2 .
k
Using (7.4) and the initial values I0 = π/2 and I1 = 1, the first few values of Ik are computed and
listed in Table 1.
k Ik k Ik
0 π/2 1 1
2 (1/2)(π/2) 3 2/3
4 (3/8)(π/2) 5 8/15
6 (15/48)(π/2) 7 48/105
Table 1.
and with the change of variables t = (cos θ)2 for 0 ≤ θ ≤ π/2, the integral on the right is equal to
R π/2
2 0 (cos θ)k dθ = 2Ik , so (7.5) is the same as
Γ( 2n+1 1 2n+2 1
2 )Γ( 2 ) Γ( 2 )Γ( 2 )
I2n I2n+1 =
2Γ( 2n+2
2 ) 2Γ( 2n+3
2 )
Γ( 2n+1 1 2
2 )Γ( 2 )
=
4Γ( 2n+1
2 + 1)
Γ( 2n+1 1 2
2 )Γ( 2 )
=
4 2n+1 2n+1
2 Γ( 2 )
Γ( 12 )2
= .
2(2n + 1)
THE GAUSSIAN INTEGRAL 7
√ √
By (7.5), π = Γ(1/2)2 . We saw in the fifth proof that Γ(1/2) = π if and only if J = π/2.
This function comes out of nowhere, so our first task is to motivate the introduction of this function.
We seek a meromorphic function f (z) to integrate around the rectangular contour γR in the
figure below, with vertices at −R, R, R + ib, and −R + ib, where b will be fixed and we let R → ∞.
Suppose f (z) → 0 along the right and left sides of γR uniformly as R → ∞. Then by applying
the residue theorem and letting R → ∞, we would obtain (if the integrals converge)
Z ∞ Z −∞ X
f (x) dx + f (x + ib) dx = 2πi Resz=a f (z),
−∞ ∞ a
where the sum is over poles of f (z) with imaginary part between 0 and b. This is equivalent to
Z ∞ X
(f (x) − f (x + ib)) dx = 2πi Resz=a f (z).
−∞ a
• Using integration by parts on the Fourier transform of f , with u = f (x) and dv = e−ixy dx,
we obtain
(F(f 0 ))(y) = iy(Ff )(y).
• If we apply the Fourier transform twice then we recover the original function up to interior
and exterior scaling:
(11.2) (F 2 f )(x) = 2πf (−x).
The 2π is admittedly a nonobvious scaling factor here, and the proof of (11.2)
√ is nontrivial. We’ll
show the appearance of 2π in (11.2) is equivalent to the evaluation of I as 2π.
2
Fixing a > 0, set f (x) = e−ax , so
f 0 (x) = −2axf (x).
Applying the Fourier transform to both sides of this equation implies iy(Ff )(y) = −2a −i 1
(Ff )0 (y),
which simplifies to (Ff )0 (y) = − 2a
1
y(Ff )(y). The general solution of g 0 (y) = − 2a
1
yg(y) is g(y) =
−y 2 /(4a)
Ce , so
2 2
f (x) = e−ax =⇒ (Ff )(y) = Ce−y /(4a)
2 /2
for some constant C. We have 1/(4a) = a when a = 1/2, so set a = 1/2: if f (x) = e−x then
−y 2 /2
(11.3) (Ff )(y) = Ce = Cf (y).
R∞ 2
Setting y = 0 in (11.3), the left side is (Ff )(0) = −∞ e−x /2 dx = I, so I = Cf (0) = C.
Applying the Fourier transform to both sides of (11.3) with C = I √ and using (11.2), we get
2πf (−x) = I(Ff )(x) = I 2 f (x). At x = 0 this becomes 2π = I 2 , so I = 2π since I > 0. That is
the Gaussian integral calculation. If we didn’t know that the constant on the right side of (11.2) is
2π, whatever its value is would 2
√ wind up being I , so saying 2π appears on the right side of (11.2)
is equivalent to saying I = 2π.
There are other ways to define the Fourier transform besides (11.1), such as
Z ∞ Z ∞
1
√ f (x)e−ixy dx or f (x)e−2πixy dx.
2π −∞ −∞
These transforms have properties similar to the transform as defined in (11.1), so they can be used
in its place to compute the Gaussian integral. Let’s see how such a proof looks using the second
alternative definition, which we’ll write as
Z ∞
(Ff
e )(y) = f (x)e−2πixy dx.
−∞
For this Fourier transform, the analogue of the three properties above for F are
e )0 (y) = −2πi(F(xf
• (Ff e (x)))(y).
0
• (F(f
e ))(y) = 2πiy(Ff e )(y).
• (Fe2 f )(x) = f (−x).
The last property for Fe looks nicer than that for F, since there is no overall 2π-factor on the right
side (it has been hidden in the definition of F).
e On the other hand, the first two properties for Fe
have overall factors of 2π on the right side while the first two properties of F do not. You can’t
escape a role for π or 2π somewhere in every possible definition of a Fourier transform.
2
Now let’s run through the proof again with Fe in place of F. For a > 0, set f (x) = e−ax .
Applying Fe to both sides of the equation f 0 (x) = −2axf (x), 2πiy(Ff e )(y) = −2a 1 (Ff )0 (y),
−(2πi)
12 KEITH CONRAD
2 2
e )0 (y) = − 2π y(Ff )(y). Solutions of g 0 (y) = − 2π yg(y) all look like
and that is equivalent to (Ff a a
2 2
Ce−(π /a)y , so
2
f (x) = e−ax =⇒ (Ff e )(y) = Ce−(π2 /a)y2
2 /a)y 2 2
for a constant C. We want π 2 /a = π so that e−(π = e−πy = f (y), which occurs for a = π.
2
Thus when f (x) = e−πx we have
2
(11.4) e )(y) = Ce−πy = Cf (y).
(Ff
R∞ 2
When y = 0 in (11.4), this becomes −∞ e−πx dx = C, so C = K: see the top of the first page for
2
the definition of K as the integral of e−πx over R.
Applying Fe to both sides of (11.4) with C = K and using (Fe2 f )(x) = f (−x), we get f (−x) =
K(Ffe )(x) = K 2 f (x). At x = 0 this becomes 1 = K 2 , so K = 1 since K > 0. That K = 1, or
R∞ 2
in more explicit form −∞ e−πx dx = 1, is equivalent to the evaluation of the Gaussian integral I
√
with the change of variables y = 2πx in the integral for K.
above by 1/x2 .
Now we prove Theorem A.1.
THE GAUSSIAN INTEGRAL 13
Proof. Step 1. For b > 1 and c > 1, we’ll show the improper integral can be truncated to an integral
over [0, b] × [0, c] plus error terms:
Z ∞ Z ∞ Z b Z c
−(t2 +1)y 2 −(t2 +1)y 2 1 1
ye dt dy = ye dt dy + O √ +O .
0 0 0 0 c b
R∞
Subtract the integral on the right from the integral on the left and split the outer integral 0
Rb R∞
into 0 + b :
Z ∞ Z ∞ Z b Z c Z b Z ∞
2 2 2 2 2 2
ye−(t +1)y dt dy − ye−(t +1)y dt dy = ye−(t +1)y dt dy
0 0 0 0 0 c
Z ∞ Z ∞
−(t2 +1)y 2
+ ye dt dy.
b 0
√
On the right side, we will show the first iterated integral is O(1/ c) and the second iterated integral
is O(1/b). The second iterated integral is simpler:
Z ∞ Z ∞ Z ∞ Z ∞
−(t2 +1)y 2 −(yt)2 2
ye dt dy = e dt ye−y dy
b 0 b 0
Z ∞ Z ∞
dt 2
≤ 2 t2 + 1
ye−y dy by Lemma A.2(1)
y
Zb ∞ 0
π 2
= ye−y dy by Lemma A.2(2)
b 2y
π ∞ −y2
Z
= e dy
2 b
π ∞ dy
Z
≤ by Lemma A.2(1)
2 b y2 + 1
π 1 1
= since 2 < 2,
2b y +1 y
so
Z c Z b
1 c dt 1 c
Z Z
2 2 dt
ye−(t +1)y dy dt = 2+1
− 2 + 1)e(t2 +1)b2
0 0 2 0 t 2 (t
Z0
1 1 c dt
(A.1) = arctan(c) − .
2 2 0 (t + 1)e(t2 +1)b2
2
References
[1] D. Bell, “Poisson’s remarkable calculation – a method or a trick?” Elem. Math. 65 (2010), 29–36.
[2] C. A. Berenstein and R. Gay, Complex Variables, Springer-Verlag, New York, 1991.
Z ∞ 2
[3] A. L. Delgado, “A Calculation of e−x dx,” The College Math. J. 34 (2003), 321–323.
0
[4] H. Iwasawa, “Gaussian Integral Puzzle,” Math. Intelligencer 31 (2009), 38–41.
[5] T. P. Jameson, “The Probability Integral by Volume of Revolution,” Mathematical Gazette 78 (1994), 339–340.
[6] H. Kneser, Funktionentheorie, Vandenhoeck and Ruprecht, 1958.
[7] P. S. Laplace, Théorie Analytique des Probabilités, Courcier, 1812.
[8] P. S. Laplace, “Mémoire sur la probabilité des causes par les évènemens,” Oeuvres Complétes 8, 27–65. (English
trans. by S. Stigler as “Memoir on the Probability of Causes of Events,” Statistical Science 1 (1986), 364–378.)
[9] P. M. Lee, http://www.york.ac.uk/depts/maths/histstat/normal history.pdf.
[10] L. Mirsky, The Probability Integral, Math. Gazette 33 (1949), 279. URL http://www.jstor.org/stable/
3611303.
[11] C. P. Nicholas and R. C. Yates, “The Probability Integral,” Amer. Math. Monthly 57 (1950), 412–413.
[12] G. Polya, “Remarks on Computing the Probability Integral in One and Two Dimensions,” pp. 63–78 in Berkeley
Symp. on Math. Statist. and Prob., Univ. California Press, 1949.
[13] R. Remmert, Theory of Complex Functions, Springer-Verlag, 1991.
[14] M. Rozman, “Evaluate Gaussian integral using differentiation under the integral sign,” Course notes for Physics
2400 (UConn), Spring 2016.
[15] W. Rudin, Principles of Mathematical Analysis, 3rd ed., McGraw-Hill, 1976.
[16] M. Spivak, Calculus, W. A. Benjamin, 1967.
[17] S. Stigler, “Laplace’s 1774 Memoir on Inverse Probability,” Statistical Science 1 (1986), 359–363.
[18] J. van Yzeren, “Moivre’s and Fresnel’s Integrals by Simple Integration,” Amer. Math. Monthly 86 (1979),
690–693.
[19] G. N. Watson, Complex Integration and Cauchy’s Theorem, Cambridge Univ. Press, Cambridge, 1914.
[20] http://gowers.wordpress.com/2007/10/04/when-are-two-proofs-essentially-the-same/#comment-239.
[21] http://math.stackexchange.com/questions/34767/int-infty-infty-e-x2-dx-with-complex-analysis.
[22] http://math.stackexchange.com/questions/390850/integrating-int-infty-0-e-x2-dx-using-feynmans-
parametrization-trick