Imagine a thin piece of wire, which only gains or loses heat through its ends. The temperature is a function of the position x along the wire and the time t. It is a solution to the heat equation
The solution to the one-dimensional heat equation with an arbitrary initial distribution is an infinite sum
If is the initial distribution, then the 's are given by
That is, is expressed as an infinite sum of sines.
A Fourier series represents a function as an infinite sum of trigonometric functions:
Consider integrable functions f and g defined on the interval , where . Define the inner product of f and g by
Example. Verify the linearity and symmetry axioms for an inner product for
Let and and be integrable functions on . Using properties of integrals, I have
Symmetry is easy:
Note that this should be called "inner product" (with quotes), since it isn't positive definite. Since , it follows that
But you could have a function which was nonzero --- for example, a function which was 0 at every point in , but nonzero at a single point --- such that .
We can get around this problem by confining ourselves to continuous functions. Unfortunately, many of the functions we'd like to apply Fourier series to aren't continuous.
Theorem. ( Orthogonality Relations)
Example. I'll illustrate the first formula with some numerical examples.
If I take different cosines, I should get 0. I'll try
Use the identity
I get
So
(Remember that the sine of a multiple of is 0!)
If I do the integral with both cosines the same, I use the double angle formula:
There is nothing essentially different in the derivation of the general formulas.
In terms of the (almost) inner product defined above, the orthogonality relations are:
In other words, the sine and cosine functions form an orthnormal set (again, allowing that we don't quite have an inner product here).
As a constant function, I'll use . You can verify that this is perpendicular to the sines and cosines above, and that it has length 1.
For vectors in an inner product space, you can get the components of a vectors by taking the inner product of the vector with elements of an orthonormal basis. For example, if is an orthonormal basis, then a vector x can be written as a linear combination of the u's by taking inner product of x with the u's:
By analogy, I might try to expand a function in terms of the sines and cosines by taking the inner product of f with the sines and cosines, and the constant function . Doing so, I get
The cosine term is .
The sine term is .
If I set in the formula, then since , I'd get
So for the constant function , the coefficient is
But my constant function is , so the constant term in the series will be
Putting everything together gives the Fourier series for :
You can interpret it as an expression for f as an infinite "linear combination" of the orthnormal sine and cosine functions.
There are several issues here. For one thing, this is an infinite sum, not the usual finite linear combination which expresses a vector in terms of a basis. Questions of convergence always arise with infinite sums. Moreover, even if the infinite sum converges, why should it converge to ?
In addition, we're only doing this by analogy with our results on inner product spaces, because this "inner product" only satisfies two of the inner product axioms.
It's important to know that these issues exist, but their treatment will be deferred to an advanced course in analysis.
Before doing some examples, here are some notes about computations.
You can often use the complex exponential (DeMoivre's formula) to simplify computations:
Specifically, let
Then
This allows me to compute a single integral, then find the real and imaginary parts of the result to get the sine and cosine coefficients.
On some occasions, symmetry can be used to obtain the values of some of the coefficients.
1. If a function is even --- that is, if for all x, so the graph is symmetric about the y-axis --- then for all n.
2. If a function is odd --- that is, if for all x, so the graph is symmetric about the origin --- then for all n.
This makes sense, since the cosine functions are even and the sine functions are odd.
Here's a final remark before I get to the examples. A sum of periodic functions is periodic. Since sine and cosine are periodic, you can only expect to represent periodic functions by a Fourier series. Therefore, outside the interval , you must "repeat" in periodic fashion (so for all x).
For example, consider the function , . , so I must "repeat" the function every two units. Picture:
The Fourier expansion of only converges to on the interval . Outside of that interval, it converges to the periodic function in the picture (except perhaps at the jump discontinuities).
Example. Find the Fourier expansion of
This graph is called a square wave.
This is an odd function, so for all k --- that is, all the cosine terms vanish. I'll compute all the coefficients anyway so you can see this directly.
Compute the constant term first:
Next, compute the terms:
Now
So
Now
So the Fourier expansion is
Here are the graphs of the first, fifth, and tenth partial sums:
Note that if , all the terms of the series are 0, and the series converges to 0. In fact, under reasonable conditions --- specifically, if f is of bounded variation in an interval around a point c --- the Fourier series will converges to the average of the left and right-hand limits at the point. In this case, the left-hand limit is -1, the right-hand limit is 1, and their average is 0. On the other hand, was defined to be 1.
Example. Find the Fourier expansion of
Compute the constant term first:
Next, compute the higher order terms:
Therefore,
Now take real and imaginary parts:
The Fourier series is
Example. Find the Fourier expansion of
Compute the constant term first:
Next, compute the higher order terms. As in the computation of , I only need the integral from to 0, since the function equals 0 from 0 to :
Thus,
The series is
Copyright 2016 by Bruce Ikenaga