When the coefficients of a linear homogeneous equation are constants, the equation can be solved using exponentials. Write the differential equation this way.
a0y + a1y′ + a2y′′ + … (up to an × the nth order derivative) = 0
Let p(x) be the corresponding polynomial a0+a1x+a2x2+…anxn. Let r be a root of p(x), and let f = Erx. Plug f into our differential equation and get p(r)×Erx. Since p(r) = 0, f is a solution.
Now assume r has multiplicity m. In other words, p(x) has m factors of x-r. Let's try to build a new solution via f = q(x)×erx. If q = 1 then f is the function described above, and we're done.
Show that f′ = (rq+q′)Erx, and f′′ = (r2q+2rq′+q′′)Erx. In general, the "next" derivative is found by taking the factor on the left of Erx, call it w, and replacing it with rw+w′. Thus the third derivative is (r3q+3r2q′+3rq′′+q′′′)Erx. The expression mirrors the binomial theorem.
When we search for a solution, the factor on the right, Erx, is always nonzero. Divide it out; it simply goes away. All that remains is the first factor, the one that looks like a binomial expansion of q and r.
As the derivatives march along, f′ f′′ f′′′ etc, pull out the first term in each expansion. These are multiplied by a1 a2 a3 etc. This gives p(r)×q(x), which is known to be 0. Thus the first term in the expansion of the derivatives goes away.
Multiply the second term in each expansion by its coefficient and get something like this.
a1q′ + 2a2rq′ + 3a3r2q′ + 4a4r3q′ + … (up to an)
This is the same as p′(x), evaluated at r, and multiplied by q′(x).
Invoke the theory of formal derivatives. If p has multiple roots at r, then p′(r) = 0. This entire sum drops out.
Now take the third term in the expansion of each derivative. Premultiply by the coefficients and get this.
a2q′′ + 3a3rq′′ + 6a4r2q′′ + 10a5r3q′′ + …
Confused about the "extra" coefficients? They come from the binomial theorem: 2 choose 2, 3 choose 2, 4 choose 2, 5 choose 2, and so on. In other words, k×(k-1)/2. Multiply the above expression through by 2 to get this.
2×1×a2q′′ + 3×2×a3rq′′ + 4×3×a4r2q′′ + 5×4×a5r3q′′ + …
This is equal to p′′(x), evaluated at r, times q′′(x). If the multiplicity of r is at least 3, i.e. p has at least 3 roots of r, then p′′(r) = 0, and this entire expression goes away.
Collect the fourth term from each derivative, multiply by the respective coefficients, then multiply by 6. The result is p′′′(x), evaluated at r, times q′′′(x). If the multiplicity m is at least 4, p′′′(r) = 0, and this expression goes away. This continues, until we build the mth expression, whence the mth derivative of p, evaluated at r, is no longer 0.
If f is a solution, the remaining expressions must drop to 0. The best way to do that is to have the mth derivative of q disappear. In other words, the mth derivative of q = 0, and q is any polynomial of degree less than m.
By the previous theorem, the solutions form a vector space. We have identified a subspace of solutions, corresponding to the root r. The subspace is q(x)×Erx, where q is the set of polynomials of degree < m. There is a convenient basis for this subspace:
Erx, xErx, x2Erx, x3Erx, … xm-1Erx
Apply this result to each root r in p(x). The basis functions are all linearly independent, (we'll prove this below), thus building a solution space of dimension n. Again, refer to the previous theorem; the solution space has dimension n. We have found all the solutions.
A linear differential equation with constant coefficients can be solved by finding the roots and multiplicities of the corresponding polynomial.
The functions that form our basis are pure exponentials Erx, (including E0), and exponentials with polynomial modifiers (e.g. x3Erx).
Suppose a linear combination c1 produces the zero function, and suppose all the functions in c1 are pure exponentials. Evaluate Erj, for each root r, and for each integer j from 0 to n-1. This builds an n×n matrix that is vandermonde, and nonsingular. Therefore these functions are linearly independent.
Suppose there is one exponential with several modifiers. Since Erx is everywhere nonzero, divide through by Erx to find a linear combination of powers of x that is equal to 0. A polynomial cannot be the zero function, hence we have a contradiction.
At this point our linear combination has more than one exponential, and at least one of these exponentials is modified by xj. Select a linear combination that has the smallest modifier xm, and the fewest number of exponentials carrying this modifier.
If our linear combination of functions is 0, the same is true of its derivative. Remember that xjErx becomes rxjErx + jxj-1Erx. Polynomials drop in degree, pure exponentials are multiplied by r, and modified exponentials experience mitosis.
Suppose xm is the highest multiplicity. If the only representative is a power of x, it has become xm-1, and we have a linear combination with lesser multiplicity, which is a contradiction.
There is at least one exponential associated with xm; call it xmErx. Let c1 be the original linear combination of functions, and let c2 be the derivative. Both contain the basis function xmErx. Let c3 = c1 - c2/r. Notice that the term xmErx drops out. Thus c3 has fewer terms modified by xm, or it has no terms based on xm. Either way we have a contradiction, since m is minimal, and the number of terms with xm is minimal. However, we still need to prove c3 is nontrivial.
Remember that c1 is based on more than one exponential. consider any other exponential in c1, such as Esx. Assume s ≠ 0. This is multiplied by s in c2, and 1-s/r in c3. In other words, the term persists in c3, giving a nontrivial linear combination. If s = 0, the term is simply xj. The highest power xj drops out of c2, hence it remains in c3.
In all cases, c3 is a nontrivial linear combination that yields 0, and violates our selection criteria. The pathological linear combination c1 cannot exist, and these functions are linearly independent. The existence of n different solutions, drawn from distinct basis functions, builds an n dimensional solution space, and completely solves the differential equation with constant coefficients.