Linear Systems of Ordinary Differential Equations with Constant Coefficients #
Introduction #
Probably the most thoroughly studied set of math models are linear systems of differential equations with constant coefficients (LSDECCs). While most of the world is non-linear, when a linear model is applicable, we should be able to take as great advantage of it as possible. This set of notes walks through the practical aspects of solving linear systems of differential equations with constant coefficients. The emphasis is upon how to solve these problems and in that spirit, proofs are left to references.
The world of homogeneous LSDECC is largely split into two categories: diagonalizable matrices and non-diagonalizable matrices. If a matrix is diagonalizable, then it has real distinct eigenvalues and a distinct eigenvector per eigenvalue. These problems are both easy to identify and simple to solve. Non-diagonalizable matrices are a little tricker to solve, but use the same intuition as the diagonalizable matrices. However, adding a forcing function to the system can quickly make these systems unsolvable. In this work, we will first address diagonalizable matrices. Then the methods will be extended to non-diagonalizable matrices and finally two techniques to handle non-homogeneous LSDECCs will be discussed.
Real Distinct Eigenvalues #
The basic idea is that we want to project the system in to a space that makes it easy to solve, then solve the easy problem and project back into our original space (see Figure 1). In “mathy” language, that means we want to project the system into a new basis (eigenspace) such that the dynamics nicely lay on the axises. So we turn to eigenvalues and eigenvectors. Things get a little tricky when eigenvalues have multiplicity or are complex. In those cases, we create a “subspace” within the eigenbasis such such that we still have orthogonality, but for now we will assume our matrices are nicely diagonalizable.
Consider the equation
where
and
where . If the systems of equations was a scalar equation, the solution would be an exponential. In the systems of equations, that is also the case:
where is called the fundamental solution matrix. Then the solution to the IVP is,
Now, the question is, what is the exponential operator upon a matrix? There are many different proofs and ways of answering that question that won’t be addressed here. In the case, has real distinct eigenvalues, here’s what to do.
where is a diagonal matrix of eigenvalues and is is the related eigenbasis (i.e., every eigenvector is a column of ). Thus,
where . Thus, the solution to the IVP is
Hermitian Matricies #
The derivation above assumes that is diagonalizable. Unfortunately, most systems are not diagonalizable, but it is pragmatic to know when a system is guaranteed to be diagonalizable. The study of diagonalizable systems is actually spectral theory and it often focuses on self-adjoint operators. While that field is too large to cover here, a little bit is quite helpful.
Definition - Hermitian Operator The Hermitian matrix operator ( ) is the complex conjugate transpose.
Definition - Hermitian Matrix A matrix ( ) is Hermitian if it is equal to it’s Hermitian:
For example,
is Hermitian, but
is not. This leads us to an important theorem.
Theorem 1.1 - A matrix, , is guaranteed to be diagonalizable (i.e., it has real, distinct eigenvalues) if and only if the matrix is Hermitian.
Corollary 1.1 - The eigenbasis of a Hermitian matrix is a unitary matrix.
Definition - Unitary Matrix A Unitary matrix is a matrix who’s inverse is equal to it’s transpose. That is, is unitary if and only if
The theorem and corollary make identifying and computing the solution to diagonalizable matrices much easier. For reference, a Hermitian matrix is a self-adjoint operator. Self-adjoint operators are important in the Dirac-von Neumann formulation of quantum mechanics where observables are self-adjoint operators and in particular, the Hamiltonian operator is self-adjoint.
Alternative Formulation #
MT note: I think format is more common among physicists and engineers than mathematicians.
Sometimes, authors will move the transform, , around in these equations. So it can get a little tricky and requires some care. One often useful formulation (which only works well for diagonalizable matrices) is to first realize that there are two different projections in the solution equation. First, the exponential solution in the eigenbasis is projected back into the original basis with . Second, the initial conditions, , are projected into the eigenbasis with . So, now consider setting initial conditions in the eigenbasis as . Then, for the solution may be written as:
where is the eigenvector associated with eigenvalue .
Examples #
Example 1 #
Todo
Example 2 #
Use Zero eigen value.
Real Eigenvalues with Multiplicity #
Now suppose we have real eigenvalues, but they are repeated. This means, we do not have distinct eigenvectors per eigenvalue and therefore the matrix is not diagonalizable. But never fear, we can reduce the matrix to a Jordon canonical form. With distinct eigenvalues, the idea is to project into a orthogonal space that decouples the dimensions. Here the idea is the same, except with repeated eigenvalues some dimensions are coupled so the idea is to create a subspace for each of the coupled dimensions such that within the subspace, the projection of the coupled dimensions are orthogonal. Before we discuss the particulars, we must build a little background.
Matrix Exponential #
In the previous discussion, the matrix exponential of the diagonal matrix was given as: . Here, we would like to give some intuition about where that comes from. Consider the scaler and the exponential . The exponential can be expanded with a Maclaurin series as
Notice, this definition only depends upon multiplication which is well defined for square matrices. Thus, we can use the Maclaurin series to define the exponential.
Definition - Matrix Exponential Given the matrix (see note 1), the exponential of is defined as
Notice that when is diagonal, the series becomes
as given before. However, in most cases, the summation is infinite and nasty, but there are a few properties to help out.
The matrix exponential has the following properties
- is defined and .
- Given , .
- .
- exists for all .
- .
While most of these properties are follow from the definition of the scaler exponential, the biggest difference is property (2) which requires and to commute. Forgetting that requirement can easily lead to the wrong answer.
Jordon Canonical Form #
Consider the equation
where , and is the initial condition. Suppose has unique eigenvalues, , such that is repeated times. Thus, . Repeated eigenvalues means that dimensions are coupled and cannot be separated through projection. We can, however, create an orthogonal subspace to project the dimensions into. First, we must find the generaized eigenvectors. The set of generalized eigenvectors associated with eigenvalue is given by
where , and . The trivial case is ignored and . Let be the projection from the system basis into the basis formed by generalized eigenvalues associated with eigenvalue defined by
Thus, we can form a generalized eigenbasis with as
where is an an matrix. Using the generalized eigenbasis, the matrix can be composed of
where is the Jordon matrix with Jordon blocks defined by
where is the identity matrix and is the nilpotent matrix given by,
Before continuing, let’s clarify the dimensions and indices of everything.
- - the dimension of and therefore and .
- - the number of unique eigenvalues of .
- - the multiplicity of the eigenvalue of .
- And
Further notice
- is the set of generalized eigenvectors associated with .
- is the Jordon block associated with .
- If for all then the Jordon matrix of the same form as when is diagonalizable.
By applying the decomposition of , the fundamental solution matrix is given by
and the particular solution is
The exponential where
since and commute. Notice that the exponential summation is finite because is nilpotent.
Whew, that is a lot. So what have we done? The matrix can’t be diagonalized so we’ve created subspaces for each set of coupled dimensions using the orthogonality of polynomials which all together form a generalized eigenbasis. We then us the generalized eigenvalues to project into the generalized eigenbasis, solve the system in the generalized eigenbasis, then project back into the original basis.
Complex Eigenvalues - Real Jordon Canonical Form #
If is real, then and may have complex entries, but any complex eigenvalues will occur in conjugate pairs; we can use this fact to simplify multiplicity of complex eigenvalues using the real Jordon canonical form.
Consider the complex eigenvalue pair such that with multiplicity (e.g., for , appears three times and appears three times). Rather than using two Jordon Blocks, and , use one block, defined as
where is the identity matrix and
Then the Jordon matrix uses for real eigenvalues and for complex eigenvalues.
Examples #
Example 1 #
Solve the homogeneous systems of differential equations
for initial conditions
Solution From the system of differential equations, we find
Then find the eigenvalues:
Thus, with multiplicity 2. Next, find generalized eigenvectors
and
Then the generalized eigenbasis is
and the inverse (using Gaussian elimination)
From theory, we know the Jordon matrix should have one block and be
but if we forget (or wish to verify), we can simply use the generalized eigenbasis to find
Then Thus, the fundamental solution matrix is
and the solution to the IVP is
Example 2 #
Multiplicity of three
Example 3 #
Complex eigenvalues and the real form
Non-Homogeneous Linear Systems of ODEs #
Now we can make things more fun by adding in forcing functions. Consider the non-homogenious system of differential equations
where , and . The function is a forcing function. Just like for scaler ODS’s there are largely two methods for handling forcing functions: method of undetermined coefficients (i.e., guessing) and method of variation of parameters. Unfortunately, the easier, method of undetermined coefficients is far more limited and the difficult (sometimes computationally impossible) method of variation of parameters is the only option.
Method of Undetermined Coefficients #
The method of undetermined coefficients (i.e., guessing) can be quite effective for scalar differential equations, but is rather limited for systems of differential equations. The only guesses that work out nicely are for constants, positive integer powers of , exponential functions and or .
First, (as always) solve the homogeneous system. This is especially important for exponential functions and or because one or more of the eigenvalues of may conflict with the forcing function. Then, the form of the particular solution can be “guessed” with now vectors for coefficients and back solved into the equation to find the coefficients. Just like for scaler ODEs, the general solution is the sum of the homogeneous and and particular solution
Constants #
The easiest guess is when is constant. The appropriate guess is
where . Then substitute the particular solution into the non-homogeneous equation to find the constants .
Positive Powers of #
Positive integer powers of is also not too tricky. If is order , then the appropriate guess for the form of the particular solution is,
where are the constants to solve by substitution.
Exponential and Sinusoid Functions #
It’s usually easiest to use Euler’s formula to make and into complex exponential functions, so we will consider both exponential and sinusoids together. For an exponential forcing function of the form , the appropriate guess for the particular solution is
where are the coefficients to back solve, and where is the multiplicity of the eigenvalue of the system such that . The basic idea is that we must ensure that the particular solution is orthogonal to the eigenspace of the homogeneous solution and we use polynomials to do that. Let’s look at a few cases to understand this better. In the case is not an eigenvalue of the system (i.e, ), then and
In the case is an eigenvalue of the system with multiplicity , then
Notice, by multiplying by , it ensures that the particular solution is orthogonal to the eigenspace of the homogeneous solution and does not interfere. When conflicts with eigenvalues of the homogeneous solution of higher order, we simply use a one higher polynomial power as that of the associated Jordon Block to ensure orthogonality.
Method of Variation of Parameters #
Variation of parameters is guaranteed to be valid. However, often the integration is impossible. Let and be a continuous map from time into . Then for every , the initial value problem
has a unique global solution given by the formula
Examples #
TODO
Notes #
- Actually, can be any linear operation on a finite normed vector space.
Ref #
[1] Sideris 2013
[2] Kreyszig 2011
[3] Schaeffer 2016
[4] Hirsch 2013