Author: Thane Downing thanedowning@gmail.com
The matrix exponential is a useful tool in solving systems of linear differential equations, helping to describe the behavior of linear systems, as well as having several other intersting uses. The matrix exponential works much like the scalar exponential and has many of the same properties. While perhaps initially unfamiliar, the matrix exponential developes readily from very familiar concepts and contributes to a straightforward understanding of how matrix math fits into a larger framework.
The eponential of any square matrix A can be writen as eA, and can be understood in a similar way that we understand the scalar equivalent exponential ex. From calculus, we know that ddxex=xex
One classic property of scalar exponentials that doesn't translate generally to matrix exponentials is the property eaeb=ea+b
An expanded form to represent the matrix exponential, that only works for diagonalizable matrices, is given as eA=XeΛX−1, where Λ is the diagonal matrix of the eigenvalues of A and X is an invertable matrix made up of the eigenvectors of A. Taking this one step futher we can find that eΛ=(eλ1eλ2⋱eλn)
The matrix exponential is handy in solving systems of linear differential equations, such as those encountered in continuous time state space equations. Given some linear ordinary differential equation dxdt=ax, we can determine that the solution is x(t)=eat. Now suppose we have a system of linear differential equations such that dxdt=Ax. The solution to this system is x(t)=eAtx(0). Using the expanded form given above, this solution can be rewriten x(t)=eAtx(0)=XeΛtX−1x(0). We can use the eigenvalues of A to determine the behavior of the solution similar to the way we can determine system dynamics in other settings. The solution will decay if Re(λi)<0, it will just oscillate if Re(λi)=0 (λi is purley imaginary), and it blow up if Re(λi)>0, where λi is the eigen value associated with a given eigenvector in the system.
Occasionally, this solution using the matrix exponential will be represented in terms of the Laplace transform as in eAt=L−1(sI−A)−1
In the context of continuous time LTI system defined by the state space equations ˙x=Ax(t)+Bf(t)y(t)=Cx(t)+Df(t)
Although there are contrived examples where the solution to the matrix exponential is nice and tidy, in general the solution will be rather complex. For instance, the general solution to a eA, where A is a 2x2 matrix A=(abcd)
There are several examples we can use to verify the results of the preceeding section. First, we generate a random square matrix and use the matrix exonential function from the scipy.linalg library to calculate the matrix exponential.
import numpy as np
from scipy.linalg import expm
from math import factorial, sinh, cosh, exp, sqrt, cos, ceil
import random
import matplotlib.pyplot as plt
A = np.array([[random.random(),random.random(),random.random(),random.random()],
[random.random(),random.random(),random.random(),random.random()],
[random.random(),random.random(),random.random(),random.random()],
[random.random(),random.random(),random.random(),random.random()]])
print "A =\n", A
print "e^A =\n", expm(A)
A = [[ 0.41777503 0.53103038 0.04252436 0.41216531] [ 0.69812543 0.79625546 0.76295725 0.41021156] [ 0.53157639 0.82114634 0.18914389 0.5079495 ] [ 0.16940129 0.18442986 0.58999039 0.67698445]] e^A = [[ 2.11211077 1.38598808 0.72110255 1.16504324] [ 2.09831048 3.57898164 1.9539366 1.83430856] [ 1.59212573 2.1362369 2.19697908 1.61725546] [ 0.88873006 1.12036052 1.30709399 2.58273717]]
Next, we can test our Taylor series expansion method (making sure to use enough iterations to get a good approximation) to verify that this approach will give us the same thing as the library function does.
taylor_estimate = np.eye(4) + A
add = A
for i in range(2,200):
add = np.matmul(add,A)/i
taylor_estimate = taylor_estimate + add
print taylor_estimate
[[ 2.11211077 1.38598808 0.72110255 1.16504324] [ 2.09831048 3.57898164 1.9539366 1.83430856] [ 1.59212573 2.1362369 2.19697908 1.61725546] [ 0.88873006 1.12036052 1.30709399 2.58273717]]
We can also test the expanded form of the matrix exponential by using the scipy.linalg library to find the eigenvalues and eigenvectors of A to verify that this method also produces what we expect to see.
lam, X = np.linalg.eig(A)
eigen_decomp = np.matmul(np.matmul(X,expm(np.diag(lam))),np.linalg.inv(X))
print eigen_decomp
[[ 2.11211077 1.38598808 0.72110255 1.16504324] [ 2.09831048 3.57898164 1.9539366 1.83430856] [ 1.59212573 2.1362369 2.19697908 1.61725546] [ 0.88873006 1.12036052 1.30709399 2.58273717]]
We can check the basic property (eA)−1=e−A
print "e^-A =\n", expm(-A)
print "(e^A)^-1 =\n", np.linalg.inv(expm(A))
e^-A = [[ 0.76307929 -0.33076492 0.17330506 -0.21782085] [-0.30979663 0.71708154 -0.50396241 -0.05396847] [-0.25085094 -0.497817 1.11972926 -0.23443581] [-0.0012397 0.05469565 -0.40770436 0.60419532]] (e^A)^-1 = [[ 0.76307929 -0.33076492 0.17330506 -0.21782085] [-0.30979663 0.71708154 -0.50396241 -0.05396847] [-0.25085094 -0.497817 1.11972926 -0.23443581] [-0.0012397 0.05469565 -0.40770436 0.60419532]]
A quick check will show that indeed, in general eAeB≠eA+B
B = np.array([[random.random(),random.random(),random.random(),random.random()],
[random.random(),random.random(),random.random(),random.random()],
[random.random(),random.random(),random.random(),random.random()],
[random.random(),random.random(),random.random(),random.random()]])
C = expm(A)*expm(B)
D = expm(A+B)
print "(e^A)*(e^B) =\n",C
print "e^(A+B) =\n",D
(e^A)*(e^B) = [[ 2.89313406 0.6620096 0.70024447 0.35380883] [ 4.2540861 12.28011182 5.30642275 3.5365327 ] [ 2.72511048 4.51514568 6.84983879 2.1878806 ] [ 1.759763 2.07105209 2.90330633 5.14254724]] e^(A+B) = [[ 7.49914959 7.82533417 7.16019721 5.87058551] [ 17.40289883 21.92329833 19.12141699 15.75151267] [ 14.0086614 17.02401286 15.89451394 12.54361878] [ 12.57003962 14.49902883 13.63391563 12.01335549]]
We can even test the formula for computing a 2x2 matrix (abcd) using the terms a, b, c, and d.
a = random.random()
b = random.random()
c = random.random()
d = random.random()
H = expm(np.array([[a,b],[c,d]]))
G = (1/sqrt((a-d)**2+4*b*c))*np.array([
[exp((a+d)/2)*(sqrt((a-d)**2+4*b*c)*cosh(.5*sqrt((a-d)**2+4*b*c))+(a-d)*sinh(.5*sqrt((a-d)**2+4*b*c))),
2*b*exp(((a+d)/2))*sinh(.5*sqrt((a-d)**2+4*b*c))],
[2*c*exp(((a+d)/2))*sinh(.5*sqrt((a-d)**2+4*b*c)),
exp(((a+d)/2))*(sqrt((a-d)**2+4*b*c)*cosh(.5*sqrt((a-d)**2+4*b*c))+(d-a)*sinh(.5*sqrt((a-d)**2+4*b*c)))]])
print "H =\n", H
print "G =\n", G
H = [[ 1.76853665 1.72718204] [ 1.81716025 3.22362205]] G = [[ 1.76853665 1.72718204] [ 1.81716025 3.22362205]]
One of the most applicable uses of the matrix exponential is in propagating the states of a system described by state space equations. We can choose from thousands of examples of physical systems that have well defined state space representations to use as the basis for this exercise. The following code is flexible enough to implement any LTI constant coefficient system. Only the A, B, C, and D matrices need to be changed (along with the corresponding dimensions of u, x, and y).
The following example models an RLC circuit with two capacitors in parallel, where the input is an AC voltage source with an amplitude of five volts, the intermidiary states are the voltages over the two capcitors, and the output is the current through the first resistor, as shown in the picture below.
c1 = 10*10**-6 # In Farads
c2 = 20*10**-6 # In Farads
r1 = 200 # In Ohms
r2 = 100 # In Ohms
A = np.array([[-1/(c1*r1)-1/(c1*r2),1/(c1*r2)],
[1/(c2*r2),-1/(c2*r2)]])
B = np.array([[1/(c1*r1)],[0]])
C = np.array([[-1/r1,0]])
D = np.array([[1/r1]])
x_0 = np.array([[0],[4]])
t_step = 0.001
t_start = 0
t_end = 15
bins = int(ceil((t_end-t_start)/t_step))
u_vals = np.zeros((bins,1))
x_vals = np.zeros((bins,2))
y_vals = np.zeros((bins,1))
P = expm(A*t_step) # This is the propagation matrix
x_val = x_0
for i in range(bins):
u_val = 5*cos(i*t_step)
y_val = np.matmul(C,x_val)+D*u_val
u_vals[i] = u_val
x_vals[i,0] = x_val[0]
x_vals[i,1] = x_val[1]
y_vals[i] = y_val
x_val = np.matmul(P,x_val)+np.matmul(P,B)*u_val*t_step
In the following plots, the red line shows the voltage of the input signal u(t) (AC voltage source), the blue and black dashed lines show the voltage on capacitors 1 and 2 respectively, and the green line shows the current through resistor 1 in amps. The x axis is the number of time samples.
plt.figure(1)
plt.title("System Dynamics of an RLC Circuit")
plt.subplot(211)
plt.plot(u_vals, 'r-', x_vals[:,0], 'b--', x_vals[:,1])
plt.title("System Dynamics of an RLC Circuit")
plt.ylabel('Voltage (in Volts)')
plt.axis([0, 15000, -5.5, 5.5])
plt.subplot(212)
plt.plot( y_vals, 'g-')
plt.ylabel('Current (in Amps)')
plt.axis([0, 15000, -5.5, 5.5])
plt.show
<function matplotlib.pyplot.show>
"Mathematical Methods and Algorithms for Signal Processing" by Todd K. Moon and Wynn C. Stirling
http://mathworld.wolfram.com/MatrixExponential.html
http://web.mit.edu/18.06/www/Spring17/Matrix-Exponentials.pdf
http://web.mit.edu/16.unified/www/FALL/signalssystems/Lecture11_12.pdf