Author: Timothy Devon Morris, devonmorris1992@gmail.com
%matplotlib notebook
import numpy as np
from mpl_toolkits.mplot3d import Axes3D
import matplotlib.pyplot as plt
from matplotlib import cm
A Quadratic form is simply a homogeneous polynomial of degree two. Quadratic forms appear in various branches of mathematics, such as differential geometry, optimization theory, probability theory, control theory and stability of dynamical systems. The geometry of quadratic forms is rich in both its applications and theory.
A homogeneous polynomial is polynomial whos nonzero terms have the same degree. In the case of quadratic forms, this degree must be two. For example, consider the polynomial
R(x,y)=x2+xy+y2.In this polynomial, every nonzero term has degree two, therefore it is a quadratic form. We can rewrite this equation as a matrix equation R(x,y)=[xy][10.50.51][xy]
The geometry of quadratic forms is rich in its structure. We can only visualize this geometry in two dimensions, but the ideas easily scale to the higher dimensional case. In the simplest case, we have A=[1001]
def plot_qform(A, title):
axx = A[0,0]
ayy = A[1,1]
axy = 2*A[0,1]
x = np.linspace(-2, 2)
y = np.linspace(-2, 2)
X, Y = np.meshgrid(x,y)
Z = axx*X**2 + axy*X*Y + ayy*Y**2
fig, ax = plt.subplots(figsize=(6,6))
CS = ax.contour(X, Y, Z, colors='k')
ax.clabel(CS, inline=1, fontsize=10)
plt.title(title)
plt.axis('equal')
plt.show()
A = np.eye(2)
plot_qform(A, "Simple Quadratic Form")
Since the level curves are all positive in this picture we have that xHAx>0, ∀x∈Rn∖{0}. We call such a quadtratic form positive definite. We can also plot this quadratic form as a surface in three dimensions.
def plot_qform3d(A, title):
axx = A[0,0]
ayy = A[1,1]
axy = 2*A[0,1]
x = np.linspace(-2, 2)
y = np.linspace(-2, 2)
X, Y = np.meshgrid(x,y)
Z = axx*X**2 + axy*X*Y + ayy*Y**2
fig = plt.figure(figsize=(6,6))
ax = fig.gca(projection='3d')
surf = ax.plot_surface(X, Y, Z, cmap=cm.coolwarm,
linewidth=0, antialiased=False)
plt.title(title)
plt.axis('equal')
plt.show()
plot_qform3d(A, "Simple Quadratic Form")
Just as there are positive definite quadratic forms, we can make negative definite quadtratic forms. These are quadratic forms that satisfy For example let xTAx<0, ∀x∈Rn∖{0}. For example let A=[−100−1].
A = -np.eye(2)
plot_qform(A, "Negative Definite Quadratic Form")
The dashed lines show that the quadratic form is negative. Plotting the surface in three dimensions we see an inverted picture from the positive definite case.
plot_qform3d(A, "Negative Definite Quadratic Form")
If the quadratic form is neither positive definite nor negative definite we have an indefinite quadratic form. For example, consider the case where A=[100−1].
A = np.eye(2)
A[1,1] = -1
plot_qform(A, "Indefinite Quadratic Form")
Plotting this form in three dimensions we see that it attains both positive and negative values.
plot_qform3d(A, "Indefinite Quadratic Form")
This surface makes the shape of a saddle, that's precisely why we call the point [0,0] a saddle point. Indefinite quadratic forms appear frequently in higher dimesions. There are also quadratic forms that are positive semi-definite and negative semi-definite, this means that xTAx≥0, ∀x∈Rn and xTAx≤0, ∀x∈Rn. The set of positive definite quadratic forms are a subset of positive semi-definite quadratic forms and similarly for negative definite quadratic forms. We also say that a matrix A is positive (negative) definite if its associated quadratic form is positive (negative) definite.
The beauty of quadratic forms is that we can determine their definiteness by analyzing the spectral properties of their matrix A. This is much easier than plotting their contours and analyzing their surfaces. We have the following theorem
Theorem: A matrix A is positive (negative) definite if and only if all of its eigenvalues are positive (negative)
Proof: Let A be a positive definite matrix and let v be an unit eigenvector of A, with associated eigenvalue λ. We have that vTAv=vTλv=λ>0
A similar result holds for positive semi-definite and negative semi-definite matrices, we just add the caveat that the eigenvalues can equal 0. In the previous section, the quadratic forms were fairly easily to analyze by inspection, since we used simple transformations of the identity matrix. However, when presented with a matrix such as A=[1.4−0.3−0.30.6],
A = np.array([[1.4, -0.3], [-0.3, 0.6]])
plot_qform(A, "More Complicated Quadratic Form")
plot_qform3d(A, "More complicated Quadratic Form")
However, we don't always have an advanced plotting system available to us. The eigenvalues and eigenvectors of this matrix are λ1=1.5v1=[0.9487−0.3162]
Now let Q=[0.94870.3162−0.31620.9387]
Thus we have our eigendecomposition A=QΛQT. We can rewrite our quadratic form as xTAx=xTQΛQTx=yTΛy
A = np.array([[1.5, 0],[0, 0.5]])
plot_qform(A, "Rotated Quadratic Form")
This plot shows that the eigenbasis constitutes a rotation of the original quadratic form. We can also plot the eigenbasis on the original quadratic form.
A = np.array([[1.4, -0.3], [-0.3, 0.6]])
Q = np.array([[0.9486833 , 0.31622777],[-0.31622777, 0.9486833]])
def plot_qform_axes(A, Q, title):
axx = A[0,0]
ayy = A[1,1]
axy = 2*A[0,1]
x = np.linspace(-2, 2)
y = np.linspace(-2, 2)
X, Y = np.meshgrid(x,y)
Z = axx*X**2 + axy*X*Y + ayy*Y**2
fig, ax = plt.subplots(figsize=(6,6))
CS = ax.contour(X, Y, Z, colors='k')
ax.clabel(CS, inline=1, fontsize=10)
origin = [0], [0]
ax.quiver(*origin, Q[0,:], Q[1,:], color='k', scale=3)
plt.title(title)
plt.axis('equal')
plt.show()
plot_qform_axes(A, Q, "Eigenbasis of Quadratic Form")
From this plot, we can see that the ellipse has a longer axis in the direction of the eigenvector with smaller associated eigenvalue. This is because the quadratic form grows slowest in the direction of the eigenvector with smallest eigenvalue. Similarly, the ellipse's shorter axis is the eigenvector with the larger eigenvalue. This is because the quadratic form grows quickest in this direction. This concept generalizes to higher dimensions.
Now that we know we can analyze the quadratic form associated with a matrix by simply lookin at its spectral properties, we can use python to find the spectrum of a matrix. Consider the matrix A=[2−10−12−10−12]
numpy.linalg.eig
function.
import numpy.linalg as la
A = np.array([[2., -1., 0.],
[-1., 2., -1.],
[0., -1., 2.]])
la.eig(A)
(array([3.41421356, 2. , 0.58578644]), array([[-5.00000000e-01, -7.07106781e-01, 5.00000000e-01], [ 7.07106781e-01, 4.05925293e-16, 7.07106781e-01], [-5.00000000e-01, 7.07106781e-01, 5.00000000e-01]]))
Thus we have that this matrix is positive definite. This would have been impossible to plot and time consuming to solve by hand. Now consider the matrix
[21−114−3−1−34]Using the numpy.linalg.eig
function gives us
A = np.array([[2., 1., -1],
[1., 2., -3.],
[-1., -3., 4.]])
la.eig(A)
(array([ 6.58774145, 1.60179257, -0.18953402]), array([[ 0.29152938, 0.94907855, -0.11941745], [ 0.56721933, -0.07099407, 0.82050111], [-0.77024208, 0.30693606, 0.55903255]]))
Which tells us that the matrix is indefinite.
We will now show how quadratic forms can be applied to stability theory of dynamical systems. First, we must start with a few definitions. Consider the autonomous (time-invariant) system given by
˙x=f(x)and suppose that this system meets the criteria to have a unique solution given some initial condition x0 (namely that it is lipschitz). A point ˉx is an equilibrium point if f(ˉx)=0
It is important to be able to classify these equilibrium points so that we can qualitatively determine how the system will behave. Stability is a desirable property of equilibrium points since we can conclude that within some region the system will stay near this equilibrium point. An equilibrium point ˉx=0 is stable if for every ϵ>0 we can find a δ=δ(ϵ)>0 such that ||x0||<δ→||x(t)||<ϵ, ∀t>0
A lyapunov function is a positive definite function V(x)>0 ∀x∈Rn∖{0}, that can be used to prove the stability of a dynamical system. Since postive definite matrices have an associated positive definite quadratic form, in many cases we can use positive definite matrices to prove stability of dynamical systems. The basic idea is as follows. Let V(x) be a positive definite function. If ˙V(x)<0 ∀x∈Rn∖{0}, then the equilibrium point ˉx=0 is asymptotically stable. This is proven explicitly in Khalil's Nonlinear Systems. We can, however, make an argument of the validity of this theorem if we analyze ˙V(x). By the chain rule, we have that ˙V(x)=n∑i=1∂V∂xidxidt=∇V(x)T˙x=∇V(x)Tf(x)
The power of quadratic forms really shines in Lyapunov theory when we consider linear systems ˙x=Ax. If the system is not linear, then we can linearize it by taking the jacobian and evaluating our jacobian at the equilibrium point. Therefore we will consider "locally linear systems." We will define our lyapunov function V(x)=xTPx
To see how these concepts come together geometrically, consider the dynamical system of a pendulum. ml2¨θ=mglsinθ−b˙θ
We can specify a negative definite Q matrix to find the appropriate lyapunov function for our system. Let
Q=[−100−1]Using python, we can solve ATP+PAT=Q for P which yields
from scipy.linalg import solve_lyapunov as lyap
g = 9.81
l = 1.
m = 1.
b = 0.5
A = np.array([[0, 1],[-g/l, -b/(m*l**2)]])
Q = -np.eye(2)
Q[1,1] = -2
P = lyap(A.T,Q)
P
array([[20.6454842, 0.0509684], [ 0.0509684, 2.1019368]])
So we have found a matrix P that represents a quadratic form that is a valid lyapunov function for this system. Thus the equilibrium point is stable.
All this math may seem hopelessly complex, but fortunately we can visualize the solution of this dynamical system and its associated lyapunov function. Using python we have the following plot.
from scipy.integrate import odeint
t = np.linspace(0,30,num=10000)
Ax = lambda x, t: A.dot(x)
x0 = [0.5, -0.2]
xt = odeint(Ax, x0, t)
title = "Lyapunov Function"
axx = P[0,0]
ayy = P[1,1]
axy = 2*P[0,1]
x = np.linspace(-1, 1)
y = np.linspace(-1, 1)
X, Y = np.meshgrid(x,y)
Z = axx*X**2 + axy*X*Y + ayy*Y**2
zt = axx*xt[:,0]**2 + axy*xt[:,0]*xt[:,1] + ayy*xt[:,1]**2
fig = plt.figure(figsize=(6,6))
ax = fig.gca(projection='3d')
surf = ax.plot_surface(X, Y, Z, cmap=cm.gray,
linewidth=0, antialiased=False)
ax.plot3D(xt[:,0],xt[:,1],zt, 'b')
plt.title(title)
plt.axis('equal')
plt.show()
Note that the solution of this system is always decreasing along the surface of the lyapunov function. This is precisely what it means to have a valid lyapunov function and for the equilibrium point to be stable!