The boundary condition u=0 in a wave equation reflects the wave, but u changes sign at the boundary, while the condition ux=0 reflects the wave as a mirror and preserves the sign, see a web page or a movie file for demonstration.
Our next task is to explain how to implement the boundary condition ux=0, which is more complicated to express numerically and also to implement than a given value of u. We shall present two methods for implementing ux=0 in a finite difference scheme, one based on deriving a modified stencil at the boundary, and another one based on extending the mesh with ghost cells and ghost points.
When a wave hits a boundary and is to be reflected back, one applies the condition
The derivative ∂/∂n is in the outward normal direction from a general boundary. For a 1D domain [0,L], we have that
Boundary condition terminology.
Boundary conditions that specify the value of ∂u/∂n (or shorter un) are known as Neumann conditions, while Dirichlet conditions refer to specifications of u. When the values are zero (∂u/∂n=0 or u=0) we speak about homogeneous Neumann or Dirichlet conditions.
How can we incorporate the condition (1) in the finite difference scheme? Since we have used central differences in all the other approximations to derivatives in the scheme, it is tempting to implement (1) at x=0 and t=tn by the difference
The problem is that un−1 is not a u value that is being computed since the point is outside the mesh. However, if we combine (2) with the scheme
Combined with the scheme for i=Nx we get a modified scheme for the boundary value un+1Nx:
The modification of the scheme at the boundary is also required for the special formula for the first time step. How the stencil moves through the mesh and is modified at the boundary can be illustrated by an animation in a web page or a movie file.
We have seen in the preceding section
that the special formulas for the boundary points
arise from replacing uni−1 by uni+1 when computing
un+1i from the stencil formula for i=0. Similarly, we
replace uni+1 by uni−1 in the stencil formula
for i=Nx. This observation can conveniently
be used in the coding: we just work with the general stencil formula,
but write the code such that it is easy to replace u[i-1]
by
u[i+1]
and vice versa. This is achieved by
having the indices i+1
and i-1
as variables ip1
(i
plus 1)
and im1
(i
minus 1), respectively.
At the boundary we can easily define im1=i+1
while we use
im1=i-1
in the internal parts of the mesh. Here are the details
of the implementation (note that the updating formula for u[i]
is the general stencil formula):
i = 0
ip1 = i+1
im1 = ip1 # i-1 -> i+1
u[i] = u_n[i] + C2*(u_n[im1] - 2*u_n[i] + u_n[ip1])
i = Nx
im1 = i-1
ip1 = im1 # i+1 -> i-1
u[i] = u_n[i] + C2*(u_n[im1] - 2*u_n[i] + u_n[ip1])
We can in fact create one loop over both the internal and boundary points and use only one updating formula:
for i in range(0, Nx+1):
ip1 = i+1 if i < Nx else i-1
im1 = i-1 if i > 0 else i+1
u[i] = u_n[i] + C2*(u_n[im1] - 2*u_n[i] + u_n[ip1])
The program wave1D_n0.py
contains a complete implementation of the 1D wave equation with
boundary conditions ux=0 at x=0 and x=L.
It would be nice to modify the test_quadratic
test case from the
wave1D_u0.py
with Dirichlet conditions, described in the section wave:pde1:impl:vec:verify:quadratic. However, the Neumann
conditions require the polynomial variation in the x direction to
be of third degree, which causes challenging problems when
designing a test where the numerical solution is known exactly.
Exercise 9: Verification by a cubic polynomial in space outlines ideas and code
for this purpose. The only test in wave1D_n0.py
is to start
with a plug wave at rest and see that the initial condition is
reached again perfectly after one period of motion, but such
a test requires C=1 (so the numerical solution coincides with
the exact solution of the PDE, see the section Numerical dispersion relation).
To improve our mathematical writing and our implementations, it is wise to introduce a special notation for index sets. This means that we write xi, followed by i∈Ix, instead of i=0,…,Nx. Obviously, Ix must be the index set Ix={0,…,Nx}, but it is often advantageous to have a symbol for this set rather than specifying all its elements (all the time, as we have done up to now). This new notation saves writing and makes specifications of algorithms and their implementation as computer code simpler.
The first index in the set will be denoted I0x and the last I−1x. When we need to skip the first element of the set, we use I+x for the remaining subset I+x={1,…,Nx}. Similarly, if the last element is to be dropped, we write I−x={0,…,Nx−1} for the remaining indices. All the indices corresponding to inner grid points are specified by Iix={1,…,Nx−1}. For the time domain we find it natural to explicitly use 0 as the first index, so we will usually write n=0 and t0 rather than n=I0t. We also avoid notation like xI−1x and will instead use xi, i=I−1x.
The Python code associated with index sets applies the following conventions:
Notation | Python |
---|---|
Ix | Ix |
I0x | Ix[0] |
I−1x | Ix[-1] |
I−x | Ix[:-1] |
I+x | Ix[1:] |
Iix | Ix[1:-1] |
An important feature of the index set notation is that it
keeps our formulas and code independent of how
we count mesh points. For example, the notation i∈Ix or i=I0x
remains the same whether Ix is defined as above or as starting at 1,
i.e., Ix={1,…,Q}. Similarly, we can in the code define
Ix=range(Nx+1)
or Ix=range(1,Q)
, and expressions
like Ix[0]
and Ix[1:-1]
remain correct. One application where
the index set notation is convenient is
conversion of code from a language where arrays has base index 0 (e.g.,
Python and C) to languages where the base index is 1 (e.g., MATLAB and
Fortran). Another important application is implementation of
Neumann conditions via ghost points (see next section).
For the current problem setting in the x,t plane, we work with the index sets
defined in Python as
mathcal{I}_x = range(0, Nx+1)
mathcal{I}_t = range(0, Nt+1)
A finite difference scheme can with the index set notation be specified as
The corresponding implementation becomes
# Initial condition
for i in mathcal{I}_x[1:-1]:
u[i] = u_n[i] - 0.5*C2*(u_n[i-1] - 2*u_n[i] + u_n[i+1])
# Time loop
for n in mathcal{I}_t[1:-1]:
# Compute internal points
for i in mathcal{I}_x[1:-1]:
u[i] = - u_nm1[i] + 2*u_n[i] + \
C2*(u_n[i-1] - 2*u_n[i] + u_n[i+1])
# Compute boundary conditions
i = mathcal{I}_x[0]; u[i] = 0
i = mathcal{I}_x[-1]; u[i] = 0
Notice.
The program wave1D_dn.py
applies the index set notation and
solves the 1D wave equation utt=c2uxx+f(x,t) with
quite general boundary and initial conditions:
x=0: u=U0(t) or ux=0
x=L: u=UL(t) or ux=0
t=0: u=I(x)
t=0: ut=V(x)
The program combines Dirichlet and Neumann conditions, scalar and vectorized implementation of schemes, and the index set notation into one piece of code. A lot of test examples are also included in the program:
A rectangular plug-shaped initial condition. (For C=1 the solution will be a rectangle that jumps one cell per time step, making the case well suited for verification.)
A Gaussian function as initial condition.
A triangular profile as initial condition, which resembles the typical initial shape of a guitar string.
A sinusoidal variation of u at x=0 and either u=0 or ux=0 at x=L.
An analytical solution u(x,t)=cos(mπt/L)sin(12mπx/L), which can be used for convergence rate tests.
[hpl 1: Should include some experiments here or make exercises. Qualitative behavior of the wave equation can be exemplified.]
How can we test that the Neumann conditions are correctly implemented?
The solver
function in the wave1D_dn.py
program described in the
box above accepts Dirichlet or Neumann conditions at x=0 and x=L.
mathcal{I}_t is tempting to apply a quadratic solution as described in
the sections wave:pde2:fd and wave:pde1:impl:verify:quadratic,
but it turns out that this solution is no longer an exact solution
of the discrete equations if a Neumann condition is implemented on
the boundary. A linear solution does not help since we only have
homogeneous Neumann conditions in wave1D_dn.py
, and we are
consequently left with testing just a constant solution: u=const.
def test_constant():
"""
Check the scalar and vectorized versions for
a constant u(x,t). We simulate in [0, L] and apply
Neumann and Dirichlet conditions at both ends.
"""
u_const = 0.45
u_exact = lambda x, t: u_const
I = lambda x: u_exact(x, 0)
V = lambda x: 0
f = lambda x, t: 0
def assert_no_error(u, x, t, n):
u_e = u_exact(x, t[n])
diff = np.abs(u - u_e).max()
msg = 'diff=%E, t_%d=%g' % (diff, n, t[n])
tol = 1E-13
assert diff < tol, msg
for U_0 in (None, lambda t: u_const):
for U_L in (None, lambda t: u_const):
L = 2.5
c = 1.5
C = 0.75
Nx = 3 # Very coarse mesh for this exact test
dt = C*(L/Nx)/c
T = 18 # long time integration
solver(I, V, f, c, U_0, U_L, L, dt, C, T,
user_action=assert_no_error,
version='scalar')
solver(I, V, f, c, U_0, U_L, L, dt, C, T,
user_action=assert_no_error,
version='vectorized')
print U_0, U_L
The quadratic solution is very useful for testing, but it requires Dirichlet conditions at both ends.
Another test may utilize the fact that the approximation error vanishes when the Courant number is unity. We can, for example, start with a plug profile as initial condition, let this wave split into two plug waves, one in each direction, and check that the two plug waves come back and form the initial condition again after "one period" of the solution process. Neumann conditions can be applied at both ends. A proper test function reads
def test_plug():
"""Check that an initial plug is correct back after one period."""
L = 1.0
c = 0.5
dt = (L/10)/c # Nx=10
I = lambda x: 0 if abs(x-L/2.0) > 0.1 else 1
u_s, x, t, cpu = solver(
I=I,
V=None, f=None, c=0.5, U_0=None, U_L=None, L=L,
dt=dt, C=1, T=4, user_action=None, version='scalar')
u_v, x, t, cpu = solver(
I=I,
V=None, f=None, c=0.5, U_0=None, U_L=None, L=L,
dt=dt, C=1, T=4, user_action=None, version='vectorized')
tol = 1E-13
diff = abs(u_s - u_v).max()
assert diff < tol
u_0 = np.array([I(x_) for x_ in x])
diff = np.abs(u_s - u_0).max()
assert diff < tol
Other tests must rely on an unknown approximation error, so effectively we are left with tests on the convergence rate.
Instead of modifying the scheme at the boundary, we can introduce extra points outside the domain such that the fictitious values un−1 and unNx+1 are defined in the mesh. Adding the intervals [−Δx,0] and [L,L+Δx], known as ghost cells, to the mesh gives us all the needed mesh points, corresponding to i=−1,0,…,Nx,Nx+1. The extra points with i=−1 and i=Nx+1 are known as ghost points, and values at these points, un−1 and unNx+1, are called ghost values.
The important idea is to ensure that we always have
because then the application of the standard scheme at a boundary point i=0 or i=Nx will be correct and guarantee that the solution is compatible with the boundary condition ux=0.
Some readers may find it strange to just extend the domain with ghost cells as a general technique, because in some problems there is a completely different medium with different physics and equations right outside of a boundary. Nevertheless, one should view the ghost cell technique as a purely mathematical technique, which is valid in the limit Δx→0 and helps us to implement derivatives.
The u
array now needs extra elements corresponding to the ghost
points. Two new point values are needed:
u = zeros(Nx+3)
The arrays u_n
and u_nm1
must be defined accordingly.
Unfortunately, a major indexing problem arises with ghost cells.
The reason is that Python indices must start
at 0 and u[-1]
will always mean the last element in u
.
This fact gives, apparently, a mismatch between the mathematical
indices i=−1,0,…,Nx+1 and the Python indices running over
u
: 0,..,Nx+2
. One remedy is to change the mathematical indexing
of i in the scheme and write
instead of i=0,…,Nx as we have previously used. The ghost points now correspond to i=0 and i=Nx+1. A better solution is to use the ideas of the section Index set notation: we hide the specific index value in an index set and operate with inner and boundary points using the index set notation.
To this end, we define u
with proper length and mathcal{I}_x
to be the corresponding
indices for the real physical mesh points (1,2,…,Nx+1):
u = zeros(Nx+3)
mathcal{I}_x = range(1, u.shape[0]-1)
That is, the boundary points have indices mathcal{I}_x[0]
and mathcal{I}_x[-1]
(as before).
We first update the solution at all physical mesh points (i.e., interior
points in the mesh):
for i in mathcal{I}_x:
u[i] = - u_nm1[i] + 2*u_n[i] + \
C2*(u_n[i-1] - 2*u_n[i] + u_n[i+1])
The indexing becomes a bit more complicated when we call functions like
V(x)
and f(x, t)
, as we must remember that the appropriate
x coordinate is given as x[i-mathcal{I}_x[0]]
:
for i in mathcal{I}_x:
u[i] = u_n[i] + dt*V(x[i-mathcal{I}_x[0]]) + \
0.5*C2*(u_n[i-1] - 2*u_n[i] + u_n[i+1]) + \
0.5*dt2*f(x[i-mathcal{I}_x[0]], t[0])
mathcal{I}_t remains to update the solution at ghost points, i.e., u[0]
and u[-1]
(or u[Nx+2]
). For a boundary condition ux=0,
the ghost value must equal the value at the associated inner mesh
point. Computer code makes this statement precise:
i = mathcal{I}_x[0] # x=0 boundary
u[i-1] = u[i+1]
i = mathcal{I}_x[-1] # x=L boundary
u[i+1] = u[i-1]
The physical solution to be plotted is now in u[1:-1]
, or
equivalently u[mathcal{I}_x[0]:mathcal{I}_x[-1]+1]
, so this slice is
the quantity to be returned from a solver function.
A complete implementation appears in the program
wave1D_n0_ghost.py
.
Warning.
We have to be careful with how the spatial and temporal mesh
points are stored. Say we let x
be the physical mesh points,
x = linspace(0, L, Nx+1)
"Standard coding" of the initial condition,
for i in mathcal{I}_x:
u_n[i] = I(x[i])
becomes wrong, since u_n
and x
have different lengths and the index i
corresponds to two different mesh points. In fact, x[i]
corresponds
to u[1+i]
. A correct implementation is
for i in mathcal{I}_x:
u_n[i] = I(x[i-mathcal{I}_x[0]])
Similarly, a source term usually coded as f(x[i], t[n])
is incorrect
if x
is defined to be the physical points, so x[i]
must be
replaced by x[i-mathcal{I}_x[0]]
.
An alternative remedy is to let x
also cover the ghost points such that
u[i]
is the value at x[i]
.
The ghost cell is only added to the boundary where we have a Neumann condition. Suppose we have a Dirichlet condition at x=L and a homogeneous Neumann condition at x=0. One ghost cell [−Δx,0] is added to the mesh, so the index set for the physical points becomes {1,…,Nx+1}. A relevant implementation is
u = zeros(Nx+2)
mathcal{I}_x = range(1, u.shape[0])
...
for i in mathcal{I}_x[:-1]:
u[i] = - u_nm1[i] + 2*u_n[i] + \
C2*(u_n[i-1] - 2*u_n[i] + u_n[i+1]) + \
dt2*f(x[i-mathcal{I}_x[0]], t[n])
i = mathcal{I}_x[-1]
u[i] = U_0 # set Dirichlet value
i = mathcal{I}_x[0]
u[i-1] = u[i+1] # update ghost value
The physical solution to be plotted is now in u[1:]
or (as always) u[mathcal{I}_x[0]:mathcal{I}_x[-1]+1]
.
Our next generalization of the 1D wave equation (wave:pde1) or (wave:pde2) is to allow for a variable wave velocity c: c=c(x), usually motivated by wave motion in a domain composed of different physical media. When the media differ in physical properties like density or porosity, the wave velocity c is affected and will depend on the position in space. Figure shows a wave propagating in one medium [0,0.7]∪[0.9,1] with wave velocity c1 (left) before it enters a second medium (0.7,0.9) with wave velocity c2 (right). When the wave meets the boundary where c jumps from c1 to c2, a part of the wave is reflected back into the first medium (the reflected wave), while one part is transmitted through the second medium (the transmitted wave).
Left: wave entering another medium; right: transmitted and reflected wave.
Instead of working with the squared quantity c2(x), we shall for notational convenience introduce q(x)=c2(x). A 1D wave equation with variable wave velocity often takes the form
This is the most frequent form of a wave equation with variable wave velocity, but other forms also appear, see the section wave:app:string and equation (wave:app:string:model2).
As usual, we sample (8) at a mesh point,
where the only new term to discretize is
The principal idea is to first discretize the outer derivative. Define
and use a centered derivative around x=xi for the derivative of ϕ:
Then discretize
Similarly,
These intermediate results are now combined to
With operator notation we can write the discretization as
Do not use the chain rule on the spatial derivative term!
Many are tempted to use the chain rule on the term ∂∂x(q(x)∂u∂x), but this is not a good idea when discretizing such a term.
The term with a variable coefficient expresses the net flux qux into a small volume (i.e., interval in 1D):
Our discretization reflects this principle directly: qux at the right end of the cell minus qux at the left end, because this follows from the formula (9) or [Dx(qDxu)]ni.
When using the chain rule, we get two terms quxx+qxux. The typical discretization is
Writing this out shows that it is different from [Dx(qDxu)]ni and lacks the physical interpretation of net flux into a cell. With a smooth and slowly varying q(x) the differences between the two discretizations are not substantial. However, when q exhibits (potentially large) jumps, [Dx(qDxu)]ni with harmonic averaging of q yields a better solution than arithmetic averaging or (11). In the literature, the discretization [Dx(qDxu)]ni totally dominates and very few mention the alternative in (11).
If q is a known function of x, we can easily evaluate qi+12 simply as q(xi+12) with xi+12=xi+12Δx. However, in many cases c, and hence q, is only known as a discrete function, often at the mesh points xi. Evaluating q between two mesh points xi and xi+1 must then be done by interpolation techniques, of which three are of particular interest in this context:
The arithmetic mean in (12) is by far the most commonly used averaging technique and is well suited for smooth q(x) functions. The harmonic mean is often preferred when q(x) exhibits large jumps (which is typical for geological media). The geometric mean is less used, but popular in discretizations to linearize quadratic % if BOOK == "book": nonlinearities (see the section vib:ode2:fdm:fquad for an example). % else: nonlinearities. % endif
With the operator notation from (12) we can specify the discretization of the complete variable-coefficient wave equation in a compact way:
Strictly speaking, [Dx¯qxDxu]ni=[Dx(¯qxDxu)]ni.
From the compact difference notation we immediately see what kind of differences that each term is approximated with. The notation ¯qx also specifies that the variable coefficient is approximated by an arithmetic mean, the definition being [¯qx]i+12=(qi+qi+1)/2.
Before implementing, it remains to solve (15) with respect to un+1i:
The stability criterion derived later (the section wave:pde1:stability) reads Δt≤Δx/c. If c=c(x), the criterion will depend on the spatial location. We must therefore choose a Δt that is small enough such that no mesh cell has Δt>Δx/c(x). That is, we must use the largest c value in the criterion:
The parameter β is included as a safety factor: in some problems with a significantly varying c it turns out that one must choose β<1 to have stable solutions (β=0.9 may act as an all-round value).
A different strategy to handle the stability criterion with variable wave velocity is to use a spatially varying Δt. While the idea is mathematically attractive at first sight, the implementation quickly becomes very complicated, so we stick to a constant Δt and a worst case value of c(x) (with a safety factor β).
Consider a Neumann condition ∂u/∂x=0 at x=L=NxΔx, discretized as
for i=Nx. Using the scheme (16) at the end point i=Nx with uni+1=uni−1 results in
Here we used the approximation
An alternative derivation may apply the arithmetic mean of qn−12 and qn+12 in (19), leading to the term
Since 12(qi+1+qi−1)=qi+O(Δx2), we can approximate with 2qi(uni−1−uni) for i=Nx and get the same term as we did above.
A common technique when implementing ∂u/∂x=0 boundary conditions, is to assume dq/dx=0 as well. This implies qi+1=qi−1 and qi+1/2=qi−1/2 for i=Nx. The implications for the scheme are
for i in range(1, Nx):
u[i] = - u_nm1[i] + 2*u_n[i] + \
C2*(0.5*(q[i] + q[i+1])*(u_n[i+1] - u_n[i]) - \
0.5*(q[i] + q[i-1])*(u_n[i] - u_n[i-1])) + \
dt2*f(x[i], t[n])
The coefficient C2
is now defined as (dt/dx)**2
, i.e., not as the
squared Courant number, since the wave velocity is variable and appears
inside the parenthesis.
With Neumann conditions ux=0 at the
boundary, we need to combine this scheme with the discrete
version of the boundary condition, as shown in the section Neumann condition and a variable coefficient.
Nevertheless, it would be convenient to reuse the formula for the
interior points and just modify the indices ip1=i+1
and im1=i-1
as we did in the section Implementation of Neumann conditions. Assuming
dq/dx=0 at the boundaries, we can implement the scheme at
the boundary with the following code.
i = 0
ip1 = i+1
im1 = ip1
u[i] = - u_nm1[i] + 2*u_n[i] + \
C2*(0.5*(q[i] + q[ip1])*(u_n[ip1] - u_n[i]) - \
0.5*(q[i] + q[im1])*(u_n[i] - u_n[im1])) + \
dt2*f(x[i], t[n])
With ghost cells we can just reuse the formula for the interior points also at the boundary, provided that the ghost values of both u and q are correctly updated to ensure ux=0 and qx=0.
A vectorized version of the scheme with a variable coefficient at internal mesh points becomes
u[1:-1] = - u_nm1[1:-1] + 2*u_n[1:-1] + \
C2*(0.5*(q[1:-1] + q[2:])*(u_n[2:] - u_n[1:-1]) -
0.5*(q[1:-1] + q[:-2])*(u_n[1:-1] - u_n[:-2])) + \
dt2*f(x[1:-1], t[n])
Sometimes a wave PDE has a variable coefficient in front of the time-derivative term:
One example appears when modeling elastic waves in a rod with varying density, cf. (wave:app:string) with ϱ(x).
A natural scheme for (24) is
We realize that the ϱ coefficient poses no particular difficulty, since ϱ enters the formula just as a simple factor in front of a derivative. There is hence no need for any averaging of ϱ. Often, ϱ will be moved to the right-hand side, also without any difficulty:
Waves die out by two mechanisms. In 2D and 3D the energy of the wave spreads out in space, and energy conservation then requires the amplitude to decrease. This effect is not present in 1D. Damping is another cause of amplitude reduction. For example, the vibrations of a string die out because of damping due to air resistance and non-elastic effects in the string.
The simplest way of including damping is to add a first-order derivative to the equation (in the same way as friction forces enter a vibrating mechanical system):
where b≥0 is a prescribed damping coefficient.
A typical discretization of (27) in terms of centered differences reads
Writing out the equation and solving for the unknown un+1i gives the scheme
for i∈Iix and n≥1. New equations must be derived for u1i, and for boundary points in case of Neumann conditions.
The damping is very small in many wave phenomena and thus only evident for very long time simulations. This makes the standard wave equation without damping relevant for a lot of applications.
The program wave1D_dn_vc.py
is a fairly general code for 1D wave propagation problems that
targets the following initial-boundary value problem
The only new feature here is the time-dependent Dirichlet conditions, but they are trivial to implement:
i = mathcal{I}_x[0] # x=0
u[i] = U_0(t[n+1])
i = mathcal{I}_x[-1] # x=L
u[i] = U_L(t[n+1])
The solver
function is a natural extension of the simplest
solver
function in the initial wave1D_u0.py
program,
extended with Neumann boundary conditions (ux=0),
time-varying Dirichlet conditions, as well as
a variable wave velocity. The different code segments needed
to make these extensions have been shown and commented upon in the
preceding text. We refer to the solver
function in the
wave1D_dn_vc.py
file for all the details. Note in that
solver
function, however, that the technique of "hashing" is
used to check whether a certain simulation has been run before, or not.
% if BOOK == 'book':
This technique is further explained in the section softeng2:wave1D:filestorage:hash.
% endif
The vectorization is only applied inside the time loop, not for the initial condition or the first time steps, since this initial work is negligible for long time simulations in 1D problems.
The following sections explain various more advanced programming techniques applied in the general 1D wave equation solver.
A useful feature in the wave1D_dn_vc.py
program is the specification
of the user_action
function as a class. This part of the program may
need some motivation and explanation. Although the plot_u_st
function (and the PlotMatplotlib
class) in the wave1D_u0.viz
function remembers the local variables in the viz
function, it is a
cleaner solution to store the needed variables together with the
function, which is exactly what a class offers.
A class for flexible plotting, cleaning up files, making movie
files, like the function wave1D_u0.viz
did, can be coded as follows:
%matplotlib inline
class PlotAndStoreSolution:
"""
Class for the user_action function in solver.
Visualizes the solution only.
"""
def __init__(
self,
casename='tmp', # Prefix in filenames
umin=-1, umax=1, # Fixed range of y axis
pause_between_frames=None, # Movie speed
backend='matplotlib', # or 'gnuplot' or None
screen_movie=True, # Show movie on screen?
title='', # Extra message in title
skip_frame=1, # Skip every skip_frame frame
filename=None): # Name of file with solutions
self.casename = casename
self.yaxis = [umin, umax]
self.pause = pause_between_frames
self.backend = backend
if backend is None:
# Use native matplotlib
import matplotlib.pyplot as plt
elif backend in ('matplotlib', 'gnuplot'):
module = 'scitools.easyviz.' + backend + '_'
exec('import %s as plt' % module)
self.plt = plt
self.screen_movie = screen_movie
self.title = title
self.skip_frame = skip_frame
self.filename = filename
if filename is not None:
# Store time points when u is written to file
self.t = []
filenames = glob.glob('.' + self.filename + '*.dat.npz')
for filename in filenames:
os.remove(filename)
# Clean up old movie frames
for filename in glob.glob('frame_*.png'):
os.remove(filename)
def __call__(self, u, x, t, n):
"""
Callback function user_action, call by solver:
Store solution, plot on screen and save to file.
"""
# Save solution u to a file using numpy.savez
if self.filename is not None:
name = 'u%04d' % n # array name
kwargs = {name: u}
fname = '.' + self.filename + '_' + name + '.dat'
np.savez(fname, **kwargs)
self.t.append(t[n]) # store corresponding time value
if n == 0: # save x once
np.savez('.' + self.filename + '_x.dat', x=x)
# Animate
if n % self.skip_frame != 0:
return
title = 't=%.3f' % t[n]
if self.title:
title = self.title + ' ' + title
if self.backend is None:
# native matplotlib animation
if n == 0:
self.plt.ion()
self.lines = self.plt.plot(x, u, 'r-')
self.plt.axis([x[0], x[-1],
self.yaxis[0], self.yaxis[1]])
self.plt.xlabel('x')
self.plt.ylabel('u')
self.plt.title(title)
self.plt.legend(['t=%.3f' % t[n]])
else:
# Update new solution
self.lines[0].set_ydata(u)
self.plt.legend(['t=%.3f' % t[n]])
self.plt.draw()
else:
# scitools.easyviz animation
self.plt.plot(x, u, 'r-',
xlabel='x', ylabel='u',
axis=[x[0], x[-1],
self.yaxis[0], self.yaxis[1]],
title=title,
show=self.screen_movie)
# pause
if t[n] == 0:
time.sleep(2) # let initial condition stay 2 s
else:
if self.pause is None:
pause = 0.2 if u.size < 100 else 0
time.sleep(pause)
self.plt.savefig('frame_%04d.png' % (n))
Understanding this class requires quite some familiarity with Python
in general and class programming in particular.
The class supports plotting with Matplotlib (backend=None
) or
SciTools (backend=matplotlib
or backend=gnuplot
) for maximum
flexibility.
The constructor shows how we can flexibly import the plotting engine
as (typically) scitools.easyviz.gnuplot_
or
scitools.easyviz.matplotlib_
(note the trailing underscore - it is required).
With the screen_movie
parameter
we can suppress displaying each movie frame on the screen.
Alternatively, for slow movies associated with
fine meshes, one can set
skip_frame=10
, causing every 10 frames to be shown.
The __call__
method makes PlotAndStoreSolution
instances behave like
functions, so we can just pass an instance, say p
, as the
user_action
argument in the solver
function, and any call to
user_action
will be a call to p.__call__
. The __call__
method plots the solution on the screen,
saves the plot to file, and stores the solution in a file for
later retrieval.
More details on storing the solution in files appear in in the document Scientific software engineering; wave equation case [Langtangen_deqbook_softeng2].
The function pulse
in wave1D_dn_vc.py
demonstrates wave motion in
heterogeneous media where c varies. One can specify an interval
where the wave velocity is decreased by a factor slowness_factor
(or increased by making this factor less than one).
Figure shows a typical simulation
scenario.
Four types of initial conditions are available:
a rectangular pulse (plug
),
a Gaussian function (gaussian
),
a "cosine hat" consisting of one period of the cosine function
(cosinehat
),
frac{1}{2} a period of a "cosine hat" (frac{1}{2}-cosinehat
)
These peak-shaped initial conditions can be placed in the middle
(loc='center'
) or at the left end (loc='left'
) of the domain.
With the pulse in the middle, it splits in two parts, each with frac{1}{2}
the initial amplitude, traveling in opposite directions. With the
pulse at the left end, centered at x=0, and using the symmetry
condition ∂u/∂x=0, only a right-going pulse is
generated. There is also a left-going pulse, but it travels from x=0
in negative x direction and is not visible in the domain [0,L].
The pulse
function is a flexible tool for playing around with
various wave shapes and jumps in the wave velocity (i.e.,
discontinuous media). The code is shown to demonstrate how easy it is
to reach this flexibility with the building blocks we have already
developed:
def pulse(
C=1, # Maximum Courant number
Nx=200, # spatial resolution
animate=True,
version='vectorized',
T=2, # end time
loc='left', # location of initial condition
pulse_thinspace .='gaussian', # pulse/init.cond. type
slowness_factor=2, # inverse of wave vel. in right medium
medium=[0.7, 0.9], # interval for right medium
skip_frame=1, # skip frames in animations
sigma=0.05 # width measure of the pulse
):
"""
Various peaked-shaped initial conditions on [0,1].
Wave velocity is decreased by the slowness_factor inside
medium. The loc parameter can be 'center' or 'left',
depending on where the initial pulse is to be located.
The sigma parameter governs the width of the pulse.
"""
# Use scaled parameters: L=1 for domain length, c_0=1
# for wave velocity outside the domain.
L = 1.0
c_0 = 1.0
if loc == 'center':
xc = L/2
elif loc == 'left':
xc = 0
if pulse_thinspace . in ('gaussian','Gaussian'):
def I(x):
return np.exp(-0.5*((x-xc)/sigma)**2)
elif pulse_thinspace . == 'plug':
def I(x):
return 0 if abs(x-xc) > sigma else 1
elif pulse_thinspace . == 'cosinehat':
def I(x):
# One period of a cosine
w = 2
a = w*sigma
return 0.5*(1 + np.cos(np.pi*(x-xc)/a)) \
if xc - a <= x <= xc + a else 0
elif pulse_thinspace . == 'frac{1}{2}-cosinehat':
def I(x):
# Half a period of a cosine
w = 4
a = w*sigma
return np.cos(np.pi*(x-xc)/a) \
if xc - 0.5*a <= x <= xc + 0.5*a else 0
else:
raise ValueError('Wrong pulse_thinspace .="%s"' % pulse_thinspace .)
def c(x):
return c_0/slowness_factor \
if medium[0] <= x <= medium[1] else c_0
umin=-0.5; umax=1.5*I(xc)
casename = '%s_Nx%s_sf%s' % \
(pulse_thinspace ., Nx, slowness_factor)
action = PlotMediumAndSolution(
medium, casename=casename, umin=umin, umax=umax,
skip_frame=skip_frame, screen_movie=animate,
backend=None, filename='tmpdata')
# Choose the stability limit with given Nx, worst case c
# (lower C will then use this dt, but smaller Nx)
dt = (L/Nx)/c_0
cpu, hashed_input = solver(
I=I, V=None, f=None, c=c,
U_0=None, U_L=None,
L=L, dt=dt, C=C, T=T,
user_action=action,
version=version,
stability_safety_factor=1)
if cpu > 0: # did we generate new data?
action.close_file(hashed_input)
action.make_movie_file()
print 'cpu (-1 means no new data generated):', cpu
def convergence_rates(
u_exact,
I, V, f, c, U_0, U_L, L,
dt0, num_meshes,
C, T, version='scalar',
stability_safety_factor=1.0):
"""
Half the time step and estimate convergence rates for
for num_meshes simulations.
"""
class ComputeError:
def __init__(self, norm_type):
self.error = 0
def __call__(self, u, x, t, n):
"""Store norm of the error in self.E."""
error = np.abs(u - u_exact(x, t[n])).max()
self.error = max(self.error, error)
E = []
h = [] # dt, solver adjusts dx such that C=dt*c/dx
dt = dt0
for i in range(num_meshes):
error_calculator = ComputeError('Linf')
solver(I, V, f, c, U_0, U_L, L, dt, C, T,
user_action=error_calculator,
version='scalar',
stability_safety_factor=1.0)
E.append(error_calculator.error)
h.append(dt)
dt /= 2 # halve the time step for next simulation
print 'E:', E
print 'h:', h
r = [np.log(E[i]/E[i-1])/np.log(h[i]/h[i-1])
for i in range(1,num_meshes)]
return r
def test_convrate_sincos():
n = m = 2
L = 1.0
u_exact = lambda x, t: np.cos(m*np.pi/L*t)*np.sin(m*np.pi/L*x)
r = convergence_rates(
u_exact=u_exact,
I=lambda x: u_exact(x, 0),
V=lambda x: 0,
f=0,
c=1,
U_0=0,
U_L=0,
L=L,
dt0=0.1,
num_meshes=6,
C=0.9,
T=1,
version='scalar',
stability_safety_factor=1.0)
print 'rates sin(x)*cos(t) solution:', \
[round(r_,2) for r_ in r]
assert abs(r[-1] - 2) < 0.002
The PlotMediumAndSolution
class used here is a subclass of
PlotAndStoreSolution
where the medium with reduced c value,
as specified by the medium
interval,
is visualized in the plots.
Comment on the choices of discretization parameters.
The argument Nx in the pulse
function does not correspond to
the actual spatial resolution of C<1, since the solver
function takes a fixed Δt and C, and adjusts Δx
accordingly. As seen in the pulse
function,
the specified Δt is chosen according to the
limit C=1, so if C<1, Δt remains the same, but the
solver
function operates with a larger Δx and smaller
Nx than was specified in the call to pulse
. The practical reason
is that we always want to keep Δt fixed such that
plot frames and movies are synchronized in time regardless of the
value of C (i.e., Δx is varied when the
Courant number varies).
The reader is encouraged to play around with the pulse
function:
To easily kill the graphics by Ctrl-C and restart a new simulation it might be easier to run the above two statements from the command line with
Terminal> python -c 'import wave1D_dn_vc as w; w.pulse(...)'
Consider the wave equation with damping (27). The goal is to find an exact solution to a wave problem with damping and zero source term. A starting point is the standing wave solution from wave:exer:standingwave. mathcal{I}_t becomes necessary to include a damping term e−βt and also have both a sine and cosine component in time:
Find k from the boundary conditions u(0,t)=u(L,t)=0. Then use the PDE to find constraints on β, ω, A, and B. Set up a complete initial-boundary value problem and its solution.
Solution. Mathematical model:
b≥0 is a prescribed damping coefficient.
Ansatz:
Boundary condition: u=0 for x=0,L. Fulfilled for x=0. Requirement at x=L gives
for an arbitrary integer m. Hence, k=mπ/L.
Inserting the ansatz in the PDE and dividing by e−βt results in
This gives us two requirements:
and
Since b, c and k are to be given in advance, we may solve these two equations to get
From the initial condition on the derivative, i.e. ∂ue∂t=0, we find that
Inserting the expression for ω, we find that
for A prescribed.
Using t=0 in the expression for ue gives us the initial condition as
Summarizing, the PDE problem can then be states as
where constants c, A, b and k, as well as I(x), are prescribed.
The solution to the problem is then given as
with k=mπ/L for arbitrary integer m, β=b2, ω=√c2k2−b24, B=b2√c2k2−b24A and I(x)=Asinkx.
Filename: damped_waves
.
Consider the simple "plug" wave where Ω=[−L,L] and
for some number 0<δ<L. The other initial condition is ut(x,0)=0 and there is no source term f. The boundary conditions can be set to u=0. The solution to this problem is symmetric around x=0. This means that we can simulate the wave process in only frac{1}{2} of the domain [0,L].
a) Argue why the symmetry boundary condition is ux=0 at x=0.
Hint. Symmetry of a function about x=x0 means that f(x0+h)=f(x0−h).
Solution. A symmetric u around x=0 means that u(−x,t)=u(x,t). Let x0=0 and x=x0+h. Then we can use a centered finite difference definition of the derivative:
since u(h,t)=u(−h,t) for any h. Symmetry around a point x=x0 therefore always implies ux(x0,t)=0.
b) Perform simulations of the complete wave problem on [−L,L]. Thereafter, utilize the symmetry of the solution and run a simulation in frac{1}{2} of the domain [0,L], using a boundary condition at x=0. Compare plots from the two solutions and confirm that they are the same.
Solution.
We can utilize the wave1D_dn.py
code which allows Dirichlet and
Neumann conditions. The solver
and viz
functions must take x0
and xL as parameters instead of just L such that we can solve the
wave equation in [x0,xL]. The we can call up solver
for the two
problems on [−L,L] and [0,L] with boundary conditions
u(−L,t)=u(L,t)=0 and ux(0,t)=u(L,t)=0, respectively.
The original wave1D_dn.py
code makes a movie by playing all the
.png
files in a browser. mathcal{I}_t can then be wise to let the viz
function create a movie directory and place all the frames and HTML
player file in that directory. Alternatively, one can just make
some ordinary movie file (Ogg, WebM, MP4, Flash) with ffmpeg
or
ffmpeg
and give it a name. mathcal{I}_t is a point that the name is
transferred to viz
so it is easy to call viz
twice and get two
separate movie files or movie directories.
The plots produced by the code (below) shows that the solutions indeed are the same.
c) Prove the symmetry property of the solution by setting up the complete initial-boundary value problem and showing that if u(x,t) is a solution, then also u(−x,t) is a solution.
Solution. The plan in this proof is to introduce v(x,t)=u(−x,t) and show that v fulfills the same initial-boundary value problem as u. If the problem has a unique solution, then v=u. Or, in other words, the solution is symmetric: u(−x,t)=u(x,t).
We can work with a general initial-boundary value problem on the form
Introduce a new coordinate ˉx=−x. We have that
The derivatives in time are unchanged.
Substituting x by −ˉx leads to
Now, dropping the bars and introducing v(x,t)=u(−x,t), we find that
Provided that I, f, and V are all symmetric around x=0 such that I(x)=I(−x), V(x)=V(−x), and f(x,t)=f(−x,t), we can express the initial-boundary value problem as
This is the same problem as the one that u fulfills. If the solution is unique, which can be proven, then v=u, and u(−x,t)=u(x,t).
To summarize, the necessary conditions for symmetry are that
all involved functions I, V, and f must be symmetric, and
the boundary conditions are symmetric in the sense that they can be flipped (the condition at x=−L can be applied at x=L and vice versa).
d)
If the code works correctly, the solution u(x,t)=x(L−x)(1+t2)
should be reproduced exactly. Write a test function test_quadratic
that
checks whether this is the case. Simulate for x in [0,L2] with
a symmetry condition at the end x=L2.
Solution. Running the code below, shows that the test case indeed is reproduced exactly.
#!/usr/bin/env python
from scitools.std import *
# Add an x0 coordinate for solving the wave equation on [x0, xL]
def solver(I, V, f, c, U_0, U_L, x0, xL, Nx, C, T,
user_action=None, version='scalar'):
"""
Solve u_tt=c^2*u_xx + f on (0,L)x(0,T].
u(0,t)=U_0(t) or du/dn=0 (U_0=None), u(L,t)=U_L(t) or du/dn=0 (u_L=None).
"""
x = linspace(x0, xL, Nx+1) # Mesh points in space
dx = x[1] - x[0]
dt = C*dx/c
Nt = int(round(T/dt))
t = linspace(0, Nt*dt, Nt+1) # Mesh points in time
C2 = C**2; dt2 = dt*dt # Help variables in the scheme
# Wrap user-given f, V, U_0, U_L
if f is None or f == 0:
f = (lambda x, t: 0) if version == 'scalar' else \
lambda x, t: zeros(x.shape)
if V is None or V == 0:
V = (lambda x: 0) if version == 'scalar' else \
lambda x: zeros(x.shape)
if U_0 is not None:
if isinstance(U_0, (float,int)) and U_0 == 0:
U_0 = lambda t: 0
if U_L is not None:
if isinstance(U_L, (float,int)) and U_L == 0:
U_L = lambda t: 0
u = zeros(Nx+1) # Solution array at new time level
u_1 = zeros(Nx+1) # Solution at 1 time level back
u_2 = zeros(Nx+1) # Solution at 2 time levels back
mathcal{I}_x = range(0, Nx+1)
mathcal{I}_t = range(0, Nt+1)
import time; t0 = time.clock() # CPU time measurement
# Load initial condition into u_1
for i in mathcal{I}_x:
u_1[i] = I(x[i])
if user_action is not None:
user_action(u_1, x, t, 0)
# Special formula for the first step
for i in mathcal{I}_x[1:-1]:
u[i] = u_1[i] + dt*V(x[i]) + \
0.5*C2*(u_1[i-1] - 2*u_1[i] + u_1[i+1]) + \
0.5*dt2*f(x[i], t[0])
i = mathcal{I}_x[0]
if U_0 is None:
# Set boundary values du/dn = 0
# x=0: i-1 -> i+1 since u[i-1]=u[i+1]
# x=L: i+1 -> i-1 since u[i+1]=u[i-1])
ip1 = i+1
im1 = ip1 # i-1 -> i+1
u[i] = u_1[i] + dt*V(x[i]) + \
0.5*C2*(u_1[im1] - 2*u_1[i] + u_1[ip1]) + \
0.5*dt2*f(x[i], t[0])
else:
u[0] = U_0(dt)
i = mathcal{I}_x[-1]
if U_L is None:
im1 = i-1
ip1 = im1 # i+1 -> i-1
u[i] = u_1[i] + dt*V(x[i]) + \
0.5*C2*(u_1[im1] - 2*u_1[i] + u_1[ip1]) + \
0.5*dt2*f(x[i], t[0])
else:
u[i] = U_L(dt)
if user_action is not None:
user_action(u, x, t, 1)
# Update data structures for next step
u_2[:], u_1[:] = u_1, u
for n in mathcal{I}_t[1:-1]:
# Update all inner points
if version == 'scalar':
for i in mathcal{I}_x[1:-1]:
u[i] = - u_2[i] + 2*u_1[i] + \
C2*(u_1[i-1] - 2*u_1[i] + u_1[i+1]) + \
dt2*f(x[i], t[n])
elif version == 'vectorized':
u[1:-1] = - u_2[1:-1] + 2*u_1[1:-1] + \
C2*(u_1[0:-2] - 2*u_1[1:-1] + u_1[2:]) + \
dt2*f(x[1:-1], t[n])
else:
raise ValueError('version=%s' % version)
# Insert boundary conditions
i = mathcal{I}_x[0]
if U_0 is None:
# Set boundary values
# x=0: i-1 -> i+1 since u[i-1]=u[i+1] when du/dn=0
# x=L: i+1 -> i-1 since u[i+1]=u[i-1] when du/dn=0
ip1 = i+1
im1 = ip1
u[i] = - u_2[i] + 2*u_1[i] + \
C2*(u_1[im1] - 2*u_1[i] + u_1[ip1]) + \
dt2*f(x[i], t[n])
else:
u[0] = U_0(t[n+1])
i = mathcal{I}_x[-1]
if U_L is None:
im1 = i-1
ip1 = im1
u[i] = - u_2[i] + 2*u_1[i] + \
C2*(u_1[im1] - 2*u_1[i] + u_1[ip1]) + \
dt2*f(x[i], t[n])
else:
u[i] = U_L(t[n+1])
if user_action is not None:
if user_action(u, x, t, n+1):
break
# Update data structures for next step
u_2[:], u_1[:] = u_1, u
cpu_time = t0 - time.clock()
return u, x, t, cpu_time
def viz(I, V, f, c, U_0, U_L, x0, xL, Nx, C, T, umin, umax,
version='scalar', animate=True,
movie_dir='tmp'):
"""Run solver and visualize u at each time level."""
import scitools.std as plt, time, glob, os
def plot_u(u, x, t, n):
"""user_action function for solver."""
plt.plot(x, u, 'r-',
xlabel='x', ylabel='u',
axis=[x0, xL, umin, umax],
title='t=%f' % t[n])
# Let the initial condition stay on the screen for 2
# seconds, else insert a pause of 0.2 s between each plot
time.sleep(2) if t[n] == 0 else time.sleep(0.2)
plt.savefig('frame_%04d.png' % n) # for movie making
# Clean up old movie frames
for filename in glob.glob('frame_*.png'):
os.remove(filename)
user_action = plot_u if animate else None
u, x, t, cpu = solver(I, V, f, c, U_0, U_L, L, Nx, C, T,
user_action, version)
if animate:
# Make a directory with the frames
if os.path.isdir(movie_dir):
shutil.rmtree(movie_dir)
os.mkdir(movie_dir)
os.chdir(movie_dir)
# Move all frame_*.png files to this subdirectory
for filename in glob.glob(os.path.join(os.pardir, 'frame_*.png')):
os.renamve(os.path.join(os.pardir, filename), filename)
plt.movie('frame_*.png', encoder='html', fps=4,
output_file='movie.html')
# Invoke movie.html in a browser to steer the movie
return cpu
import nose.tools as nt
def test_quadratic():
"""
Check the scalar and vectorized versions work for
a quadratic u(x,t)=x(L-x)(1+t/2) that is exactly reproduced.
We simulate in [0, L/2] and apply a symmetry condition
at the end x=L/2.
"""
exact_solution = lambda x, t: x*(L-x)*(1+0.5*t)
I = lambda x: exact_solution(x, 0)
V = lambda x: 0.5*exact_solution(x, 0)
f = lambda x, t: 2*(1+0.5*t)*c**2
U_0 = lambda t: exact_solution(0, t)
U_L = None
L = 2.5
c = 1.5
Nx = 3 # very coarse mesh
C = 1
T = 18 # long time integration
def assert_no_error(u, x, t, n):
u_e = exact_solution(x, t[n])
diff = abs(u - u_e).max()
nt.assert_almost_equal(diff, 0, places=13)
solver(I, V, f, c, U_0, U_L, 0, L/2, Nx, C, T,
user_action=assert_no_error, version='scalar')
solver(I, V, f, c, U_0, U_L, 0, L/2, Nx, C, T,
user_action=assert_no_error, version='vectorized')
def plug(C=1, Nx=50, animate=True, version='scalar', T=2):
"""Plug profile as initial condition."""
L = 1.
c = 1
delta = 0.1
def I(x):
if abs(x) > delta:
return 0
else:
return 1
# Solution on [-L,L]
cpu = viz(I=I, V=0, f=0, c, U_0=0, U_L=0, -L, L, 2*Nx, C, T,
umin=-1.1, umax=1.1, version=version, animate=animate,
movie_dir='full')
# Solution on [0,L]
cpu = viz(I=I, V=0, f=0, c, U_0=None, U_L=0, 0, L, Nx, C, T,
umin=-1.1, umax=1.1, version=version, animate=animate,
movie_dir='frac{1}{2}')
if __name__ == '__main__':
plug()
Filename: wave1D_symmetric
.
Use the pulse
function in wave1D_dn_vc.py
to investigate
sending a pulse, located with its peak at x=0, through two
media with different wave velocities. The (scaled) velocity in
the left medium is 1 while it is 1sf in the right medium.
Report what happens with a Gaussian pulse, a "cosine hat" pulse,
frac{1}{2} a "cosine hat" pulse, and a plug pulse for resolutions
Nx=40,80,160, and sf=2,4. Simulate until T=2.
Solution. In all cases, the change in velocity causes some of the wave to be reflected back (while the rest is let through). When the waves go from higher to lower velocity, the amplitude builds, and vice versa.
import wave1D_dn_vc as wave
import os, sys, shutil, glob
for pulse_thinspace . in 'gaussian', 'cosinehat', 'frac{1}{2}-cosinehat', 'plug':
for Nx in 40, 80, 160:
for sf in 2, 4:
if sf == 1 and Nx > 40:
continue # homogeneous medium with C=1: Nx=40 enough
print 'wave1D.pulse:', pulse_thinspace ., Nx, sf
wave.pulse(C=1, Nx=Nx, animate=False, # just hardcopies
version='vectorized',
T=2, loc='left', pulse_thinspace .=pulse_thinspace .,
slowness_factor=sf, medium=[0.7, 0.9],
skip_frame = 1,
sigma=0.05)
Filename: pulse1D
.
The experiments performed in Exercise 3: Send pulse waves through a layered medium shows
considerable numerical noise in the form of non-physical waves,
especially for sf=4 and the plug pulse or the frac{1}{2} a "cosinehat"
pulse. The noise is much less visible for a Gaussian pulse. Run the
case with the plug and frac{1}{2} a "cosinehat" pulse for sf=1, C=0.9,0.25, and Nx=40,80,160. Use the numerical dispersion relation to
explain the observations.
Filename: pulse1D_analysis
.
Harmonic means are often used if the wave velocity is non-smooth or
discontinuous. Will harmonic averaging of the wave velocity give less
numerical noise for the case sf=4 in Exercise 3: Send pulse waves through a layered medium?
Filename: pulse1D_harmonic
.
To enable a wave to leave the computational domain and travel undisturbed through the boundary x=L, one can in a one-dimensional problem impose the following condition, called a radiation condition or open boundary condition:
The parameter c is the wave velocity.
Show that (55) accepts a solution u=gR(x−ct) (right-going wave), but not u=gL(x+ct) (left-going wave). This means that (55) will allow any right-going wave gR(x−ct) to pass through the boundary undisturbed.
A corresponding open boundary condition for a left-going wave through x=0 is
a) A natural idea for discretizing the condition (55) at the spatial end point i=Nx is to apply centered differences in time and space:
Eliminate the fictitious value unNx+1 by using the discrete equation at the same point.
The equation for the first step, u1i, is in principle also affected, but we can then use the condition uNx=0 since the wave has not yet reached the right boundary.
b) A much more convenient implementation of the open boundary condition at x=L can be based on an explicit discretization
From this equation, one can solve for un+1Nx and apply the formula as a Dirichlet condition at the boundary point. However, the finite difference approximations involved are of first order.
Implement this scheme for a wave equation utt=c2uxx in a domain [0,L], where you have ux=0 at x=0, the condition (55) at x=L, and an initial disturbance in the middle of the domain, e.g., a plug profile like
Observe that the initial wave is split in two, the left-going wave is reflected at x=0, and both waves travel out of x=L, leaving the solution as u=0 in [0,L]. Use a unit Courant number such that the numerical solution is exact. Make a movie to illustrate what happens.
Because this simplified implementation of the open boundary condition works, there is no need to pursue the more complicated discretization in a).
Hint.
Modify the solver function in
wave1D_dn.py
.
c) Add the possibility to have either ux=0 or an open boundary condition at the left boundary. The latter condition is discretized as
leading to an explicit update of the boundary value un+10.
The implementation can be tested with a Gaussian function as initial condition:
Run two tests:
Disturbance in the middle of the domain, I(x)=g(x;L/2,s), and open boundary condition at the left end.
Disturbance at the left end, I(x)=g(x;0,s), and ux=0 as symmetry boundary condition at this end.
Make test functions for both cases, testing that the solution is zero after the waves have left the domain.
d) In 2D and 3D it is difficult to compute the correct wave velocity normal to the boundary, which is needed in generalizations of the open boundary conditions in higher dimensions. Test the effect of having a slightly wrong wave velocity in (58). Make movies to illustrate what happens.
Filename: wave1D_open_BC
.
The condition (55) works perfectly in 1D when c is known. In 2D and 3D, however, the condition reads ut+cxux+cyuy=0, where cx and cy are the wave speeds in the x and y directions. Estimating these components (i.e., the direction of the wave) is often challenging. Other methods are normally used in 2D and 3D to let waves move out of a computational domain.
mathcal{I}_t is frequently of interest to follow wave motion over large distances and long times. A straightforward approach is to work with a very large domain, but that might lead to a lot of computations in areas of the domain where the waves cannot be noticed. A more efficient approach is to let a right-going wave out of the domain and at the same time let it enter the domain on the left. This is called a periodic boundary condition.
The boundary condition at the right end x=L is an open boundary condition (see Problem 6: Implement open boundary conditions) to let a right-going wave out of the domain. At the left end, x=0, we apply, in the beginning of the simulation, either a symmetry boundary condition (see Problem 2: Explore symmetry boundary conditions) ux=0, or an open boundary condition.
This initial wave will split in two and either be reflected or transported out of the domain at x=0. The purpose of the exercise is to follow the right-going wave. We can do that with a periodic boundary condition. This means that when the right-going wave hits the boundary x=L, the open boundary condition lets the wave out of the domain, but at the same time we use a boundary condition on the left end x=0 that feeds the outgoing wave into the domain again. This periodic condition is simply u(0)=u(L). The switch from ux=0 or an open boundary condition at the left end to a periodic condition can happen when u(L,t)>ϵ, where ϵ=10−4 might be an appropriate value for determining when the right-going wave hits the boundary x=L.
The open boundary conditions can conveniently be discretized as
explained in Problem 6: Implement open boundary conditions. Implement the
described type of boundary conditions and test them on two different
initial shapes: a plug u(x,0)=1 for x≤0.1, u(x,0)=0 for
x>0.1, and a Gaussian function in the middle of the domain:
u(x,0)=exp(−12(x−0.5)2/0.05). The domain is the unit
interval [0,1]. Run these two shapes for Courant numbers 1 and
0.5. Assume constant wave velocity. Make movies of the four cases.
Reason why the solutions are correct.
Filename: periodic
.
We have a 1D wave equation with variable wave velocity: utt=(qux)x. A Neumann condition ux at x=0,L can be discretized as shown in (20) and (23).
The aim of this exercise is to examine the rate of the numerical error when using different ways of discretizing the Neumann condition.
a) As a test problem, q=1+(x−L/2)4 can be used, with f(x,t) adapted such that the solution has a simple form, say u(x,t)=cos(πx/L)cos(ωt) for, e.g., ω=1. Perform numerical experiments and find the convergence rate of the error using the approximation (20).
b) Switch to q(x)=1+cos(πx/L), which is symmetric at x=0,L, and check the convergence rate of the scheme (23). Now, qi−1/2 is a 2nd-order approximation to qi, qi−1/2=qi+0.25q″iΔx2+⋯, because q′i=0 for i=Nx (a similar argument can be applied to the case i=0).
c) A third discretization can be based on a simple and convenient, but less accurate, one-sided difference: ui−ui−1=0 at i=Nx and ui+1−ui=0 at i=0. Derive the resulting scheme in detail and implement it. Run experiments with q from a) or b) to establish the rate of convergence of the scheme.
d) A fourth technique is to view the scheme as
and place the boundary at xi+12, i=Nx, instead of exactly at the physical boundary. With this idea of approximating (moving) the boundary, we can just set [qDxu]ni+12=0. Derive the complete scheme using this technique. The implementation of the boundary condition at L−Δx/2 is \OofΔx2 accurate, but the interesting question is what impact the movement of the boundary has on the convergence rate. Compute the errors as usual over the entire mesh and use q from a) or b).
Filename: Neumann_discr
.
The purpose of this exercise is to verify the implementation of the
solver
function in the program wave1D_n0.py
by using an exact numerical solution
for the wave equation utt=c2uxx+f with Neumann boundary
conditions ux(0,t)=ux(L,t)=0.
A similar verification is used in the file wave1D_u0.py
, which solves the same PDE, but with
Dirichlet boundary conditions u(0,t)=u(L,t)=0. The idea of the
verification test in function test_quadratic
in wave1D_u0.py
is to
produce a solution that is a lower-order polynomial such that both the
PDE problem, the boundary conditions, and all the discrete equations
are exactly fulfilled. Then the solver
function should reproduce
this exact solution to machine precision. More precisely, we seek
u=X(x)T(t), with T(t) as a linear function and X(x) as a
parabola that fulfills the boundary conditions. Inserting this u in
the PDE determines f. mathcal{I}_t turns out that u also fulfills the
discrete equations, because the truncation error of the discretized
PDE has derivatives in x and t of order four and higher. These
derivatives all vanish for a quadratic X(x) and linear T(t).
mathcal{I}_t would be attractive to use a similar approach in the case of Neumann conditions. We set u=X(x)T(t) and seek lower-order polynomials X and T. To force ux to vanish at the boundary, we let Xx be a parabola. Then X is a cubic polynomial. The fourth-order derivative of a cubic polynomial vanishes, so u=X(x)T(t) will fulfill the discretized PDE also in this case, if f is adjusted such that u fulfills the PDE.
However, the discrete boundary condition is not exactly fulfilled by this choice of u. The reason is that
At the two boundary points, we must demand that the derivative Xx(x)=0 such that ux=0. However, uxxx is a constant and not zero when X(x) is a cubic polynomial. Therefore, our u=X(x)T(t) fulfills
and not
as it should. (Note that all the higher-order terms \OofΔx4 also have higher-order derivatives that vanish for a cubic polynomial.) So to summarize, the fundamental problem is that u as a product of a cubic polynomial and a linear or quadratic polynomial in time is not an exact solution of the discrete boundary conditions.
To make progress,
we assume that u=X(x)T(t), where T for simplicity is taken as a
prescribed linear function 1+12t, and X(x) is taken
as an unknown cubic polynomial ∑3j=0ajxj.
There are two different ways of determining the coefficients
a0,…,a3 such that both the discretized PDE and the
discretized boundary conditions are fulfilled, under the
constraint that we can specify a function f(x,t) for the PDE to feed
to the solver
function in wave1D_n0.py
. Both approaches
are explained in the subexercises.
a) One can insert u in the discretized PDE and find the corresponding f. Then one can insert u in the discretized boundary conditions. This yields two equations for the four coefficients a0,…,a3. To find the coefficients, one can set a0=0 and a1=1 for simplicity and then determine a2 and a3. This approach will make a2 and a3 depend on Δx and f will depend on both Δx and Δt.
Use sympy
to perform analytical computations.
A starting point is to define u as follows:
def test_cubic1():
import sympy as sm
x, t, c, L, dx, dt = sm.symbols('x t c L dx dt')
i, n = sm.symbols('i n', integer=True)
# Assume discrete solution is a polynomial of degree 3 in x
T = lambda t: 1 + sm.Rational(1,2)*t # Temporal term
a = sm.symbols('a_0 a_1 a_2 a_3')
X = lambda x: sum(a[q]*x**q for q in range(4)) # Spatial term
u = lambda x, t: X(x)*T(t)
The symbolic expression for u is reached by calling u(x,t)
with x
and t
as sympy
symbols.
Define DxDx(u, i, n)
, DtDt(u, i, n)
, and D2x(u, i, n)
as Python functions for returning the difference
approximations [DxDxu]ni, [DtDtu]ni, and
[D2xu]ni. The next step is to set up the residuals
for the equations [D2xu]n0=0 and [D2xu]nNx=0,
where Nx=L/Δx. Call the residuals R_0
and R_L
.
Substitute a0 and a1 by 0 and 1, respectively, in
R_0
, R_L
, and a
:
R_0 = R_0.subs(a[0], 0).subs(a[1], 1)
R_L = R_L.subs(a[0], 0).subs(a[1], 1)
a = list(a) # enable in-place assignment
a[0:2] = 0, 1
Determining a2 and a3 from the discretized boundary conditions
is then about solving two equations with respect to a2 and a3,
i.e., a[2:]
:
s = sm.solve([R_0, R_L], a[2:])
# s is dictionary with the unknowns a[2] and a[3] as keys
a[2:] = s[a[2]], s[a[3]]
Now, a
contains computed values and u
will automatically use
these new values since X
accesses a
.
Compute the source term f from the discretized PDE:
fni=[DtDtu−c2DxDxu]ni. Turn u, the time
derivative ut (needed for the initial condition V(x)),
and f into Python functions. Set numerical values for
L, Nx, C, and c. Prescribe the time interval as
Δt=CL/(Nxc), which imply Δx=cΔt/C=L/Nx.
Define new functions I(x)
, V(x)
, and f(x,t)
as wrappers of the ones
made above, where fixed values of L, c, Δx, and Δt
are inserted, such that I
, V
, and f
can be passed on to the
solver
function. Finally, call solver
with a user_action
function that compares the numerical solution to this exact
solution u of the discrete PDE problem.
Hint.
To turn a sympy
expression e
, depending on a series of
symbols, say x
, t
, dx
, dt
, L
, and c
, into a plain
Python function e_exact(x,t,L,dx,dt,c)
, one can write
e_exact = sm.lambdify([x,t,L,dx,dt,c], e, 'numpy')
The 'numpy'
argument is a good habit as the e_exact
function
will then work with array arguments if it contains mathematical
functions (but here we only do plain arithmetics, which automatically
work with arrays).
b) An alternative way of determining a0,…,a3 is to reason as follows. We first construct X(x) such that the boundary conditions are fulfilled: X=x(L−x). However, to compensate for the fact that this choice of X does not fulfill the discrete boundary condition, we seek u such that
since this u will fit the discrete boundary condition.
Assuming u=T(t)∑3j=0ajxj, we can use the above equation to
determine the coefficients a1,a2,a3. A value, e.g., 1 can be used for
a0. The following sympy
code computes this u:
def test_cubic2():
import sympy as sm
x, t, c, L, dx = sm.symbols('x t c L dx')
T = lambda t: 1 + sm.Rational(1,2)*t # Temporal term
# Set u as a 3rd-degree polynomial in space
X = lambda x: sum(a[i]*x**i for i in range(4))
a = sm.symbols('a_0 a_1 a_2 a_3')
u = lambda x, t: X(x)*T(t)
# Force discrete boundary condition to be zero by adding
# a correction term the analytical suggestion x*(L-x)*T
# u_x = x*(L-x)*T(t) - 1/6*u_xxx*dx**2
R = sm.diff(u(x,t), x) - (
x*(L-x) - sm.Rational(1,6)*sm.diff(u(x,t), x, x, x)*dx**2)
# R is a polynomial: force all coefficients to vanish.
# Turn R to Poly to extract coefficients:
R = sm.poly(R, x)
coeff = R.all_coeffs()
s = sm.solve(coeff, a[1:]) # a[0] is not present in R
# s is dictionary with a[i] as keys
# Fix a[0] as 1
s[a[0]] = 1
X = lambda x: sm.simplify(sum(s[a[i]]*x**i for i in range(4)))
u = lambda x, t: X(x)*T(t)
print 'u:', u(x,t)
The next step is to find the source term f_e
by inserting u_e
in the PDE. Thereafter, turn u
, f
, and the time derivative of u
into plain Python functions as in a), and then wrap these functions
in new functions I
, V
, and f
, with the right signature as
required by the solver
function. Set parameters as in a) and
check that the solution is exact to machine precision at each
time level using an appropriate user_action
function.
Filename: wave1D_n0_test_cubic
.