A python package for Krylov and Magnus-Krylov based time propagation.
Project description
mkprop
by Tobias Jawecki
This package provides some Python routines for adaptive time integration based on Magnus-Krylov and Krylov methods.
This package is still under development. Future versions might have different API. The current version is not fully documented.
Getting started
Install the latest release from the pypi package manager python -m pip install mkprop
Mathematical problem description
Let $\psi(t)\in\mathbb{C}^n$ refer to the solution of the system of ODE's $$\psi'(t)=\mathrm{i}H\psi(t),\qquad \psi(t_0)\in\mathbb{C}^{n},$$ where $H\in\mathbb{C}^{n\times n}$, $t\in\mathbb{R}$, and $t_0\in\mathbb{R}$ is a given initial time. Then $\psi(t_0+t)$ for some time-step $t$ is given by the action of the matrix exponential $$\psi(t_0+t) = \exp(\mathrm{i}tH)\psi(t_0).$$ For large $n$ the matrix exponential of $\mathrm{i}tH$ can not be computed directly. The present package provides Krylov methods to compute the action of the matrix exponential, perfcetly fitting to the case that $n$ is large and the action of the matrix $\psi \mapsto H\psi$ is available.
In a similar manner, we also consider non-autonomous system of ODE's. The time propagation for the solution $\psi(\tau)\in\mathbb{C}^n$ of the system of ODE's $$\psi'(t)=\mathrm{i}H(t)\psi(t),\qquad \psi(t_0)\in\mathbb{C}^{n},$$ where $H(t)\in\mathbb{C}^{n\times n}$ depends on the time $t$, can be described by a matrix exponential of the Magnus expansion $\Omega(t,t_0)$. Namely, $$\psi(t_0+t) = \exp(\mathrm{i}\Omega(t,t_0))\psi(t_0).$$ The Magnus expansion corresponds to an infinite series, $$\Omega(t,t_0)=\sum_{j=1}^\infty\Omega_j(t,t_0) = \int_{t_0}^{t_0+t} H(s) ,\mathrm{d}s + \int_{t_0}^{t_0+t} \int_{t_0}^{t_0+s_1} [H(s_1),H(s_2)] ,\mathrm{d}s_2 ,\mathrm{d}s_1 + \ldots ,$$ where the matrix commutator is defined by $[A,B] = AB-BA\in\mathbb{C}^{n\times n}$ for two matrices $A,B\in\mathbb{C}^{n\times n}$. This infinite series converges for a sufficiently small time step $t$. In the simple case of $H$ having the commutator $[H(t),H(s)]=0$ for all times $t,s\in\mathbb{R}$, the Magnus expansion simplifies to $\Omega(t,t_0)=\int_{t_0}^{t_0+t} H(s) ,\mathrm{d}s$. Magnus integrators refer to methods which make use of approximations to this expansions, to approximate the time propagation $\psi(t_0)$ to $\psi(t_0+t)$. Typically, Magnus integrators first provide an approximation $$S(t,t_0) \approx \exp(\mathrm{i}\Omega(t,t_0))$$
Krylov methods
The routine expimv_pKry
provides an approximation to the action of the matrix exponential on a vector $u\in\mathbb{C}^n$,
$$y_K(t,H,u) \approx \exp(\mathrm{i}tH)u.$$
The approximation is computed to satisfy the error bound
$$\lVert y_K(t,H,u) -\exp(\mathrm{i}tH)u\rVert\leq \varepsilon t,$$
where $\varepsilon>0$ is a given tolerance.
The approximation is using Krylov method (Arnoldi or Lanczos) and requires the matrix $H$ or the action of the matrix $\psi \mapsto H\psi$.
usage: examples/basicKrylov.ipynb
import scipy.sparse
import scipy.linalg
import numpy as np
import mkprop
nrm = lambda x : np.linalg.norm(x,2)
n = 1000
e = np.ones(n)
e1 = np.ones(n-1)+1e-5*np.random.rand(n-1)
# M is finite difference discretization of 1d Laplace operator
M = scipy.sparse.diags([e1.conj(),-2*e,e1], [-1,0,1])
u=np.random.rand(n)
u=u/nrm(u)
dt = 50
tol = 1e-6
yref = scipy.sparse.linalg.expm_multiply(1j*dt*M,u)
y,_,_,mused = mkprop.expimv_pKry(M,u,t=dt,tol=tol)
print("approximation error = %.2e, tolerance = %.2e" % (nrm(yref-y)/dt, tol))
# output: approximation error = 5.06e-08, tolerance = 1.00e-06
Magnus integrators
The solution of the non-autonomous system of ODE's is given by a Magnus expansion $\psi(t_0+t) = \exp(\mathrm{i}t\Omega(t,t_0))\psi(t_0)$ which exists for sufficiently small time steps. The matrix $\Omega(t,t_0)$ corresponds to an infinite sum of integrals over terms depending on $H$ and commutators of $H$ evaluated at different times. This series can be truncated to construct approximations to the Magnus expansion, and integrals can be approximated by quadrature rules. E.g., for an initial vector $u\in\mathbb{C}^n$, applying the midpoint quadrature rule to the first term of the Magnus expansion we arrive at $$\exp(\mathrm{i}t\Omega(t,t_0))u\approx\exp(\mathrm{i}tH(t_0+t/2))u,$$ which is a good approximation for a sufficiently small time-step $t$. A very general approach consists of avoiding commutator terms in higher order approximations. E.g., the fourth order method $$\exp(\mathrm{i}t\Omega(t,t_0))u\approx\exp(\mathrm{i}tB_2(t,t_0))\exp(\mathrm{i}tB_1(t,t_0))u,$$ where $B_1$ and $B_2$ correspond to linear combinitions of $H$ evaluated at different times. In general, higher order methods allow taking larger time steps $t$.
Commutator free Magnus (CFM) integrators are of the form $S(t,t_0)=\exp(\mathrm{i}tB_J(t,t_0))\cdots\exp(\mathrm{i}tB_1(t,t_0)),$ where $$B_j=\sum_{k=1}^Ka_{jk}H(t_0+c_kt),$$ for coefficients
c=(c_1,\ldots,c_K)\in[0,1]^K\quad\text{and}\quad a=\begin{pmatrix}a_{11}&a_{12}&\ldots&a_{1K}\\\vdots&\vdots&\ddots&\vdots\\a_{J1}&a_{J2}&\ldots&a_{JK}\end{pmatrix}\in\mathbb{R}^{J\times K}.
The following methods are available: CFM integrators with table of coefficients
Currently, only methods of order $p=2$ and $p=4$ are implemented with adaptive step-size control.
Defining a problem with a time-dependent Hamiltonian H(t)
Define a simple problem, see also examples/exlaser.py
.
import numpy as np
import sympy as sym
import scipy.sparse
class doublewellproblem():
def __init__(self,n):
self.n = n
L = 5
self.x = np.linspace(-L,L,n+1)
# problem related inner product
dx = 2*L/(n+1)
self.inr = lambda x,y : dx*np.vdot(x,y)
# H_0
e = np.ones(n+1)
e1 = np.ones(n)
D2 = dx**(-2)*scipy.sparse.diags([e1.conj(),-2*e,e1], [-1,0,1])
self.H0 = D2
# a double well potential V_0
V0 = self.x**4 - 20*self.x**2
# add a time-dependent potential V(t)
# prepare V(t) and dV(t)
T = 5.0
omega = 10
x = sym.Symbol('x')
t = sym.Symbol('t')
exprV = 10*sym.sin((np.pi*t/T)**2)*sym.sin(omega*t)*x
exprdV = sym.diff(exprV, t)
evalVt = sym.lambdify([x,t], exprV, "numpy")
evaldVt = sym.lambdify([x,t], exprdV, "numpy")
self.Vt = lambda tau: (V0 + evalVt(self.x,tau))
self.dVt = lambda tau: evaldVt(self.x,tau)
def getprop(self):
# return nodes x and inner product
return self.x, self.inr
def getinitialstate(self):
# return initial state
s, x0 = 0.2, -2.5
u = (s*np.pi)**(-0.25)*np.exp(-(self.x-x0)**2/(2*s))
return u
def setupHamiltonian(self,t):
# return a routine to apply H(t) = H0 - (V0 + V(t))
# and dH(t) = - dV(t)
self.V = self.Vt(t)
self.dV = self.dVt(t)
H = self.H0 - scipy.sparse.diags([self.V], [0])
mv = lambda u : H.dot(u)
dH = -self.dV
dmv = lambda u : dH*u
return mv, dmv
def setupHamiltonianCFM(self,a,c,chat,t,dt):
# return a routine to apply sum_j aj*H(t + cj*dt)
# and dH(t) = sum_j (c_j+chat)*aj*dH(t + cj*dt)
# for some scalar chat
jexps = len(a)
V_CFM = a[0]*self.Vt(t+c[0]*dt)
dV_CFM = (c[0]+chat)*a[0]*self.dVt(t+c[0]*dt)
for j in range(jexps-1):
V_CFM += a[j+1]*self.Vt(t+c[j+1]*dt)
dV_CFM += (c[j+1]+chat)*a[j+1]*self.dVt(t+c[j+1]*dt)
self.V = V_CFM
self.dV = dV_CFM
if sum(a)!=0:
H = sum(a)*self.H0 - scipy.sparse.diags([self.V], [0])
mv = lambda u : H.dot(u)
else:
mv = lambda u : -self.V*u
dmv = lambda u : -self.dV*u
return mv, dmv
Magnus-Krylov methods
This package provides adaptive Magnus-Krylov methods, namely, using the adaptive midpoint rule and CFM integrators with error estimates based on symmetrized defects and works of Auzinger et al.. Again, Magnus-Krylov approximations $$y_{MK}(t,t_0,H,u)\approx \exp(\mathrm{i}t\Omega(t,t_0))u,$$ are computed to satisfy the error bound $$\lVert y_{MK}(t,t_0,H,u) -\exp(\mathrm{i}t\Omega(t,t_0))u\rVert\leq \varepsilon t,$$ where $\varepsilon>0$ is a given tolerance.
usage: examples/basicMagnusKrylov.ipynb
import numpy as np
import mkprop
from exlaser import doublewellproblem as prob
import matplotlib.pyplot as plt
# setup problem from exlaser.py
n=1200
Hamiltonian = prob(n)
x, inr = Hamiltonian.getprop()
nrm = lambda u : ((inr(u,u)).real)**0.5
u = Hamiltonian.getinitialstate()
# define initial and final time
tnow = 0
tend = 0.1
tau = tend-tnow
# some options for Krylov method, use Lanczos
ktype = 2
mmax = 60
# compute reference solution with adaptive fourth order CFM integrator
tolref=1e-6
dtinitref = 5e-2
yref,_,_,_,_,_,_ = mkprop.adaptiveCFMp4j2(u,tnow,tend,dtinitref,Hamiltonian,tol=tolref,
m=mmax,ktype=ktype,inr=inr)
# test adaptive midpoint rule
tol=1e-4
dtinit = 1e-3
y,_,_,_,_,_,_ = mkprop.adaptivemidpoint(u,tnow,tend,dtinit,Hamiltonian,tol=tol,
m=mmax,ktype=ktype,inr=inr)
print("approximation error = %.2e, tolerance = %.2e" % (nrm(yref-y)/tau, tol))
# output: approximation error = 6.40e-05, tolerance = 1.00e-04
Examples
examples/exlaser.py
: define Hamiltonian, etc..
examples/comparestepscosts.ipyn
: compare step sizes and computational cost of adaptive Magnus-Krylov methods. Step sizes:
Cost:
examples/testerrasymorder.ipynb
:
Errors and error estimates over the step size, asymptotic order.
Some references
A. Alvermann and H. Fehske. High-order commutator-free exponential time-propagation of driven quantum systems. J. Comput. Phys., 230(15):5930-5956, 2011. doi:10.1016/j.jcp.2011.04.006.
W. Auzinger, H. Hofstätter, and O. Koch. Symmetrized local error estimators for time-reversible one-step methods in nonlinear evolution equations. J. Comput. Appl. Math., 356:339-357, 2019. doi:10.1016/j.cam.2019.02.011.
W. Auzinger, H. Hofstätter, O. Koch, M. Quell, and M. Thalhammer. A posteriori error estimation for Magnus-type integrators. M2AN - Math. Model. Numer. Anal., 53(1):197-218, 2019. doi:10.1051/m2an/2018050.
W. Auzinger and O. Koch. An improved local error estimator for symmetric time-stepping schemes. Appl. Math. Lett., 82:106-110, 2018. doi:10.1016/j.aml.2018.03.001.
P. Bader, S. Blanes, and N. Kopylov. Exponential propagators for the Schrödinger equation with a time-dependent potential. J. Chem. Phys., 148(24):244109, 2018. doi:10.1063/1.5036838.
T. Jawecki, W. Auzinger, and O. Koch. Computable upper error bounds for Krylov approximations to matrix exponentials and associated $\varphi$-functions. BIT, 60(1):157-197, 2020. doi:10.1007/s10543-019-00771-6.
T. Jawecki. Krylov techniques and approximations to the action of matrix exponentials. PhD thesis, TU Wien, Austria, 2022. doi:10.34726/hss.2022.45083.
T. Jawecki. A study of defect-based error estimates for the Krylov approximation of $\varphi$-functions. Numer. Algorithms, 90(1):323-361, 2022. doi:10.1007/s11075-021-01190-x.
W. Magnus. On the exponential solution of differential equations for a linear operator. Comm. Pure Appl. Math., 7(4):649-673, 1954. doi:10.1002/cpa.3160070404.
S. Blanes, F. Casas, J. Oteo and J. Ros. The Magnus expansion and some of its applications. Phys. Rep., 470(5):151-238, 2009. doi:10.1016/j.physrep.2008.11.001.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.