arxiv_papers / 1001.0044.txt
alx-ai's picture
Upload 920 files
b13a737
arXiv:1001.0044v3 [math.PR] 31 Mar 2011A law of large numbers approximation for
Markov population processes with countably
many types
A. D. Barbour∗and M. J. Luczak†
Universit¨ at Z¨ urich and London School of Economics
Abstract
When modelling metapopulation dynamics, the influence of a s in-
gle patch on the metapopulation depends on the number of indi vidu-
als in the patch. Since the population size has no natural upp er limit,
this leads to systems in which there are countably infinitely many
possible types of individual. Analogous considerations ap ply in the
transmission of parasitic diseases. In this paper, we prove a law of
large numbers for quite general systems of this kind, togeth er with
a rather sharp bound on the rate of convergence in an appropri ately
chosen weighted ℓ1norm.
Keywords: Epidemic models, metapopulation processes, countably many
types, quantitative law of large numbers, Markov population proce sses
AMS subject classification: 92D30, 60J27, 60B12
Running head: A law of large numbers approximation
1 Introduction
There are many biological systems that consist of entities that diffe r in their
influence according to the number of active elements associated wit h them,
∗Angewandte Mathematik, Universit¨ at Z¨ urich, Winterthurertra sse 190, CH-8057
Z¨URICH; ADB was supported in part by Schweizerischer Nationalfond s Projekt Nr. 20–
107935/1.
†London School of Economics; MJL was supported in part by a STICE RD grant.
1and can be divided into types accordingly. In parasitic diseases (Bar bour &
Kafetzaki 1993, Luchsinger 2001a,b, Kretzschmar 1993), the in fectivity of a
host depends on the number of parasites that it carries; in metapo pulations,
the migration pressure exerted by a patch is related to the number of its
inhabitants (Arrigoni 2003); the behaviour of a cell may depend on the num-
ber of copies of a particular gene that it contains (Kimmel & Axelrod 2 002,
Chapter 7); and so on. In none of these examples is there a natura l upper
limit to the number of associated elements, so that the natural set ting for
a mathematical model is one in which there are countably infinitely man y
possible types of individual. In addition, transition rates typically incr ease
with the number of associated elements in the system — for instance , each
parasite has an individual death rate, so that the overall death ra te of par-
asites grows at least as fast as the number of parasites — and this le ads
to processes with unbounded transition rates. This paper is conce rned with
approximations to density dependent Markov models of this kind, wh en the
typical population size Nbecomes large.
In density dependent Markov population processes with only finitely
many types of individual, a law of large numbers approximation, in the f orm
ofasystemofordinarydifferentialequations, wasestablishedbyK urtz(1970),
together with a diffusion approximation (Kurtz, 1971). In the infinit e di-
mensional case, the law of large numbers was proved for some spec ific mod-
els (Barbour & Kafetzaki 1993, Luchsinger 2001b, Arrigoni 2003 , see also
L´ eonard 1990), using individually tailored methods. A more general result
was then given by Eibeck & Wagner (2003). In Barbour & Luczak (20 08),
the law of large numbers was strengthened by the addition of an err or bound
inℓ1that is close to optimal order in N. Their argument makes use of an
intermediate approximation involving an independent particles proce ss, for
which the law of large numbers is relatively easy to analyse. This proce ss is
then shown to be sufficiently close to the interacting process of act ual inter-
est, by means of a coupling argument. However, the generality of t he results
obtained is limited by the simple structure of the intermediate proces s, and
the model of Arrigoni (2003), for instance, lies outside their scop e.
In this paper, we develop an entirely different approach, which circu m-
vents the need for an intermediate approximation, enabling a much w ider
class of models to be addressed. The setting is that of families of Mar kov
population processes XN:= (XN(t), t≥0),N≥1, taking values in the
countable space X+:={X∈ZZ+
+;/summationtext
m≥0Xm<∞}. Each component repre-
2sents the number of individuals of a particular type, and there are c ountably
many types possible; however, at any given time, there are only finit ely
many individuals in the system. The process evolves as a Markov proc ess
with state-dependent transitions
X→X+Jat rate NαJ(N−1X), X∈ X+, J∈ J,(1.1)
where each jump is of bounded influence, in the sense that
J ⊂ {X∈ZZ+;/summationdisplay
m≥0|Xm| ≤J∗<∞},for some fixed J∗<∞,(1.2)
so that the number of individuals affected is uniformly bounded. Dens ity
dependence is reflected in the fact that the arguments of the fun ctionsαJ
are counts normalised by the ‘typical size’ N. Writing R:=RZ+
+, the func-
tionsαJ:R →R+are assumed to satisfy
/summationdisplay
J∈JαJ(ξ)<∞, ξ∈ R0, (1.3)
whereR0:={ξ∈ R:ξi= 0 for all but finitely many i}; this assumption
implies that the processes XNare pure jump processes, at least for some
non-zero length of time. To prevent the paths leaving X+, we also assume
thatJl≥ −1 for each l, and that αJ(ξ) = 0 ifξl= 0 for any J∈ Jsuch
thatJl=−1. Some remarks on the consequences of allowing transitions J
withJl≤ −2 for some lare made at the end of Section 4.
Thelawoflargenumbersisthenformallyexpressed intermsofthesy stem
ofdeterministic equations
dt=/summationdisplay
J∈JJαJ(ξ) =:F0(ξ), (1.4)
to be understood componentwise for those ξ∈ Rsuch that
/summationdisplay
J∈J|Jl|αJ(ξ)<∞,for alll≥0,
thus by assumption including R0. Here, the quantity F0represents the in-
finitesimal average drift of the components of the random proces s. However,
in this generality, it is not even immediately clear that equations (1.4) h ave
a solution.
3In order to make progress, it is assumed that the unbounded comp onents
in the transition rates can be assimilated into a linear part, in the sens e
thatF0can be written in the form
F0(ξ) =Aξ+F(ξ), (1.5)
again to be understood componentwise, where Ais a constant Z+×Z+
matrix. These equations are then treated as a perturbed linear sy stem
(Pazy 1983, Chapter 6). Under suitable assumptions on A, there exists a
measure µonZ+, defining a weighted ℓ1norm/⌊ard⌊l · /⌊ard⌊lµonR, and a strongly
/⌊ard⌊l·/⌊ard⌊lµ–continuoussemigroup {R(t), t≥0}oftransitionmatriceshaving point-
wise derivative R′(0) =A. IfFis locally /⌊ard⌊l·/⌊ard⌊lµ–Lipschitz and /⌊ard⌊lx(0)/⌊ard⌊lµ<∞,
this suggests using the solution xof the integral equation
x(t) =R(t)x(0)+/integraldisplayt
0R(t−s)F(x(s))ds (1.6)
as an approximation to xN:=N−1XN, instead of solving the deterministic
equations (1.4) directly. We go on to show that the solution XNof the
stochastic system can be expressed using a formula similar to (1.6), which
has an additional stochastic component in the perturbation:
xN(t) =R(t)xN(0)+/integraldisplayt
0R(t−s)F(xN(s))ds+/tildewidemN(t),(1.7)
where
/tildewidemN(t) :=/integraldisplayt
0R(t−s)dmN(s), (1.8)
andmNis the local martingale given by
mN(t) :=xN(t)−xN(0)−/integraldisplayt
0F0(xN(s))ds. (1.9)
The quantity mNcanbe expected to be small, at least componentwise, under
reasonable conditions.
To obtain tight control over /tildewidemNin all components simultaneously, suf-
ficient to ensure that sup0≤s≤t/⌊ard⌊l/tildewidemN(s)/⌊ard⌊lµis small, we derive Chernoff–like
boundsonthedeviations ofthemost significant components, witht hehelpof
a family of exponential martingales. The remaining components are t reated
usingsomegeneral a prioriboundsonthebehaviourofthestochasticsystem.
4This allows us to take the difference between the stochastic and det erministic
equations (1.7) and (1.6), after which a Gronwall argument can be c arried
through, leading to the desired approximation.
The main result, Theorem 4.7, guarantees an approximation error o f or-
derO(N−1/2√logN) in the weighted ℓ1metric/⌊ard⌊l·/⌊ard⌊lµ, except on an event of
probability of order O(N−1logN). More precisely, for each T >0, there
exist constants K(1)
T,K(2)
T,K(3)
Tsuch that, for Nlarge enough, if
/⌊ard⌊lN−1XN(0)−x(0)/⌊ard⌊lµ≤K(1)
T/radicalbigg
logN
N,
then
P/parenleftBig
sup
0≤t≤T/⌊ard⌊lN−1XN(t)−x(t)/⌊ard⌊lµ> K(2)
T/radicalbigg
logN
N/parenrightBig
≤K(3)
TlogN
N.(1.10)
Theerrorboundissharper, byafactoroflog N, thanthatgiveninBarbour&
Luczak(2008),andthetheoremisapplicabletoamuch widerclassof models.
However, the method of proof involves moment arguments, which r equire
somewhat stronger assumptions on the initial state of the system , and, in
models such as that of Barbour & Kafetzaki (1993), onthe choice of infection
distributions allowed. The conditions under which the theorem holds c an be
divided into three categories: growth conditions on the transition r ates, so
that the a prioribounds, which have the character of moment bounds, can
be established; conditions on the matrix A, sufficient to limit the growth of
the semigroup R, and (together with the properties of F) to determine the
weights defining the metric in which the approximation is to be carried o ut;
and conditions on the initial state of the system. The a priori bounds are
derived in Section 2, the semigroup analysis is conducted in Section 3, and
the approximation proper is carried out in Section 4. The paper conc ludes
in Section 5 with some examples.
The form (1.8) of the stochastic component /tildewidemN(t) in (1.7) is very simi-
lar to that of a key element in the analysis of stochastic partial differ ential
equations; see, for example, Chow (2007, Section 6.6). The SPDE a rguments
used for its control are however typically conducted in a Hilbert spa ce con-
text. Our setting is quite different in nature, and it does not seem cle ar how
to translate the SPDE methods into our context.
52 A priori bounds
We begin by imposing further conditions on the transition rates of th e pro-
cessXN, sufficient to constrain its paths to bounded subsets of X+dur-
ing finite time intervals, and in particular to ensure that only finitely ma ny
jumps can occur in finite time. The conditions that follow have the flav our
of moment conditions on the jump distributions. Since the index j∈Z+is
symbolic in nature, we start by fixing an ν∈ R, such that ν(j) reflects in
some sense the ‘size’ of j, with most indices being ‘large’:
ν(j)≥1 for allj≥0 and lim
j→∞ν(j) =∞. (2.1)
We then define the analogues of higher empirical moments using the q uanti-
tiesνr∈ R, defined by νr(j) :=ν(j)r,r≥0, setting
Sr(x) :=/summationdisplay
j≥0νr(j)xj=xTνr, x∈ R0, (2.2)
where, for x∈ R0andy∈ R,xTy:=/summationtext
l≥0xlyl. In particular, for X∈ X+,
S0(X) =/⌊ard⌊lX/⌊ard⌊l1. Note that, because of (2.1), for any r≥1,
#{X∈ X+:Sr(X)≤K}<∞for allK >0. (2.3)
To formulate the conditions that limit the growth of the empirical mom ents
ofXN(t) witht, we also define
Ur(x) :=/summationdisplay
J∈JαJ(x)JTνr;Vr(x) :=/summationdisplay
J∈JαJ(x)(JTνr)2, x∈ R.(2.4)
The assumptions that we shall need are then as follows.
Assumption 2.1 There exists a νsatisfying (2.1)andr(1)
max,r(2)
max≥1such
that, for all X∈ X+,
/summationdisplay
J∈JαJ(N−1X)|JTνr|<∞,0≤r≤r(1)
max,(2.5)
the case r= 0following from (1.2)and(1.3); furthermore, for some non-
negative constants krl, the inequalities
U0(x)≤k01S0(x)+k04,
U1(x)≤k11S1(x)+k14, (2.6)
Ur(x)≤ {kr1+kr2S0(x)}Sr(x)+kr4,2≤r≤r(1)
max;
6and
V0(x)≤k03S1(x)+k05,
Vr(x)≤kr3Sp(r)(x)+kr5,1≤r≤r(2)
max, (2.7)
are satisfied, where 1≤p(r)≤r(1)
maxfor1≤r≤r(2)
max.
The quantities r(1)
maxandr(2)
maxusually need to be reasonably large, if Assump-
tion 4.2 below is to be satisfied.
Now, for XNas in the introduction, we let tXNndenote the time of its
n-th jump, with tXN
0= 0, and set tXN∞:= lim n→∞tXNn, possibly infinite. For
0≤t < tXN∞, we define
S(N)
r(t) :=Sr(XN(t));U(N)
r(t) :=Ur(xN(t));V(N)
r(t) :=Vr(xN(t)),
(2.8)
once again with xN(t) :=N−1XN(t), and also
τ(N)
r(C) := inf {t < tXN
∞:S(N)
r(t)≥NC}, r≥0,(2.9)
where the infimum of the empty set is taken to be ∞. Our first result shows
thattXN∞=∞a.s., and limits the expectations of S(N)
0(t) andS(N)
1(t) for any
fixedt.
In what follows, we shall write F(N)
s=σ(XN(u),0≤u≤s), so that
(F(N)
s:s≥0) is the natural filtration of the process XN.
Lemma 2.2 Under Assumptions 2.1, tXN∞=∞a.s. Furthermore, for any
t≥0,
E{S(N)
0(t)} ≤(S(N)
0(0)+Nk04t)ek01t;
E{S(N)
1(t)} ≤(S(N)
1(0)+Nk14t)ek11t.
Proof. Introducing the formal generator ANassociated with (1.1),
ANf(X) :=N/summationdisplay
J∈JαJ(N−1X){f(X+J)−f(X)}, X∈ X+,(2.10)
we note that NUl(x) =ANSl(Nx). Hence, if we define M(N)
lby
M(N)
l(t) :=S(N)
l(t)−S(N)
l(0)−N/integraldisplayt
0U(N)
l(u)du, t ≥0,(2.11)
7for 0≤l≤r(1)
max, it is immediate from (2.3), (2.5) and (2.6) that the process
(M(N)
l(t∧τ(N)
1(C)), t≥0) is a zero mean F(N)–martingale for each C >0.
In particular, considering M(N)
1(t∧τ(N)
1(C)), it follows in view of (2.6) that
E{S(N)
1(t∧τ(N)
1(C))} ≤S(N)
1(0)+E/braceleftBigg/integraldisplayt∧τ(N)
1(C)
0{k11S(N)
1(u)+Nk14}du/bracerightBigg
≤S(N)
1(0)+/integraldisplayt
0(k11E{S(N)
1(u∧τ(N)
1(C))}+Nk14)du.
Using Gronwall’s inequality, we deduce that
E{S(N)
1(t∧τ(N)
1(C))} ≤(S(N)
1(0)+Nk14t)ek11t,(2.12)
uniformly in C >0, and hence that
P/bracketleftBig
sup
0≤s≤tS1(XN(s))≥NC/bracketrightBig
≤C−1(S1(xN(0))+k14t)ek11t(2.13)
also. Hence sup0≤s≤tS1(XN(s))<∞a.s. for any t, limC→∞τ(N)
1(C) =∞
a.s., and, from (2.3) and (1.3), it thus follows that tXN∞=∞a.s. The bound
onE{S(N)
1(t)}is now immediate, and that on E{S(N)
0(t)}follows by applying
the same Gronwall argument to M(N)
0(t∧τ(N)
1(C)).
The next lemma shows that, if any T >0 is fixed and Cis chosen large
enough, then, with high probability, N−1S(N)
0(t)≤Cholds for all 0 ≤t≤T.
Lemma 2.3 Assume that Assumptions 2.1 are satisfied, and that S(N)
0(0)≤
NC0andS(N)
1(0)≤NC1. Then, for any C≥2(C0+k04T)ek01T, we have
P[{τ(N)
0(C)≤T}]≤(C1∨1)K00/(NC2),
whereK00depends on Tand the parameters of the model.
Proof. It is immediate from (2.11) and (2.6) that
S(N)
0(t) =S(N)
0(0)+N/integraldisplayt
0U(N)
0(u)du+M(N)
0(t)
≤S(N)
0(0)+/integraldisplayt
0(k01S(N)
0(u)+Nk04)du+ sup
0≤u≤tM(N)
0(u).(2.14)
8Hence, from Gronwall’s inequality, if S(N)
0(0)≤NC0, then
S(N)
0(t)≤/braceleftbigg
N(C0+k04T)+ sup
0≤u≤tM(N)
0(u)/bracerightbigg
ek01t.(2.15)
Now, considering the quadratic variation of M(N)
0, we have
E/braceleftBigg
{M(N)
0(t∧τ(N)
1(C′))}2−N/integraldisplayt∧τ(N)
1(C′)
0V(N)
0(u)du/bracerightBigg
= 0 (2.16)
for anyC′>0, from which it follows, much as above, that
E/parenleftBig
{M(N)
0(t∧τ(N)
1(C′))}2/parenrightBig
≤E/braceleftbigg
N/integraldisplayt
0V(N)
0(u∧τ(N)
1(C′))du/bracerightbigg
≤/integraldisplayt
0{k03ES(N)
1(u∧τ(N)
1(C′))+Nk05}du.
Using (2.12), we thus find that
E/parenleftBig
{M(N)
0(t∧τ(N)
1(C′))}2/parenrightBig
≤k03
k11N(C1+k14T)(ek11t−1)+Nk05t,(2.17)
uniformlyforall C′. Doob’smaximal inequality appliedto M(N)
0(t∧τ(N)
1(C′))
now allows us to deduce that, for any C′,a >0,
P/bracketleftBig
sup
0≤u≤TM(N)
0(u∧τ(N)
1(C′))> aN/bracketrightBig
≤1
Na2/braceleftbiggk03
k11(C1+k14T){ek11T−1}+k05T/bracerightbigg
=:C1K01+K02
Na2,
say, so that, letting C′→ ∞,
P/bracketleftBig
sup
0≤u≤TM(N)
0(u)> aN/bracketrightBig
≤C1K01+K02
Na2
also. Taking a=1
2Ce−k01Tand putting the result into (2.15), the lemma
follows.
In the next theorem, we control the ‘higher ν-moments’ S(N)
r(t) ofXN(t).
9Theorem 2.4 Assume thatAssumptions 2.1are satisfied, andthat S(N)
1(0)≤
NC1andS(N)
p(1)(0)≤NC′
1. Then, for 2≤r≤r(1)
maxand for any C >0, we
have
E{S(N)
r(t∧τ(N)
0(C))} ≤(S(N)
r(0)+Nkr4t)e(kr1+Ckr2)t,0≤t≤T.(2.18)
Furthermore, if for 1≤r≤r(2)
max,S(N)
r(0)≤NCrandS(N)
p(r)(0)≤NC′
r,
then, for any γ≥1,
P[ sup
0≤t≤TS(N)
r(t∧τ(N)
0(C))≥NγC′′
rT]≤Kr0γ−2N−1, (2.19)
where
C′′
rT:= (Cr+kr4T+/radicalbig
(C′
r∨1))e(kr1+Ckr2)T
andKr0depends on C,Tand the parameters of the model.
Proof. Recalling (2.11), use the argument leading to (2.12) with the martin-
galesM(N)
r(t∧τ(N)
1(C′)∧τ(N)
0(C)), for any C′>0, to deduce that
ES(N)
r(t∧τ(N)
1(C′)∧τ(N)
0(C))
≤S(N)
r(0)+/integraldisplayt
0/parenleftBig
{kr1+Ckr2}E/braceleftBig
S(N)
r(u∧τ(N)
1(C′)∧τ(N)
0(C))/bracerightBig
+Nkr4/parenrightBig
du,
for 1≤r≤r(1)
max, sinceN−1S(N)
0(u)≤Cwhenu≤τ(N)
0(C): define k12= 0.
Gronwall’s inequality now implies that
ES(N)
r(t∧τ(N)
1(C′)∧τ(N)
0(C))≤(S(N)
r(0)+Nkr4t)e(kr1+Ckr2)t,(2.20)
for 1≤r≤r(1)
max, and (2.18) follows by Fatou’s lemma, on letting C′→ ∞.
Now, also from (2.11) and (2.6), we have, for t≥0 and each r≤r(1)
max,
S(N)
r(t∧τ(N)
0(C))
=S(N)
r(0)+N/integraldisplayt∧τ(N)
0(C)
0U(N)
r(u)du+M(N)
r(t∧τ(N)
0(C))
≤S(N)
r(0)+/integraldisplayt
0/parenleftBig
{kr1+Ckr2}S(N)
r(u∧τ(N)
0(C))+Nkr4/parenrightBig
du
+ sup
0≤u≤tM(N)
r(u∧τ(N)
0(C)).
10Hence, from Gronwall’s inequality, for all t≥0 andr≤r(1)
max,
S(N)
r(t∧τ(N)
0(C))≤/braceleftBig
N(Cr+kr4t)+ sup
0≤u≤tM(N)
r(u∧τ(N)
0(C))/bracerightBig
e(kr1+Ckr2)t.
(2.21)
Now, as in (2.16), we have
E/braceleftBigg
{M(N)
r(t∧τ(N)
1(C′)∧τ(N)
0(C))}2−N/integraldisplayt∧τ(N)
1(C′)∧τ(N)
0(C)
0V(N)
r(u)du/bracerightBigg
= 0,
(2.22)
from which it follows, using (2.7), that, for 1 ≤r≤r(2)
max,
E/parenleftBig
{M(N)
r(t∧τ(N)
1(C′)∧τ(N)
0(C))}2/parenrightBig
≤E/braceleftBigg
N/integraldisplayt∧τ(N)
1(C′)∧τ(N)
0(C))
0V(N)
r(u)du/bracerightBigg
≤/integraldisplayt
0{kr3ES(N)
p(r)(u∧τ(N)
1(C′)∧τ(N)
0(C))+Nkr5}du
≤N(C′
r+kp(r),4T)kr3
kp(r),1+Ckp(r),2(e(kp(r),1+Ckp(r),2t)−1)+Nkr5T,
this last by (2.20), since p(r)≤r(1)
maxfor 1≤r≤r(2)
max. Using Doob’s
inequality, it follows that, for any a >0,
P/bracketleftBig
sup
0≤u≤TM(N)
r(u∧τ(N)
0(C))> aN/bracketrightBig
≤1
Na2/braceleftbiggkr3(C′
r+kp(r),4T)
kp(r),1+Ckp(r),2(e(kp(r),1+Ckp(r),2T)−1)+kr5T/bracerightbigg
=:C′
rKr1+Kr2
Na2.
Takinga=γ/radicalbig
(C′
r∨1) and putting the result into (2.21) gives (2.19), with
Kr0= (C′
rKr1+Kr2)/(C′
r∨1).
Note also that sup0≤t≤TS(N)
r(t)<∞a.s. for all 0 ≤r≤r(2)
max, in view of
Lemma 2.3 and Theorem 2.4.
In what follows, we shall particularly need to control quantities of t he
form/summationtext
J∈JαJ(xN(s))d(J,ζ), where xN:=N−1XNand
d(J,ζ) :=/summationdisplay
j≥0|Jj|ζ(j), (2.23)
11forζ∈ Rchosen such that ζ(j)≥1 grows fast enough with j: see (4.12).
Defining
τ(N)(a,ζ) := inf/braceleftBigg
s:/summationdisplay
J∈JαJ(xN(s))d(J,ζ)≥a/bracerightBigg
,(2.24)
infinite if there is no such s, we show in the following corollary that, under
suitable assumptions, τ(N)(a,ζ) is rarely less than T.
Corollary 2.5 Assume that Assumptions 2.1 hold, and that ζis such that
/summationdisplay
J∈JαJ(N−1X)d(J,ζ)≤ {k1N−1Sr(X)+k2}b(2.25)
for some 1≤r:=r(ζ)≤r(2)
maxand some b=b(ζ)≥1. For this value
ofr, assume that S(N)
r(0)≤NCrandS(N)
p(r)(0)≤NC′
rfor some constants
CrandC′
r. Assume further that S(N)
0(0)≤NC0,S(N)
1(0)≤NC1for some
constants C0,C1, and define C:= 2(C0+k04T)ek01T. Then
P[τ(N)(a,ζ)≤T]≤N−1{Kr0γ−2
a+K00(C1∨1)C−2},
for anya≥ {k2+k1C′′
rT}b, whereγa:= (a1/b−k2)/{k1C′′
rT},Kr0andC′′
rT
are as in Theorem 2.4, and K00is as in Lemma 2.3.
Proof. In view of (2.25), it is enough to bound the probability
P[ sup
0≤t≤TS(N)
r(t)≥N(a1/b−k2)/k1].
However, Lemma 2.3 and Theorem 2.4 together bound this probability by
N−1/braceleftbig
Kr0γ−2
a+K00(C1∨1)C−2/bracerightbig
,
whereγais as defined above, as long as a1/b−k2≥k1C′′
rT.
If (2.25) is satisfied,/summationtext
J∈JαJ(xN(s))d(J,ζ)isa.s. bounded on0 ≤s≤T,
becauseS(N)
r(s) is. The corollary shows that the sum is then bounded by
{k2+k1C′′
r,T}b, except on an event of probability of order O(N−1). Usually,
one can choose b= 1.
123 Semigroup properties
We make the following initial assumptions about the matrix A: first, that
Aij≥0 for alli/ne}ationslash=j≥0;/summationdisplay
j/negationslash=iAji<∞for alli≥0,(3.1)
and then that, for some µ∈RZ+
+such that µ(m)≥1 for each m≥0, and
for some w≥0,
ATµ≤wµ. (3.2)
We then use µto define the µ-norm
/⌊ard⌊lξ/⌊ard⌊lµ:=/summationdisplay
m≥0µ(m)|ξm|onRµ:={ξ∈ R:/⌊ard⌊lξ/⌊ard⌊lµ<∞}.(3.3)
Note that there may be many possible choices for µ. In what follows, it is
important that Fbe a Lipschitz operator with respect to the µ-norm, and
this has to be borne in mind when choosing µ.
Setting
Qij:=AT
ijµ(j)/µ(i)−wδij, (3.4)
whereδis the Kronecker delta, we note that Qij≥0 fori/ne}ationslash=j, and that
0≤/summationdisplay
j/negationslash=iQij=/summationdisplay
j/negationslash=iAT
ijµ(j)/µ(i)≤w−Aii=−Qii,
using (3.2) for the inequality, so that Qii≤0. Hence Qcan be augmented to
a conservative Q–matrix, in the sense of Markov jump processes, by adding a
coffin state ∂, and setting Qi∂:=−/summationtext
j≥0Qij≥0. LetP(·) denote the semi-
group of Markov transition matrices corresponding to the minimal p rocess
associated with Q; then, in particular,
Q=P′(0) and P′(t) =QP(t) for all t≥0 (3.5)
(Reuter 1957, Theorem 3). Set
RT
ij(t) :=ewtµ(i)Pij(t)/µ(j). (3.6)
13Theorem 3.1 LetAsatisfy Assumptions (3.1)and(3.2). Then, with the
above definitions, Ris a strongly continuous semigroup on Rµ, and
/summationdisplay
i≥0µ(i)Rij(t)≤µ(j)ewtfor alljandt. (3.7)
Furthermore, the sums/summationtext
j≥0Rij(t)Ajk= (R(t)A)ikare well defined for all
i,k, and
A=R′(0)andR′(t) =R(t)Afor allt≥0.(3.8)
Proof. We note first that, for x∈ Rµ,
/⌊ard⌊lR(t)x/⌊ard⌊lµ≤/summationdisplay
i≥0µ(i)/summationdisplay
j≥0Rij(t)|xj|=ewt/summationdisplay
i≥0/summationdisplay
j≥0µ(j)Pji(t)|xj|
≤ewt/summationdisplay
j≥0µ(j)|xj|=ewt/⌊ard⌊lx/⌊ard⌊lµ, (3.9)
sinceP(t) is substochastic on Z+; henceR:Rµ→ Rµ. To show strong
continuity, we take x∈ Rµ, and consider
/⌊ard⌊lR(t)x−x/⌊ard⌊lµ=/summationdisplay
i≥0µ(i)/vextendsingle/vextendsingle/vextendsingle/vextendsingle/vextendsingle/summationdisplay
j≥0Rij(t)xj−xi/vextendsingle/vextendsingle/vextendsingle/vextendsingle/vextendsingle=/summationdisplay
i≥0/vextendsingle/vextendsingle/vextendsingle/vextendsingle/vextendsingleewt/summationdisplay
j≥0µ(j)Pji(t)xj−µ(i)xi/vextendsingle/vextendsingle/vextendsingle/vextendsingle/vextendsingle
≤(ewt−1)/summationdisplay
i≥0/summationdisplay
j≥0µ(j)Pji(t)xj+/summationdisplay
i≥0/summationdisplay
j/negationslash=iµ(j)Pji(t)xj+/summationdisplay
i≥0µ(i)xi(1−Pii(t))
≤(ewt−1)/summationdisplay
j≥0µ(j)xj+2/summationdisplay
i≥0µ(i)xi(1−Pii(t)),
from which it follows that lim t→0/⌊ard⌊lR(t)x−x/⌊ard⌊lµ= 0, by dominated conver-
gence, since lim t→0Pii(t) = 1 for each i≥0.
The inequality (3.7) follows from the definition of Rand the fact that P
is substochastic on Z+. Then
(ATRT(t))ij=/summationdisplay
k/negationslash=iQikµ(i)
µ(k)ewtµ(k)
µ(j)Pkj(t)+(Qii+w)ewtµ(i)
µ(j)Pij(t)
=µ(i)
µ(j)[(QP(t))ij+wPij(t)]ewt,
14with (QP(t))ij=/summationtext
k≥0QikPkj(t) well defined because P(t) is sub-stochastic
andQis conservative. Using (3.5), this gives
(ATRT(t))ij=µ(i)
µ(j)d
dt[Pij(t)ewt] =d
dtRT
ij(t),
and this establishes (3.8).
4 Main approximation
LetXN,N≥1, beasequence ofpure jumpMarkov processes asinSection 1,
withAandFdefined as in (1.4) and (1.5), and suppose that F:Rµ→ Rµ,
withRµas defined in (3.3), for some µsuch that Assumption (3.2) holds.
Suppose also that Fis locally Lipschitz in the µ-norm: for any z >0,
sup
x/negationslash=y:/bardblx/bardblµ,/bardbly/bardblµ≤z/⌊ard⌊lF(x)−F(y)/⌊ard⌊lµ//⌊ard⌊lx−y/⌊ard⌊lµ≤K(µ,F;z)<∞.(4.1)
Then, for x(0)∈ RµandRas in (3.6), the integral equation
x(t) =R(t)x(0)+/integraldisplayt
0R(t−s)F(x(s))ds. (4.2)
has a unique continuous solution xinRµon some non-empty time interval
[0,tmax), such that, if tmax<∞, then/⌊ard⌊lx(t)/⌊ard⌊lµ→ ∞ast→tmax(Pazy 1983,
Theorem 1.4, Chapter 6). Thus, if Awere the generator of R, the function x
would be a mild solution of the deterministic equations (1.4). We now wish
to show that the process xN:=N−1XNis close to x. To do so, we need a
corresponding representation for XN.
To find such a representation, let W(t),t≥0, be a pure jump path on X+
that has only finitely many jumps up to time T. Then we can write
W(t) =W(0)+/summationdisplay
j:σj≤t∆W(σj),0≤t≤T, (4.3)
where ∆W(s) :=W(s)−W(s−)andσj,j≥1, denote thetimes when Whas
its jumps. Now let Asatisfy (3.1) and (3.2), and let R(·) be the associated
semigroup, as defined in (3.6). Define the path W∗(t), 0≤t≤T, from the
equation
W∗(t) :=R(t)W(0)+/summationtext
j:σj≤tR(t−σj)∆j−/integraltextt
0R(t−s)AW(s)ds,
(4.4)
15where ∆ j:= ∆W(σj). Note that the latter integral makes sense, because
each of the sums/summationtext
j≥0Rij(t)Ajkis well defined, from Theorem 3.1, and
because only finitely many of the coordinates of Ware non-zero.
Lemma 4.1 W∗(t) =W(t)for all0≤t≤T.
Proof. Fix any t, and suppose that W∗(s) =W(s) for alls≤t. This is
clearly the case for t= 0. Let σ(t)> tdenote the time of the first jump
ofWaftert. Then, for any 0 < h < σ(t)−t, using the semigroup property
forRand (4.4),
W∗(t+h)−W∗(t)
= (R(h)−I)R(t)W(0)+/summationdisplay
j:σj≤t(R(h)−I)R(t−σj)∆j (4.5)
−/integraldisplayt
0(R(h)−I)R(t−s)AW(s)ds−/integraldisplayt+h
tR(t+h−s)AW(t)ds,
where, in the last integral, we use the fact that there are no jumps ofW
between tandt+h. Thus we have
W∗(t+h)−W∗(t)
= (R(h)−I)
R(t)W(0)+/summationdisplay
j:σj≤tR(t−σj)∆j−/integraldisplayt
0R(t−s)AW(s)ds
−/integraldisplayt+h
tR(t+h−s)AW(t)ds
= (R(h)−I)W(t)−/integraldisplayt+h
tR(t+h−s)AW(t)ds. (4.6)
But now, for x∈ X+,
/integraldisplayt+h
tR(t+h−s)Axds= (R(h)−I)x,
from (3.8), so that W∗(t+h) =W∗(t) for all t+h < σ(t), implying that
W∗(s) =W(s) for all s < σ(t). On the other hand, from (4.4), we have
W∗(σ(t))−W∗(σ(t)−) = ∆W(σ(t)), so that W∗(s) =W(s) for alls≤σ(t).
Thus we can prove equality over the interval [0 ,σ1], and then successively
over the intervals [ σj,σj+1], until [0 ,T] is covered.
16Now suppose that Warises as a realization of XN. ThenXNhas transi-
tion rates such that
MN(t) :=/summationdisplay
j:σj≤t∆XN(σj)−/integraldisplayt
0AXN(s)ds−/integraldisplayt
0NF(xN(s))ds(4.7)
is a zero mean local martingale. In view of Lemma 4.1, we can use (4.4) t o
write
XN(t) =R(t)XN(0)+/tildewiderMN(t)+N/integraldisplayt
0R(t−s)F(xN(s))ds,(4.8)
where
/tildewiderMN(t) :=/summationdisplay
j:σj≤tR(t−σj)∆XN(σj)
−/integraldisplayt
0R(t−s)AXN(s)ds−/integraldisplayt
0R(t−s)NF(xN(s))ds.(4.9)
Thus, comparing (4.8) and (4.2), we expect xNandxto be close, for
0≤t≤T < tmax, provided that we can show that supt≤T/⌊ard⌊l/tildewidemN(t)/⌊ard⌊lµis small,
where/tildewidemN(t) :=N−1/tildewiderMN(t). Indeed, if xN(0) andx(0) are close, then
/⌊ard⌊lxN(t)−x(t)/⌊ard⌊lµ
≤ /⌊ard⌊lR(t)(xN(0)−x(0))/⌊ard⌊lµ
+/integraldisplayt
0/⌊ard⌊lR(t−s)[F(xN(s))−F(x(s))]/⌊ard⌊lµds+/⌊ard⌊l/tildewidemN(t)/⌊ard⌊lµ
≤ewt/⌊ard⌊lxN(0)−x(0)/⌊ard⌊lµ
+/integraldisplayt
0ew(t−s)K(µ,F;2ΞT)/⌊ard⌊lxN(s)−x(s)/⌊ard⌊lµds+/⌊ard⌊l/tildewidemN(t)/⌊ard⌊lµ,(4.10)
by (3.9), with the stage apparently set for Gronwall’s inequality, ass uming
that/⌊ard⌊lxN(0)−x(0)/⌊ard⌊lµand sup0≤t≤T/⌊ard⌊l/tildewidemN(t)/⌊ard⌊lµare small enough that then
/⌊ard⌊lxN(t)/⌊ard⌊lµ≤2ΞTfor 0≤t≤T, where Ξ T:= sup0≤t≤T/⌊ard⌊lx(t)/⌊ard⌊lµ.
Bounding sup0≤t≤T/⌊ard⌊l/tildewidemN(t)/⌊ard⌊lµis, however, not so easy. Since /tildewiderMNis not
itselfamartingale, wecannotdirectlyapplymartingaleinequalitiestoc ontrol
its fluctuations. However, since
/tildewiderMN(t) =/integraldisplayt
0R(t−s)dMN(s), (4.11)
17we can hope to use control over the local martingale MNinstead. For this
and the subsequent argument, we introduce some further assum ptions.
Assumption 4.2
1. There exists r=rµ≤r(2)
maxsuch that supj≥0{µ(j)/νr(j)}<∞.
2. There exists ζ∈ Rwithζ(j)≥1for alljsuch that (2.25)is satisfied
for some b=b(ζ)≥1andr=r(ζ)such that 1≤r(ζ)≤r(2)
max, and that
Z:=/summationdisplay
k≥0µ(k)(|Akk|+1)/radicalbig
ζ(k)<∞. (4.12)
The requirement that ζsatisfies (4.12) as well as satisfying (2.25) for some
r≤r(2)
maximplies in practice that it must be possible to take r(1)
maxandr(2)
max
to be quite large in Assumption 2.1; see the examples in Section 5.
Note that part 1 of Assumption 4.2 implies that lim j→∞{µ(j)/νr(j)}= 0
for some r= ˜rµ≤rµ+1. We define
ρ(ζ,µ) := max {r(ζ),p(r(ζ)),˜rµ}, (4.13)
wherep(·) is as in Assumptions 2.1. We can now prove the following lemma,
which enables us to control the paths of /tildewiderMNby using fluctuation bounds for
the martingale MN.
Lemma 4.3 Under Assumption 4.2,
/tildewiderMN(t) =MN(t)+/integraldisplayt
0R(t−s)AMN(s)ds.
Proof. From (3.8), we have
R(t−s) =I+/integraldisplayt−s
0R(v)Adv.
Substituting this into (4.11), we obtain
/tildewiderMN(t) =/integraldisplayt
0R(t−s)dMN(s)
18=MN(t)+/integraldisplayt
0/braceleftbigg/integraldisplayt
0R(v)A1[0,t−s](v)dv/bracerightbigg
dMN(s)
=MN(t)+/integraldisplayt
0/braceleftbigg/integraldisplayt
0R(v)A1[0,t−s](v)dv/bracerightbigg
dXN(s)
−/integraldisplayt
0/braceleftbigg/integraldisplayt
0R(v)A1[0,t−s](v)dv/bracerightbigg
F0(xN(s))ds.
It remains to change the order of integration in the double integrals , for
which we use Fubini’s theorem.
In the first, the outer integral is almost surely a finite sum, and at e ach
jump time tXN
lwe havedXN(tXN
l)∈ J. Hence it is enough that, for each i,
mandt,/summationtext
j≥0Rij(t)Ajmis absolutely summable, which follows from Theo-
rem 3.1. Thus we have
/integraldisplayt
0/braceleftbigg/integraldisplayt
0R(v)A1[0,t−s](v)dv/bracerightbigg
dXN(s) =/integraldisplayt
0R(v)A{XN(t−v)−XN(0)}dv.
(4.14)
For the second, the k-th component of R(v)AF0(xN(s)) is just
/summationdisplay
j≥0Rkj(v)/summationdisplay
l≥0Ajl/summationdisplay
J∈JJlαJ(xN(s)). (4.15)
Now, from (3.7), we have 0 ≤Rkj(v)≤µ(j)ewv/µ(k), and
/summationdisplay
j≥0µ(j)|Ajl| ≤µ(l)(2|All|+w), (4.16)
becauseATµ≤wµ. Hence, puttingabsolutevaluesinthesummandsin(4.15)
yields at most
ewv
µ(k)/summationdisplay
J∈JαJ(xN(s))/summationdisplay
l≥0|Jl|µ(l)(2|All|+w).
Now, in view of (4.12) and since ζ(j)≥1 for allj, there is a constant K <∞
such that µ(l)(2|All|+w)≤Kζ(l). Furthermore, ζsatisfies (2.25), so that,
by Corollary 2.5,/summationtext
J∈JαJ(xN(s))/summationtext
l≥0|Jl|ζ(l) is a.s. uniformly bounded in
0≤s≤T. Hence we can apply Fubini’s theorem, obtaining
/integraldisplayt
0/braceleftbigg/integraldisplayt
0R(v)A1[0,t−s](v)dv/bracerightbigg
F0(xN(s))ds=/integraldisplayt
0R(v)A/braceleftbigg/integraldisplayt−v
0F0(xN(s))ds/bracerightbigg
dv,
19and combining this with (4.14) proves the lemma.
We now introduce the exponential martingales that we use to bound the
fluctuations of MN. Forθ∈RZ+bounded and x∈ Rµ,
ZN,θ(t) :=eθTxN(t)exp/braceleftBig
−/integraltextt
0gNθ(xN(s−))ds/bracerightBig
, t≥0,
is a non-negative finite variation local martingale, where
gNθ(ξ) :=/summationdisplay
J∈JNαJ(ξ)/parenleftBig
eN−1θTJ−1/parenrightBig
.
Fort≥0, we have
logZN,θ(t) =θTxN(t)−/integraldisplayt
0gNθ(xN(s−))ds
=θTmN(t)−/integraldisplayt
0ϕN,θ(xN(s−),s)ds, (4.17)
where
ϕN,θ(ξ) :=/summationdisplay
J∈JNαJ(ξ)/parenleftBig
eN−1θTJ−1−N−1θTJ/parenrightBig
,(4.18)
andmN(t) :=N−1MN(t). Note also that we can write
ϕN,θ(ξ) =N/integraldisplay1
0(1−r)D2vN(ξ,rθ)[θ,θ]dr, (4.19)
where
vN(ξ,θ′) :=/summationdisplay
J∈JαJ(ξ)eN−1(θ′)TJ,
andD2vNdenotes thematrixofsecond derivatives withrespect totheseco nd
argument:
D2vN(ξ,θ′)[ζ1,ζ2] :=N−2/summationdisplay
J∈JαJ(ξ)eN−1(θ′)TJζT
1JJTζ2(4.20)
for anyζ1,ζ2∈ Rµ.
Now choose any B:= (Bk, k≥0)∈ R, and define ˜ τ(N)
k(B) by
˜τ(N)
k(B) := inf/braceleftBigg
t≥0:/summationdisplay
J:Jk/negationslash=0αJ(xN(t−))> Bk/bracerightBigg
.
Our exponential bound is as follows.
20Lemma 4.4 For anyk≥0,
P
sup
0≤t≤T∧˜τ(N)
k(B)|mk
N(t)| ≥δ
≤2exp(−δ2N/2BkK∗T).
for all0< δ≤BkK∗T, whereK∗:=J2
∗eJ∗, andJ∗is as in(1.2).
Proof. Takeθ=e(k)β, forβto be chosen later. We shall argue by stopping
the local martingale ZN,θat timeσ(N)(k,δ), where
σ(N)(k,δ) :=T∧˜τ(N)
k(B)∧inf{t:mk
N(t)≥δ}.
Note that eN−1θTJ≤eJ∗, so long as |β| ≤N, so that
D2vN(ξ,rθ)[θ,θ]≤N−2/parenleftBigg/summationdisplay
J:Jk/negationslash=0αJ(ξ)/parenrightBigg
β2K∗.
Thus, from (4.19), we have
ϕN,θ(xN(u−))≤1
2N−1Bkβ2K∗, u≤˜τ(N)
k(B),
and hence, on the event that σ(N)(k,δ) = inf{t:mk
N(t)≥δ} ≤(T∧˜τ(N)
k(B)),
we have
ZN,θ(σ(k,δ))≥exp{βδ−1
2N−1Bkβ2K∗T}.
But since ZN,θ(0) = 1, it now follows from the optional stopping theorem
and Fatou’s lemma that
1≥E{ZN,θ(σ(N)(k,δ))}
≥P/bracketleftBig
sup
0≤t≤T∧˜τ(N)
k(B)mk
N(t)≥δ/bracketrightBig
exp{βδ−1
2N−1Bkβ2K∗T}.
We can choose β=δN/B kK∗T, as long as δ/BkK∗T≤1, obtaining
P
sup
0≤t≤T∧˜τ(N)
k(B)mk
N(t)≥δ
≤exp(−δ2N/2BkK∗T).
Repeating with
˜σ(N)(k,δ) :=T∧˜τ(N)
k(B)∧inf{t:−mk
N(t)≥δ},
21and choosing β=δN/B kK∗T, gives the lemma.
Theprecedinglemmagivesaboundforeachindividualcomponentof MN.
We need first to translate this into a statement for all components simulta-
neously. For ζas in Assumption 4.2, we start by writing
Z(1)
∗:= max
k≥1k−1#{m:ζ(m)≤k};Z(2)
∗:= sup
k≥0µ(k)(|Akk|+1)/radicalbig
ζ(k).(4.21)
Z(2)
∗is clearly finite, because of Assumption 4.2, and the same is true for Z(1)
also, since Zof Assumption 4.2 is at least # {m:ζ(m)≤k}/√
k, for each k.
Then, using the definition (2.24) of τ(N)(a,ζ), note that, for every k,
/summationdisplay
J:Jk/negationslash=0αJ(xN(t))h(k)≤/summationdisplay
J:Jk/negationslash=0αJ(xN(t))h(k)d(J,ζ)
|Jk|ζ(k)≤ah(k)
ζ(k),(4.22)
for anyt < τ(N)(a,ζ) and any h∈ R, and that, for any K ⊆Z+,
/summationdisplay
k∈K/summationdisplay
J:Jk/negationslash=0αJ(xN(t))h(k)≤/summationdisplay
k∈K/summationdisplay
J:Jk/negationslash=0αJ(xN(t))h(k)d(J,ζ)
|Jk|ζ(k)
≤a
mink∈K(ζ(k)/h(k)). (4.23)
From (4.22) with h(k) = 1 for all k, if we choose B:= (a/ζ(k), k≥0), then
τ(N)(a,ζ)≤˜τ(N)
k(B) for allk. For this choice of B, we can take
δ2
k:=δ2
k(a) :=4aK∗TlogN
Nζ(k)=4BkK∗TlogN
N(4.24)
in Lemma 4.4 for k∈κN(a), where
κN(a) :=/braceleftbig
k:ζ(k)≤1
4aK∗TN/logN/bracerightbig
={k:Bk≥4logN/K∗TN},
(4.25)
since then δk(a)≤BkK∗T. Note that then, from (4.12),
/summationdisplay
k∈κN(a)µ(k)δk(a)≤2Z/radicalbig
aK∗TN−1logN, (4.26)
withZas defined in Assumption 4.2, and that
|κN(a)| ≤1
4aZ(1)
∗K∗TN/logN. (4.27)
22Lemma 4.5 If Assumptions 4.2 are satisfied, taking δk(a)andκN(a)as
defined in (4.24)and(4.25), and for any η∈ R, we have
1.P
/uniondisplay
k∈κN(a)/braceleftBig
sup
0≤t≤T∧τ(N)(a,ζ)|mN(t)| ≥δk(a)/bracerightBig
≤aZ(1)
∗K∗T
2NlogN;
2.P
/summationdisplay
k/∈κN(a)Xk
N(t) = 0for all0≤t≤T∧τ(N)(a,ζ)
≥1−4logN
K∗N;
3. sup
0≤t≤T∧τ(N)(a,ζ)
/summationdisplay
k/∈κN(a)η(k)|Fk(xN(t))|
≤aJ∗
mink/∈κN(a)(ζ(k)/η(k)).
Proof. For part 1, use Lemma 4.4 together with (4.24) and (4.27) to give
the bound. For part 2, the total rate of jumps into coordinates w ith indices
k /∈κN(a) is
/summationdisplay
k/∈κN(a)/summationdisplay
J:Jk/negationslash=0αJ(xN(t))≤a
mink/∈κN(a)ζ(k),
ift≤τ(N)(a,ζ),using(4.23)with K= (κN(a))c,which, combinedwith(4.25),
proves the claim. For the final part, if t≤τ(N)(a,ζ),
/summationdisplay
k/∈κN(a)η(k)|Fk(xN(t))| ≤/summationdisplay
k/∈κN(a)η(k)/summationdisplay
J:Jk/negationslash=0αJ(xN(t))J∗,
and the inequality follows once more from (4.23).
LetB(1)
N(a) andB(2)
N(a) denote the events
B(1)
N(a) :=
/summationdisplay
k/∈κN(a)Xk
N(t) = 0 for all 0 ≤t≤T∧τ(N)(a,ζ)
;
B(2)
N(a) :=
/intersectiondisplay
k∈κN(a)/braceleftBig
sup
0≤t≤T∧τ(N)(a,ζ)|mN(t)| ≤δk(a)/bracerightBig
,(4.28)
and setBN(a) :=B(1)
N(a)∩B(2)
N(a). Then, by Lemma 4.5, we deduce that
P[BN(a)c]≤aZ(1)
∗K∗T
2NlogN+4logN
K∗N, (4.29)
23of order O(N−1logN) for each fixed a. Thus we have all the components
ofMNsimultaneously controlled, except on a set of small probability. We
now translate this into the desired assertion about the fluctuation s of/tildewidemN.
Lemma 4.6 If Assumptions 4.2 are satisfied, then, on the event BN(a),
sup
0≤t≤T∧τ(N)(a,ζ)/⌊ard⌊l/tildewidemN(t)/⌊ard⌊lµ≤√aK4.6/radicalbigg
logN
N,
where the constant K4.6depends on Tand the parameters of the process.
Proof. From Lemma 4.3, it follows that
sup
0≤t≤T∧τ(N)(a,ζ)/⌊ard⌊l/tildewidemN(t)/⌊ard⌊lµ (4.30)
≤sup
0≤t≤T∧τ(N)(a,ζ)/⌊ard⌊lmN(t)/⌊ard⌊lµ+ sup
0≤t≤T∧τ(N)(a,ζ)/integraldisplayt
0/⌊ard⌊lR(t−s)AmN(s)/⌊ard⌊lµds.
For the first term, on BN(a) and for 0 ≤t≤T∧τ(N)(a,ζ), we have
/⌊ard⌊lmN(t)/⌊ard⌊lµ≤/summationdisplay
k∈κN(a)µ(k)δk(a)+/integraldisplayt
0/summationdisplay
k/∈κN(a)µ(k)|Fk(xN(u))|du.
The first sum is bounded using (4.26) by 2 Z√aK∗T N−1/2√logN, the sec-
ond, from Lemma 4.5 and (4.25), by
TaJ∗
mink/∈κN(a)(ζ(k)/µ(k))≤Z(2)
∗2J∗/radicalbigg
Ta
K∗/radicalbigg
logN
N.
For the second term in (4.30), from (3.7) and (4.16), we note that
/⌊ard⌊lR(t−s)AmN(s)/⌊ard⌊lµ≤/summationdisplay
k≥0µ(k)/summationdisplay
l≥0Rkl(t−s)/summationdisplay
r≥0|Alr||mr
N(s)|
≤ew(t−s)/summationdisplay
l≥0µ(l)/summationdisplay
r≥0|Alr||mr
N(s)|
≤ew(t−s)/summationdisplay
r≥0µ(r){2|Arr|+w}|mr
N(s)|.
24OnBN(a) and for 0 ≤s≤T∧τ(N)(a,ζ), from (4.12), the sum for r∈κN(a)
is bounded using
/summationdisplay
r∈κN(a)µ(r){2|Arr|+w}|mr
N(s)|
≤/summationdisplay
r∈κN(a)µ(r){2|Arr|+w}δr(a)
≤/summationdisplay
r∈κN(a)µ(r){2|Arr|+w}/radicalBigg
4aK∗TlogN
Nζ(r)
≤(2∨w)Z/radicalbig
4aK∗T/radicalbigg
logN
N.
The remaining sum is then bounded by Lemma 4.5, on the set BN(a) and
for 0≤s≤T∧τ(N)(a,ζ), giving at most
/summationdisplay
r/∈κN(a)µ(r){2|Arr|+w}|mr
N(s)|
≤/summationdisplay
r/∈κN(a)µ(r){2|Arr|+w}/integraldisplays
0|Fr(xN(t))|dt
≤(2∨w)saJ∗
mink/∈κN(a)(ζ(k)/µ(k){|Akk|+1})
≤(2∨w)Z(2)
∗2J∗/radicalbigg
Ta
K∗/radicalbigg
logN
N.
Integrating, it follows that
sup
0≤t≤T∧τ(N)(a,ζ)/integraldisplayt
0/⌊ard⌊lR(t−s)AmN(s)/⌊ard⌊lµds
≤(2T∨1)ewT/braceleftBigg/radicalbig
4aK∗TZ+Z(2)
∗J2J∗/radicalbigg
Ta
K∗/bracerightBigg/radicalbigg
logN
N,
and the lemma follows.
This has now established the control on sup0≤t≤T/⌊ard⌊l/tildewidemN(t)/⌊ard⌊lµthat we need,
in order to translate (4.10) into a proof of the main theorem.
25Theorem 4.7 Suppose that (1.2),(1.3),(3.1),(3.2)and(4.1)are all satis-
fied, and that Assumptions 2.1 and 4.2 hold. Recalling the defi nition(4.13)
ofρ(ζ,µ), forζas given in Assumption 4.2, suppose that S(N)
ρ(ζ,µ)(0)≤NC∗
for some C∗<∞.
Letxdenote the solution to (4.2)with initial condition x(0)satisfying
Sρ(ζ,µ)(x(0))<∞. Thentmax=∞.
Fix any T, and define ΞT:= sup0≤t≤T/⌊ard⌊lx(t)/⌊ard⌊lµ. If/⌊ard⌊lxN(0)−x(0)/⌊ard⌊lµ≤
1
2ΞTe−(w+k∗)T, wherek∗:=ewTK(µ,F;2ΞT), then there exist constants c1,c2
depending on C∗,Tand the parameters of the process, such that for all N
large enough
P/parenleftBigg
sup
0≤t≤T/⌊ard⌊lxN(t)−x(t)/⌊ard⌊lµ>/parenleftBigg
ewT/⌊ard⌊lxN(0)−x(0)/⌊ard⌊lµ+c1/radicalbigg
logN
N/parenrightBigg
ek∗T/parenrightBigg
≤c2logN
N. (4.31)
Proof. AsS(N)
ρ(ζ,µ)(0)≤NC∗, it follows also that S(N)
r(0)≤NC∗for all
0≤r≤ρ(ζ,µ). Fix any T < tmax, takeC:= 2(C∗+k04T)ek01T, and observe
that, for r≤ρ(ζ,µ)∧r(2)
max, and such that p(r)≤ρ(ζ,µ), we can take
C′′
rT≤/tildewideCrT:={2(C∗∨1)+kr4T}e(kr1+Ckr2)T, (4.32)
in Theorem 2.4, since we can take C∗to bound CrandC′
r. In particular,
r=r(ζ) as defined in Assumption 4.2 satisfies both the conditions on r
for (4.32) to hold. Then, taking a:={k2+k1/tildewideCr(ζ)T}b(ζ)in Corollary 2.5, it
follows that for some constant c3>0, on the event BN(a),
P[τ(N)(a,ζ)≤T]≤c3N−1.
Then, from (4.29), for some constant c4,P[BN(a)c]≤c4N−1logN. Here,
the constants c3,c4depend on C∗,Tand the parameters of the process.
We now use Lemma 4.6 to bound the martingale term in (4.10). It fol-
lows that, on the event BN(a)∩ {τ(N)(a,ζ)> T}and on the event that
/⌊ard⌊lxN(s)−x(s)/⌊ard⌊lµ≤ΞTfor all 0≤s≤t,
/⌊ard⌊lxN(t)−x(t)/⌊ard⌊lµ≤/parenleftBigg
ewT/⌊ard⌊lxN(0)−x(0)/⌊ard⌊lµ+√aK4.6/radicalbigg
logN
N/parenrightBigg
+k∗/integraldisplayt
0/⌊ard⌊lxN(s)−x(s)/⌊ard⌊lµds,
26wherek∗:=ewTK(µ,F;2ΞT). Then from Gronwall’s inequality, on the
eventBN(a)∩{τ(N)(a,ζ)> T},
/⌊ard⌊lxN(t)−x(t)/⌊ard⌊lµ≤/parenleftBigg
ewT/⌊ard⌊lxN(0)−x(0)/⌊ard⌊lµ+√aK4.6/radicalbigg
logN
N/parenrightBigg
ek∗t,
(4.33)
for all 0≤t≤T, provided that
/parenleftBigg
ewT/⌊ard⌊lxN(0)−x(0)/⌊ard⌊lµ+√aK4.6/radicalbigg
logN
N/parenrightBigg
≤ΞTe−k∗T.
This is true for all Nsufficiently large, if /⌊ard⌊lxN(0)−x(0)/⌊ard⌊lµ≤1
2ΞTe−(w+k∗)T,
which we have assumed. We have thus proved (4.31), since, as show n above,
P(BN(a)c∪{τ(N)(a,ζ)> T}c) =O(N−1logN).
We now use this to show that in fact tmax=∞. Forx(0) as above, we
can take xj
N(0) :=N−1⌊Nxj(0)⌋ ≤xj(0), so that S(N)
ρ(ζ,µ)(0)≤NC∗forC∗:=
Sρ(ζ,µ)(x(0))<∞. Then, by (4.13), lim j→∞{µ(j)/νρ(ζ,µ)(j)}= 0, so it fol-
lowseasilyusing boundedconvergence that /⌊ard⌊lxN(0)−x(0)/⌊ard⌊lµ→0asN→ ∞.
Hence, for any T < t max, it follows from (4.31) that /⌊ard⌊lxN(t)−x(t)/⌊ard⌊lµ→D0
asN→ ∞, fort≤T, with uniform bounds over the interval, where ‘ →D’
denotes convergence in distribution. Also, by Assumption 4.2, ther e is a con-
stantc5such that /⌊ard⌊lxN(t)/⌊ard⌊lµ≤c5N−1S(N)
rµ(t) for each t, whererµ≤r(2)
maxand
rµ≤ρ(ζ,µ). Hence, using Lemma 2.3 and Theorem 2.4, sup0≤t≤2T/⌊ard⌊lxN(t)/⌊ard⌊lµ
remains bounded in probability as N→ ∞. Hence it is impossible that
/⌊ard⌊lx(t)/⌊ard⌊lµ→ ∞asT→tmax<∞,implyingthatinfact tmax=∞forsuchx(0).
Remark . The dependence on the initial conditions is considerably compli-
cated by the way the constant Cappears in the exponent, for instance in the
expression for /tildewideCrTin the proof of Theorem 4.7. However, if kr2in Assump-
tions 2.1 can be chosen to be zero, as for instance in the examples be low, the
dependence simplifies correspondingly.
Therearebiologicallyplausiblemodelsinwhichtherestrictionto Jl≥ −1
is irksome. In populations in which members of a given type lcan fight one
another, a natural possibility is to have a transition J=−2e(l)at a rate
proportional to Xl(Xl−1), which translates to αJ=α(N)
J=γxl(xl−N−1),
a function depending on N. Replacing this with αJ=γ(xl)2removes the
27N-dependence, but yields a process that can jump to negative value s ofXl.
For this reason, it is useful to be able to allow the transition rates αJto
depend on N.
Since the arguments inthis paper are not limiting arguments for N→ ∞,
it does not require many changes to derive the corresponding resu lts. Quan-
tities such as A,F,Ur(x) andVr(x) now depend on N; however, Theorem 4.7
continues toholdwithconstants c1andc2thatdo notdepend on N, provided
thatµ,w,ν, theklmfrom Assumption 2.1 and ζfrom Assumption 4.2 can
be chosen to be independent of N, and that the quantities Z(l)
∗from (4.21)
can be bounded uniformly in N. On the other hand, the solution x=x(N)
of (4.2) that acts as approximation to xNin Theorem 4.7 now itself depends
onN, through R=R(N)andF=F(N). IfA(and hence R) can be taken
to be independent of N, and lim N→∞/⌊ard⌊lF(N)−F/⌊ard⌊lµ= 0 for some fixed µ–
Lipschitz function F, a Gronwall argument can be used to derive a bound
for the difference between x(N)and the (fixed) solution xto equation (4.2)
withN-independent RandF. IfAhas to depend on N, the situation is
more delicate.
5 Examples
We begin with some general remarks, to show that the assumptions are sat-
isfied in many practical contexts. We then discuss two particular ex amples,
those of Kretzschmar (1993) and of Arrigoni (2003), that fitte d poorly or
not at all into the general setting of Barbour & Luczak (2008), th ough the
other systems referred to in the introduction could also be treate d similarly.
In both of our chosen examples, the index jrepresents a number of individ-
uals — parasites in a host in the first, animals in a patch in the second —
and we shall for now use the former terminology for the preliminary, general
discussion.
Transitions that can typically be envisaged are: births of a few para sites,
which may occur either in the same host, or in another, if infection is b eing
represented; births and immigration of hosts, with or without para sites; mi-
gration of parasites between hosts; deaths of parasites; death s of hosts; and
treatment of hosts, leading to the deaths of many of the host’s pa rasites. For
births of parasites, there is a transition X→X+J, whereJtakes the form
Jl= 1;Jm=−1;Jj= 0, j/ne}ationslash=l,m, (5.1)
28indicating that one m-host has become an l-host. For births of parasites
within a host, a transition rate of the form bl−mmXmcould be envisaged,
withl > m, the interpretation being that there are Xmhosts with parasite
burdenm, each of which gives birth to soffspring at rate bs, for some small
values of s. For infection of an m-host, a possible transition rate would be
of the form
Xm/summationdisplay
j≥0N−1Xjλpj,l−m,
since an m-host comes into contact with j-hosts at a rate proportional to
their density in the host population, and pjrrepresents the probability of a
j-host transferring rparasites to the infected host during the contact. The
probability distributions pj·can be expected to be stochastically increasing
inj. Deaths of parasites also give rise to transitions of the form (5.1),
but now with l < m, the simplest form of rate being just dmXmforl=
m−1, though d=dmcould also be chosen to increase with parasite burden.
Treatment of a host would lead to values of lmuch smaller than m, and
a rate of the form κXmfor the transition with l= 0 would represent fully
successful treatment of randomly chosen individuals. Births and d eaths of
hosts and immigration all lead to transitions of the form
Jl=±1;Jj= 0, j/ne}ationslash=l. (5.2)
Fordeaths, Jl=−1, anda typical ratewould be d′Xl. Forbirths, Jl= 1, and
a possible rate would be/summationtext
j≥0Xjb′
jl(withl= 0 only, if new-born individuals
are free of parasites). For immigration, constant rates λlcould be supposed.
Finally, for migration of individual parasites between hosts, transit ions are
of the form
Jl=Jm=−1;Jl+1= 1;Jm−1= 1;Jj= 0, j/ne}ationslash=l,m,l+1,m−1,
(5.3)
a possible rate being γmXmN−1Xl.
For all the above transitions, we can take J∗= 2 in (1.2), and (1.3) is
satisfied in biologically sensible models. (3.1) and (3.2) depend on the wa y in
which the matrix Acan be defined, which is more model specific; in practice,
(3.1) is very simple to check. The choice of µin (3.2) is influenced by the
need to have (4.1) satisfied. For Assumptions 2.1, a possible choice o fνis to
takeν(j) = (j+1) for each j≥0, withS1(X) then representing the num-
ber of hosts plus the number of parasites. Satisfying (2.5) is then e asy for
29transitions only involving the movement of a single parasite, but in gen eral
requires assumptions as to the existence of the r-th moments of the distri-
butions of the numbers of parasites introduced at birth, immigratio n and
infection events. For (2.6), in which transitions involving a net reduc tion
in the total number of parasites and hosts can be disregarded, th e parasite
birth events are those in which the rates typically have a factor mXmfor
transitions with Jm=−1, withmin principle unbounded. However, at such
events, an m-individual changes to an m+sindividual, with the number s
of offspring of the parasite being typically small, so that the value of JTνr
associated with this rate has magnitude mr−1; the product mXmmr−1, when
summed over m, then yields a contribution of magnitude Sr(X), which is al-
lowable in(2.6). Similar considerations showthat theterms N−1S0(X)Sr(X)
accommodate the migration rates suggested above. Finally, in orde r to have
Assumptions 4.2 satisfied, it is in practice necessary that Assumptio ns 2.1
are satisfied for large values of r, thereby imposing restrictions on the dis-
tributions of the numbers of parasites introduced at birth, immigra tion and
infection events, as above.
5.1 Kretzschmar’s model
Kretzschmar (1993) introduced a model of a parasitic infection, in which the
transitions from state Xare as follows:
J=e(i−1)−e(i)at rate Niµxi, i ≥1;
J=−e(i)at rate N(κ+iα)xi, i≥0;
J=e(0)at rate Nβ/summationtext
i≥0xiθi;
J=e(i+1)−e(i)at rate Nλxiϕ(x), i ≥0,
wherex:=N−1X,ϕ(x) :=/⌊ard⌊lx/⌊ard⌊l11{c+/⌊ard⌊lx/⌊ard⌊l1}−1withc >0, and/⌊ard⌊lx/⌊ard⌊l11:=/summationtext
j≥1j|x|j; here, 0≤θ≤1, andθidenotes its i-th power (our θcorresponds
to the constant ξin [7]). Both (1.2) and (1.3) are obviously satisfied. For
Assumptions (3.1), (3.2) and (4.1), we note that equation corresp onding
to (1.5) has
Aii=−{κ+i(α+µ)};AT
i,i−1=iµandAT
i0=βθi, i≥2;
A11=−{κ+α+µ};AT
10=µ+βθ;
A00=−κ+β, i≥1,
30with all other elements of the matrix equal to zero, and
Fi(x) =λ(xi−1−xi)ϕ(x), i≥1;F0(x) =−λx0ϕ(x).
Hence Assumption (3.1) isimmediate, andAssumption (3.2)holds for µ(j) =
(j+1)s, for any s≥0, withw= (β−κ)+. For the choice µ(j) =j+1,F
maps elements of RµtoRµ, and is also locally Lipschitz in the µ-norm, with
K(µ,F;Ξ) =c−2λΞ(2c+Ξ).
For Assumptions 2.1, choose ν=µ; then (2.5) is a finite sum for each
r≥0. Turning to (2.6), it is immediate that U0(x)≤βS0(x). Then, for
r≥1,
/summationdisplay
i≥0λϕ(N−1X)Xi{(i+2)r−(i+1)r} ≤λS1(X)
S0(X)/summationdisplay
i≥0rXi(i+2)r−1
≤r2r−1λSr(X),
since, by Jensen’s inequality, S1(X)Sr−1(X)≤S0(X)Sr(X). Hence we can
takekr2=kr4= 0 and kr1=β+r2r−1λin (2.6), for any r≥1, so that
r(1)
max=∞. Finally, for (2.7),
V0(x)≤(κ+β)S0(x)+αS1(x),
so thatk03=κ+β+αandk05= 0, and
Vr(x)≤r2(κS2r(x)+αS2r+1(x)+µS2r−1(x)+22(r−1)λS2r−1(x))+βS0(x),
so that we can take p(r) = 2r+1,kr3=β+r2{κ+α+µ+22(r−1)λ}, and
kr5= 0 for any r≥1, and so r(2)
max=∞. In Assumptions 4.2, we can clearly
takerµ= 1 and ζ(k) = (k+1)7, givingr(ζ) = 8,b(ζ) = 1 and ρ(ζ,µ) = 17.
5.2 Arrigoni’s model
Inthemetapopulation model ofArrigoni (2003), thetransitions f romstate X
are as follows:
J=e(i−1)−e(i)at rateNixi(di+γ(1−ρ)), i ≥2;
J=e(0)−e(1)at rateNx1(d1+γ(1−ρ)+κ);
J=e(i+1)−e(i)at rateNibixi, i ≥1;
J=e(0)−e(i)at rateNxiκ, i ≥2;
J=e(k+1)−e(k)+e(i−1)−e(i)at rateNixixkργ, k ≥0, i≥1;
31as before, x:=N−1X. Here, the total number N=/summationtext
j≥0Xj=S0(X) of
patches remains constant throughout, and the number of animals in any one
patch changes by at most one at each transition; in the final (migra tion)
transition, however, the numbers in two patches change simultane ously. In
the above transitions, γ,ρ,κare non-negative, and ( di),(bi) are sequences of
non-negative numbers.
Once again, both (1.2) and (1.3) are obviously satisfied. The equatio n
corresponding to (1.4) can now be expressed by taking
Aii=−{κ+i(bi+di+γ)};AT
i,i−1=i(di+γ);AT
i,i+1=ibi, i≥1;
A00=−κ,
with all other elements of Aequal to zero, and
Fi(x) =ργ/⌊ard⌊lx/⌊ard⌊l11(xi−1−xi), i≥1;F0(x) =−ργx0/⌊ard⌊lx/⌊ard⌊l11+κ,
where we have used the fact that N−1/summationtext
j≥0Xj= 1. Hence Assumption (3.1)
is again immediate, and Assumption (3.2) holds for µ(j) = 1 with w= 0,
forµ(j) =j+ 1 with w= max i(bi−di−γ−κ)+(assuming ( bi) and (di)
to be such that this is finite), or indeed for µ(j) = (j+1)swith any s≥2,
with appropriate choice of w. With the choice µ(j) =j+1,Fagain maps
elements of RµtoRµ, and is also locally Lipschitz in the µ-norm, with
K(µ,F;Ξ) = 3ργΞ.
To check Assumptions 2.1, take ν=µ; once again, (2.5) is a finite sum
for each r. Then, for (2.6), it is immediate that U0(x) = 0. For any r≥1,
using arguments from the previous example,
Ur(x)≤r2r−1/braceleftBigg/summationdisplay
i≥1ibixi(i+1)r−1+/summationdisplay
i≥1/summationdisplay
k≥0iργxixk(k+1)r−1/bracerightBigg
≤r2r−1{max
ibiSr(x)+ργS1(x)Sr−1(x)}
≤r2r−1{max
ibiSr(x)+ργS0(x)Sr(x)},
so that, since S0(x) = 1, we can take kr1=r2r−1(maxibi+ργ) andkr2=
kr4= 0 in (2.6), and r(1)
max=∞. Finally, for (2.7), V0(x) = 0 and, for r≥1,
Vr(x)
≤r2/braceleftBig
22(r−1)max
ibiS2r−1(x)+max
i(i−1di)S2r(x)+γ(1−ρ)S2r−1(x)
+ργ(22(r−1)S1(x)S2r−2(x)+S0(x)S2r−1(x))/bracerightBig
+κS2r(x),
32so that we can take p(r) = 2r, and (assuming i−1dito be finite)
kr3=κ+r2{22(r−1)(max
ibi+ργ)+max
i(i−1di)+γ},
andkr5= 0 for any r≥1, andr(2)
max=∞. In Assumptions 4.2, we can again
takerµ= 1 and ζ(k) = (k+1)7, givingr(ζ) = 8,b(ζ) = 1 and ρ(ζ,µ) = 16.
Acknowledgement
We wish to thank a referee for recommendations that have substa ntially
streamlined our arguments. ADB wishes to thank both the Institut e for
MathematicalSciencesoftheNationalUniversityofSingaporeand theMittag–
Leffler Institute for providing a welcoming environment while part of t his
work was accomplished. MJL thanks the University of Z¨ urich for th eir hos-
pitality on a number of visits.
References
[1]Arrigoni, F. (2003). Deterministic approximation of a stochastic
metapopulation model. Adv. Appl. Prob. 35691–720.
[2]Barbour, A. D. andKafetzaki, M. (1993). A host–parasite model
yielding heterogeneous parasite loads. J. Math. Biology 31157–176.
[3]Barbour, A. D. andLuczak, M. J. (2008). Laws of large numbers
for epidemic models with countably many types. Ann. Appl. Probab. 18
2208–2238.
[4]Chow, P.-L. (2007).Stochastic partial differential equations. Chapman
and Hall, Boca Raton.
[5]Eibeck, A. andWagner, W. (2003). Stochastic interacting particle
systems and non-linear kinetic equations. Ann. Appl. Probab. 13845–
889.
[6]Kimmel, M. andAxelrod, D. E. (2002).Branching processes in biol-
ogy.Springer, Berlin.
33[7]Kretzschmar, M. (1993).Comparison ofaninfinite dimensional model
for parasitic diseases with a related 2-dimensional system. J. Math. Anal-
ysis Applics 176235–260.
[8]Kurtz, T. G. (1970). Solutions of ordinary differential equations as
limits of pure jump Markov processes. J. Appl. Probab. 749–58.
[9]Kurtz, T. G. (1971).Limit theorems forsequences ofjumpMarkov pro-
cesses approximating ordinary differential processes. J. Appl. Probab. 8
344–356.
[10]L´eonard, C. (1990).Some epidemic systems are long range interacting
particle systems. In: Stochastic Processes in Epidemic Theory , Eds J.-P.
Gabriel, C. Lef` evre & P. Picard, Lecture Notes in Biomathematics 86
170–183: Springer, New York.
[11]Luchsinger, C. J. (1999).MathematicalModelsofaParasiticDisease,
Ph.D. thesis, University of Z¨ urich.
[12]Luchsinger, C. J. (2001a). Stochastic models of a parasitic infection,
exhibiting three basic reproduction ratios. J. Math. Biol. 42, 532–554.
[13]Luchsinger, C. J. (2001b). Approximating the long term behaviour
of a model for parasitic infection. J. Math. Biol. 42, 555–581.
[14]Pazy, A. (1983).Semigroups of Linear Operators and Applications to
Partial Differential Equations. Springer, Berlin.
[15]Reuter, G. E. H. (1957). Denumerable Markov processes and the
associated contraction semigroups on l.Acta Math. 97, 1–46.
34