Jump Markov Processes
Connexions module m44968

Zdzisław (Gustav) Meglicki, Jr
Office of the VP for Information Technology, Indiana University
RCS: Section-5.tex,v 1.67 2013/07/30 20:31:02 gustav Exp

Copyright © 2013 by Zdzisław Meglicki
July 30, 2013

Abstract

On jump Markov processes. This is just a place-holder at this stage.

Keywords: something or other

Contents

1 Preliminaries
2 Introduction and Summary
3 The Jump Propagator
4 The Jump Kramers-Moyal Equations
 4.1 The Truncation
 4.2 Continuous Markov Process as a Limit
5 The Master Equations
6 The Evolution of the Moments
7 The Next Jump Density Function
8 Homogeneous Jump Markov Processes
 8.1 Exponential jump displacement
  8.1.1 Evolution of Moments
  8.1.2 Kramers-Moyal Equation
  8.1.3 Master Equation
  8.1.4 Quadrature Formula
 8.2 Gaussian jump displacement
  8.2.1 Evolution of Moments
  8.2.2 Kramers-Moyal Equation
  8.2.3 Master Equation
  8.2.4 Quadrature Formula
 8.3 Cauchy-Lorentz jump displacement
  8.3.1 Evolution of Moments
  8.3.2 Kramers-Moyal Equation
  8.3.3 Master Equation
  8.3.4 Quadrature Formula
Index

1 Preliminaries

  1. “Probability Distributions,” Connexions module m43336,
  2. “Introduction to Markov Processes,” Connexionx module m44014,
  3. “Continuous Markov Processes,” Connexions module m44258 (in parallel with “Integral of a Markov Process,” module 44376),
  4. “Integral of a Markov Process,” Connexions module m44376 (in parallel with “Continuous Markov Processes,” module 44258).

2 Introduction and Summary

A jump Markov process proceeds by the Markov system performing finite, that is, not infinitesimal jumps on moving from x(t) to x(t + dt). We will denote the jump by Δx. Although we will be able to demonstrate that the Markov propagator density function for such a process can be fully characterized by two functions, as was also the case with the continuous Markov processes, on account of this finiteness we will no longer be able to truncate Jump Krammers-Moyal equations don’t trunctate. Kramers-Moyal equations. There is no Fokker-Planck equivalent for the jump Markov process.

Because the Kramers-Moyal equations don’t truncate, there is another machinery, the master equations, which are differential-integral equations, that is preferably used in this context. Although still intractable analytically, the master equations provide us with tools for numerical investigations of jump processes.

The jump Markov process propagator density looks as follow

Π Δx|dt|(x,t) = a(x,t)dtw Δx|(x,t) + 1 a(x,t)dtδ Δx, (1)

where

The same propagator density is sometimes expressed as

Π Δx|dt|(x,t) = W Δx|(x,t) dt + 1 ΩΔxW Δx|(x,t) dtdΔxδ Δx, (2)

where W Δx|(x,t) is called the consolidated characterizing function of a jump Markov process. It is the probability that the Markov process at x at t will jump in the next dt by between Δx and Δx + dΔx away from x. Functions a(x,t) and w Δx|(x,t) can be expressed in terms of W Δx|(x,t) as follows

a(x,t) =ΩΔxW Δx|(x,t) dΔx, w Δx|(x,t) = W Δx|(x,t) ΩΔxW Δx|(x,t) dΔx. (3)

The master equations can be expressed in terms of the consolidated characterizing functions in which case they look as follows (forward and backward):

tP (x,t)|(x0,t0) = ΩΔx W Δx|(x Δx,t) P (x Δx,t)|(x0,t0) W Δx|(x,t) P (x,t)|(x0,t0) dΔx (4)
t0P (x,t)|(x0,t0) = ΩΔxW Δx|(x0,t0) P (x,t)|(x0 Δx,t0) P (x,t)|(x0,t0) dΔx (5)

Unlike continuous Markov processes, the jump processes can be characterized by a yet another function, called the next jump density function

𝒥Δx, Δt|x,tdΔtdΔx. (6)

The function is the probability density that a jump Markov process that is at x at t will perform its next jump at about Δt past t, that is, between t + Δt and t + Δt + dΔt landing between x + Δx and x + Δx + dΔx, or, to put it in other words, that the next jump will happen within [Δt, Δt + dΔt] after t landing the system within [Δx, Δx + dΔx] away from x. This function can be expressed in terms of a(x,t) and w x|(x,t) as follows

𝒥Δx, Δt|x,t = a(x,t + Δt)e0Δta(x,t+Δt)dΔt w Δx|(x,t + Δt) . (7)

This will simplify for temporally homogeneous processes to

𝒥Δx, Δt|x,t = a(x)ea(x)Δtw Δx|x. (8)

Completely homogeneous jump Markov processes are particularly interesting, because they’re the simplest possible and so we can say something about them. The two characterizing functions, a and w, become

a x,t = a, w Δx|(x,t) = w Δx, (9)

with the next jump density function simplifying to

𝒥Δx, Δt|(x,t) = aw ΔxeaΔt. (10)

An important thing to notice is that whereas we could characterize completely homogeneous continuous Markov processes (they are called Wiener processes) by two constants, here we have one constant only and one irreducible function of the jump itself.

It is interesting to consider w Δx given by exponential, Gaussian and Cauchy-Lorentz distributions. The systems so defined become quite tractable and, more importantly, applicable to a range of physical phenomena, for example, diffusion and Brownian motion.

In this module we are also going to take another look at quantum mechanics, asking if quantum mechanics can be simulated by jump Markov processes. There was a vigorous discussion about this in the literature.

It all started with a paper by Hardy, Home, Squires and Whitaker, “Realism and the quantum-mechanical two-state oscillator,” published on the 1st of April 1992 in Physical Review A, vol. 45, no. 7, pp. 4267–4270. Nearly two years later Gillespie referred to this paper in his article “Why quantum mechanics cannot be formulated as a Markov process,” published in March 1994 in Physical Review A, vol. 49, no. 3, pp. 1607–1612. A year after that, in May 1995, Garbaczewski and Olkiewicz responded with “Why quantum dynamics can be formulated as a Markov process,” published in Physical Review A, vol. 51, no. 5, pp. 3445–3453, to which Gillespie responded in June 1996 in “Comment on ‘Why quantum dynamics can be formulated as a Markov process’”, published in Physical Review A, vol. 53, no. 6, pp. 4602–4604. Not to be outdone, Gorbaczewski and Olkiewicz published their riposte in August 1996, “Comment on ‘Why quantum mechanics cannot be formulated as a Markov Process’” in Physical Review A, vol. 54, no. 2, pp. 1733–1736. In the same issue of Physical Review A, following the article by Gorbaczewski and Olkiewicz, Gillespie presented his final one-page argument, pp. 1737–1738. But the last word in this exchange belonged to Hardy, Home, Squires and Whitaker, who in their “Comment on ‘Why quantum mechanics cannot be formulated as a Markov process’”, published in October 1997 in Physical Review A, vol. 56, no. 4, pp. 3301–3303, pointed out that they did not claim their process described in the original 1992 paper to be a Markov process. They demonstrated a concrete model that satisfied their requirements (of 1992) and pointed out various problems with Gillespie’s own arguments.

The important moral of the story is that not every stochastic jump process must be a Markov process to begin with. Markovianism is a quite special requirement that may not always apply. Another moral is that some reasoning that Gillespie is so fond of in his papers and in his book (which here we faithfully follow, because it’s an excellent introduction to Markov processes) is not necessarily unquestionable. In particular, the trick of dividing dt into dtn does stretch things a bit. Shouldn’t we treat dt as indivisible instead? In effect the Gillespie’s trick leads to smoothness conditions that may be stronger than necessary. Gillespie’s dt is not infinitesimal enough.

To understand this fascinating discussion though we must master the formalism of jump Markov processes first, and so we begin

3 The Jump Propagator

A jump process found at x at t will most likely stay at x for a while, then it’ll suddenly jump away from x. We can’t normally say how long it’s going to stay at x, but we are interested in processes for which a probability exists that the jump will occur within the next Δt. Not precisely at Δt after t, mind you, but within Δt following t. And so

q(Δt|(x,t)) (11)

is this probability. The probability function q does not have to exist, but if it does not, then there’s not much else we can say about such processes. So we stick to this assumption. There’s one other thing we can obviously assume, namely that q(0|(x,t)) = 0, meaning that the process is stuck at x at t. We will also assume that q(Δt|(x,t)) is a smooth function of times t and Δt. These two assumptions, in combination with the assumption that q exists in the first place, go quite far already. Whether we assume too much by doing so will transpire later, when we attempt to apply the theory to known physical processes. But for the time being, the assumptions are purely operational: we need them so that we can develop a tractable theory. It will turn out eventually that the assumption of Markovianism demands a certain degree of smoothness of q, although it will also turn out that this is a sufficient, not a necessary, condition.

Now, once the process has jumped, it’ll land somewhere, and where it lands may be described by another probability

w Δx|(x,t) dΔx. (12)

This is the probability that once the process has jumped it’ll land at between some finite Δx and Δx + dΔx away from x, that is, between x + Δx and x + Δx + dΔx. And here we assume also that w Δx|(x,t) is a smooth function of time t, again so that the theory is tractable, but it’ll transpire down the road that the assumption of Markovianism demands a certain degree of smoothness of w, although this will be a sufficient, not a necessary, condition too.

Both functions q and wdΔx must be non-negative, because they are probabilities. Additionally wdΔx being probability density must integrate to 1 over Δx.

Theorem 3.1.
If Δt = dt is infinitesimal, then

q(dt|(x,t)) = a(x,t)dt. (13)

The proof of this equation is rooted in the Markov definition of the underlying process and in the proportional function lemma, that was proved in module m44258:

Lemma 3.1 (Proportional Function).
If

  1. f(x) is smooth,
  2. n : f(x) = n(f(xn)), where is the set of natural numbers (positive integers),

then f(x) = αx, where α does not depend on x (but it may depend on other parameters of the problem.

We prove equation (13) as follows.

Proof. We are going to demonstrate that

q dt|(x,t) = nq dt n |(x,t) , (14)

from which equation (13) follows by the use of the Proportional Function lemma. To get there we divide the infinitesimal dt into n portions of length dtn each. If q(dt|(x,t)) is the probability that the system that’s in x at t will jump away from x in the next dt, then 1 q(dt|(x,t)) is the probability that the system will not jump in the next dt. Since we have divided dt into n segments, the probability that the system will not jump in dt must be equal to the product of probabilities that the system will not jump in each time segment, that is

1 q(dt|(x,t)) = i=1n 1 q dt n |(x,ti1) . (15)

Alas, because dt is supposed to be an infinitesimal, the times ti1 are infinitesimally close to each other, so we can replace them all with just t, which yields

1 q(dt|(x,t)) = 1 q dt n |(x,t) n. (16)

Now we make a yet another use of the fact that dt is infinitesimal. Let us recall that q(0|(x,t)) = 0 and since q is smooth, we can readily assume that q dt n |(x,t) is infinitesimal too. Therefore we can approximate

1 q dt n |(x,t) n 1 nq dt n |(x,t) , (17)

wherefrom, in combination with equation (16), equation (14) follows, which ends the proof.

In view of our comments on the discussion between Gillespie, Hardy and his collaborators, and Gorbaczewski and Olkiewicz, it is important that we recognize that, in this case, we have assumed the smoothness of q from the beginning. Whether this assumption is more than is strictly required by Markovianism remains to be seen. The reasoning here is quite similar to how we demonstrated in m44258 that

x(dt, (x,t)) = μ(x,t)dt, var x(dt, (x,t)) = D(x,t)dt (18)

for the continuous Markov processes, wherefrom we deduced that the propagator density function in this case had to be a Gaussian. The assumptions behind equation (13) preclude, for example, a sharp spike in probability q at the end of dt, or indeed, within a finite time interval Δt thereafter.

Given the meaning of q, the meaning of a(x,t)dt is the probability that the process in x at t will jump away from x in the next dt. Furthermore, because dt is infinitesimal, we expect only one jump to occur within dt or none at all. The probability of two jumps occurring would be proportional to (dt)2.

The jump Markov process propagator density function can now be written in the following form

Π Δx|dt|(x,t) dΔx = a(x,t)dtw Δx|(x,t) dΔx + 1 a(x,t)dtδ(Δx)dΔx, (19)

which we read as follows. On the left side we have the probability of the Markov process finding itself removed by Δx from x which it occupied at t upon having advanced the time by dt. As Δx is finite, this is unmistakably a jump propagator. On the right side of the equation, we express the same as the probability that the process in x at t will jump in the next dt times the probability that having jumped it will land by a finite Δx away from x or—and this is what the plus stands for—the probability that the process will not jump, in which case it will stay at x, which is to say, it will displace by Δx = 0, which is what is signified here by the delta function. So, the process will jump, or it will not.

Dividing both sides by dΔx yields

Π Δx|dt|(x,t) = a(x,t)dtw Δx|(x,t) + 1 a(x,t)dtδ(Δx). (20)

In the following, we’re going to look more closely at two issues. First, we’ll derive a formula for q(Δt|(x,t)) in terms of a(x,t). Second, we’re going to see that equation (20) satisfies the Chapman-Komogorov equation to the first order in dt, wherefrom we’ll also obtain a more precise idea as to how smooth the functions a(x,t) and w Δx|(x,t) have to be.

Expression for q(Δt|(x,t))

We cannot infer from equation (13) that
q(Δt|(x,t)) =0Δta(x,t)dt (21)

It is more tricky, because we deal here with probabilities and they have their own special rules. Let ¬q(Δt|(x,t)) be the probability that the system that’s in x at t will not jump away from x in Δt. This probability, of course, is

¬q(Δt|(x,t)) = 1 q(Δt|(x,t)). (22)

The probability that the system still will not jump in the next dt is ¬q(Δt + dt|(x,t)) and it can be decomposed into the product of the probability that the system will not jump in Δt and the probability that it will not jump in the following dt either:

¬q(Δt + dt|(x,t)) = ¬q(Δt|(x,t)) ׬q(dt|(x,t + Δt)) = ¬q(Δt|(x,t))(1 q(dt|(x,t + Δt))) = ¬q(Δt|(x,t))(1 a(x,t + Δt)dt), (23)

wherefrom

d¬q(Δt|(x,t)) ¬q(Δt|(x,t)) = a(x,t + Δt)dt. (24)

This then integrates to

¬q(Δt|(x,t)) = e0Δta(x,t+t)dt . (25)

Now we drop the ¬, reverting to q itself, which yields

q(Δt|(x,t)) = 1 e0Δta(x,t+t)dt . (26)

For the very small Δt dt we should expect a(x,t)dt to be small too. The exponential function then will be approximated by

1 a(x,t)dt (27)

and equation (26) becomes (13), our starting point. We see here that the infinitesimal relationship given by equation (13) itself was not enough to reconstruct the correct relationship between q and a for the finite Δt. Additional information, provided by equation (23), was needed to get it right.

Smoothness of a and w

In order to obtain smoothness conditions on a and w we need to inspect what is required of these two functions to ensure that the propagator density function Π Δx|dt|(x,t) satisfies the Chapman-Kolmogorov condition
Π Δx|dt|(x,t) = ΩΔxΠ Δx Δx|(1 λ)dt|x + Δx,t + λdt Π Δx|λdt|x,tdΔx. (28)

This condition assumes that the infinitesimal dt can be split into two sub-infinitesimals, [0,λdt] and [λdt,dt] and that the process itself must be Markov-divisible on top of these two intervals. This is a weighty assumption, especially given that λ is an arbitrary real number in [0, 1]. It will result in further continuity conditions down the road.

We begin by substituting equation (20) in equation (28). The first Π under the integral becomes

a x + Δx,t + λdt (1 λ)dtw Δx Δx|x + Δx,t + λdt + 1 a x + Δx,t + λdt (1 λ)dtδ Δx Δx. (29)

The second Π is

a x,tλdtw Δx|(x,t) + 1 a(x,t)λdtδ Δx. (30)

We have to multiply them now, but we only keep terms linear in dt, because dt is infinitesimal after all. The first component in each sum is proportional to dt, so we can neglect first times first right away. But the second component of each sum contains a term that is not proportional to dt, which is the δ term itself. In summary, we’ll end up with the first component of the first sum times the delta from the second sum, plus the first component of the second sum times the delta from the first sum, plus a term that’ll contain a product of the two deltas. This last term looks as follows

δ Δx Δxδ Δx1 a x + Δx,t + λdt (1 λ)dt a(x,t)λdt (31)

to which we add

a x + Δx,t + λdt (1 λ)dtw Δx Δx|x + Δx,t + λdtδ Δx + a x,tλdtw Δx|(x,t) δ Δx Δx. (32)

This is to be integrated over Ω Δx. The deltas make the integration easy. δ Δx simply converts Δx to zero and δ Δx Δx converts Δx to Δx, which yields

a x,t + λdt (1 λ)dtw Δx|(x,t + λdt) + a(x,t)λdtw Δx|(x,t) + 1 a x,t + λdt (1 λ)dt a(x,t)λdtδ Δx. (33)

We want this to be equal to

a(x,t)dtw Δx|(x,t) + 1 a(x,t)dtδ Δx. (34)

The requirement imposes certain conditions on a and w. For example, comparing the δ terms yields

a x,t + λdt (1 λ) + a(x,t)λ = a(x,t). (35)

This must hold in the dt 0 limit.

If we assume that

a(x,t + λdt) = a(x,t) + a(x,t) t λdt + 𝒪(λdt)2 (36)

then

a x,t + λdt (1 λ) + a(x,t)λ = (1 λ) a(x,t) + a(x,t) t λdt + λa(x,t) + 𝒪(λdt)2 = a(x,t) + (1 λ)a(x,t) t λdt + 𝒪(λdt)2 dt 0a(x,t). (37)

We add a similar assumption regarding w, that is,

w Δx|(x,t + λdt) = w Δx|(x,t) + w Δx|(x,t) t λdt + 𝒪(λdt)2 (38)

to demonstrate similarly that

a(x,t + λdt)(1 λ)w Δx|(x,t + λdt) + a(x,t)λw Δx|(x,t) dt 0a(x,t)w Δx|(x,t) . (39)

But by now we can see it even without the explicit computation. If both a and w are to be expanded as per equations (36) and (38), the first term, the one that does not involve dt, is simply a(x,t)w Δx|(x,t). The terms that are proportional to λ cancel out. All other terms have dt in them, so they all vanish in the limit dt 0.

Consequently, we see that assumptions given by equations (36) and (38) are sufficient to ensure that the propagator in the form of equation (20) satisfies the Chapman-Kolmogorov condition given by equation (28). However, we have not proven that they are necessary. There may exist less restrictive conditions that would still ensure the satisfaction of equation (28).

Functions a(x,t) and w Δx|(x,t) are called the characterizing functions of the jump Markov process. Whereas the continuous Markov process is also described by two functions (see module m44258, where we call them μ(x,t), the drift function, and D(x,t), the diffusion function), the functions depend on two variables x and t. Here, for jump processes, we have three variables. The third one is Δx, the finite jump length. The resulting description is going to be more elaborate. We will also show that in a certain limit jump processes become continuous processes.

Looking at equation (20) we see that the two functions enter the expression for the propagator in a special way, namely through the product of a and w and then through a itself in the second summand, the one proportional to the Dirac delta. Because w x|(x,t) is the probability density in x, it must integrate to 1. In turn, a(x,t) does not depend on x. We can therefore define

W x|(x,t) = a(x,t)w x|(x,t) , (40)

for which we’ll find that

a(x,t) =Ω(x)W x|(x,t) dx (41)

and

w x|(x,t) = W x|(x,t) Ω(x)W x|(x,t) dx. (42)

Function W x|(x,t) is called the consolidated characterizing function of the jump Markov process and it can be used instead of a and w to encode the jump process propagator density as follows

Π x|dt|(x,t) = W x|(x,t) dt + 1 Ω(x)W x|(x,t) dxdtδ x. (43)

The expression

W x|(x,t) dtdx (44)

is the probability that the system in x at t will jump in the next dt by between x and x + dx away from x.

Now, we switch to one dimention, x x and x x.

Once we have the jump propagator density function, we can calculate propagator moments, namely

Πn(x,t)dt =xnΠ x|dt|(x,t) dx =xn a(x,t)dtw x|(x,t) + (1 a(x,t)dt)δ(x) dx = a(x,t)dtxnw x|(x,t) dx. (45)

Here we have made use of

xnδ xdx = 0 (46)

to kill the second term, the one with 1 a. From the above then

Πn(x,t) = a(x,t)xnw x|(x,t) dx. (47)

Defining the nth moments of w and W, the consolidated characterizing function of the jump process, by

wn(x,t) =xnw x|(x,t) dx (48)

and

Wn(x,t) =xna(x,t)w x|(x,t) dx =xnW x|(x,t) dx, (49)

we can rewrite equation (47) as

Πn(x,t) = a(x,t)wn(x,t) = Wn(x,t). (50)

For the above definitions to be meaningful, the corresponding integrals must be convergent: the moments Πn exist if the corresponding moments of w (or W) exist.

4 The Jump Kramers-Moyal Equations

The Kramers-Moyal equations (forward and backward) were derived in module m44014 from the Kolmogorov equations. They allowed us to express the time derivatives of the Markov process probability density function P (x,t)|(x0,t0) in terms of the propagator density moments, thus providing us with some sort of evolutionary equations, not unlike the Schrödinger equation of quantum mechanics—though, in general, as we’ll see in the case of the jump processes, they may not be explicitly solvable. In the case of the continuous Markov processes, the forward Kramers-Moyal equation turned out to be the familiar Fokker-Planck equation that with some additional non-Markovian assumptions could be converted into the Schrödinger equation—the “non-Markovian” phrase here being important.

The forward Kramers-Moyal equation is

tP (x,t)|(x0,t0) = n=1(1)n n! n xn Πn(x,t)P (x,t)|(x0,t0) . (51)

Substituting equation (50) yields

tP (x,t)|(x0,t0) = n=1(1)n n! n xn a(x,t)wn(x,t)P (x,t)|(x0,t0) or = n=1(1)n n! n xn Wn(x,t)P (x,t)|(x0,t0) . (52)

The backward Kramers-Moyal equation is

t0P (x,t)|(x0,t0) = n=1 1 n!Πn x0,t0 n x0nP (x,t)|(x0,t0) . (53)

Substituting equation (50) yields

t0P (x,t)|(x0,t0) = n=1 1 n!a x0,t0 wn x0,t0 n x0nP (x,t)|(x0,t0) or = n=1 1 n!Wn x0,t0 n x0nP (x,t)|(x0,t0) . (54)

The Kramers-Moyal equations for the jump Markov processes are partial differential equations of the infinite order in the x or x0 variable. Unlike in the continuous Markov process case, they cannot be truncated in general, so they are nothing but trouble. They can be truncated approximately when the jump Markov process can be considered a continuous one, approximately too, in which case they turn into the Fokker-Planck equations. We are going to demonstrate this in the following.

4.1 The Truncation

To see that the Kramers-Moyal equations do not truncate we make use of the concavity of xn for x 0. Concavity means that xn curves away from a tangent line for any x [0,]. In other words, any tangent provides us with a low estimate for xn.

Let us then consider x = a > 0. The slope of the tangent at x = a is

dxn dx x=a = nxn1 x=a = nan1 = α. (55)

The equation for the tangent line is

f(x) = αx + β, (56)

where α is as above and β is given by the requirement that f(a) = an, that is

αa + β = an β = an αa. (57)

In summary,

f(x) = αx + an αa = nan1x + an nan1a = nan1x + (1 n)an (58)

and we can state that

x0a0xn nan1x + (1 n)an. (59)

Now, let us consider a random variable (x,P(x)), such that P(x < 0) = 0. In this case

xn =0xnP(x)dx 0nan1x + (1 n)an P(x)dx = nan10xP(x)dx + (1 n)an0P(x)dx = nan1 x + (1 n)an. (60)

This holds for any a 0, including a = x. So, let us substitute the latter, which yields

xn nxn1x + (1 n)xn = xn. (61)

Now, this holds only for positive random variables, that is variables for which P(x) = 0 if x < 0. But for an arbitrary random variable x,P(x), its even powers satisfy this requirement, therefore we can state that for an arbitrary random variable

x2n x2 n. (62)

Let us return now to the definition of wn(x,t), which is a moment of the characterizing function w x|(x,t), that is

wn(x,t) =xnw x|(x,t) dx, (63)

where w is the probability density of x parametrized by (x,t). The above considerations therefore apply and we can state that

w2n(x,t) w2(x,t) n. (64)

The strict inequality in this case holds whenever w2(x,t) > 0. In turn, w2(x,t) = 0 only if w x|(x,t) = δ(x), which corresponds to there being no jumps at all, which is not an interesting case. And so, excluding this case, we can state that

w2n(x,t) > w2(x,t) n > 0. (65)

This tells us that the successive terms in the Kramers-Moyal equations, at least the even ones, do not become identically zero, implying that the equations do not truncate, and so cannot be used, well, not easily, to solve for P (x,t)|(x0,t0). Of course, we can still state that the infinite sums on the right sides of the equations must be convergent, by construction, because the left sides are finite and well defined.

4.2 Continuous Markov Process as a Limit

If the jumps are short and viewed from a long distance, they may not be discernible individually. In this case we may see the system progress continuously in some random fashion. It is in this limit that the Kramers-Moyal equations for the jump Markov process may turn into the Fokker-Planck equations, but only if certain conditions are met. We are now going to explore what these conditions may be.

If (x(t),P(x)) is a random variable, then z(t) = x(t)Ω, where Ω is a volume within which the Markov process unfolds, is also a random variable with the probability density Q(z) such that

Ω(x)P(x)dx =Ω(z)P(x(z)) dx dzdz =Ω(z)Q(z)dz, (66)

where Ω Ω. From this we obtain

P(x(z)) dx dz = ΩP(x(z)) = Q(z), (67)

or

P(x(z)) = 1 ΩQ(z). (68)

Also, because x = Ωz, we get that

x = 1 Ω zand n xn = 1 Ωn n zn. (69)

For a large volume Ω, 1Ω is small, so this is the parameter we are going to expand the Kramers-Moyal equation in. We begin with the forward form given by equation (52), and substitute QΩ in place of P and zΩ in place of x. The left side becomes

tP (x,t)|(x0,t0) = 1 Ω tQ (z,t)|(z0,t0) . (70)

And on the right side, using the form with moments of the consolidated characteristic function, W x|(x,t), we obtain

n=1(1)n n! n xn Wn(x,t)P (x,t)|(x0,t0) = n=1(1)n n! 1 Ωn n zn Wn(Ωz,t) 1 ΩQ (z,t)|(z0,t0) . (71)

Combining equations (70) and (71) yields

tQ (z,t)|(z0,t0) = n=1(1)n n! n zn Wn(Ωz,t) Ωn Q (z,t)|(z0,t0) . (72)

Whether the right side of this equation can be truncated approximately depends on how Wn(Ωz,t)Ωn scales with Ω . Let us look again at the definition of Wn:

Wn(Ωz,t) =znW z|(Ωz,t) dz. (73)

Now, if, for example,

W z|(Ωz,t) Ω ΩW z|(z,t) (74)

then

Wn(Ωz,t) Ω ΩznW z|(z,t) dz = ΩW n(z,t). (75)

Therefore the right side of equation (72) becomes in the limit Ω

n=1(1)n n! n zn Wn(z,t) Ωn1 Q (z,t)|(z0,t0) = z W1(z,t)Q (z,t)|(z0,t0) + 1 2 2 z2 W2(z,t) Ω Q (z,t)|(z0,t0) + 𝒪1 Ω 2. (76)

Let us compare this with the forward Fokker-Planck equation, which in this case would read

tQ (z,t)|(z0,t0) = z μ(z,t)Q (z,t)|(z0,t0) + 1 2 2 z2 D(z,t)Q (z,t)|(z0,t0) . (77)

Comparing equations 76 and 77 tells us that if we were to identify

W1(z,t) Ω μ(z,t)andW2(z,t) Ω Ω D(z,t) (78)

and assuming that W2Ω was still finite for Ω , whereas the higher terms vanished in the limit, the observed process would look like a continuous Markov process. But this is predicated on equation (74). In other words, it depends on how W scales with Ω.

5 The Master Equations

Since the Kramers-Moyal equations, forward and backward, don’t truncate, in general neither exactly nor approximately, they are not of much use in the case of the jump Markov processes. The master equations, on the other hand, are differential-integral equations that at least can be written in a compact form and that are tractable numerically. Yet, surprisingly perhaps, both Kramers-Moyal and master equations derive from the same Chapman-Kolmogorov equations, that are in essence differential-integral equations too. We arrive at Kramers-Moyal equations by expanding the function in the Chapman-Kolmogorov integral in the Taylor series and replacing the resulting infinite series of integrals with an infinite series in which the integrals are merely encapsulated in Πn:

Ω(x) xnΠ x|dt|(x,t) dt dx = Π n(x,t). (79)

This therefore is a renaming exercise, with the basic problem just swept under the carpet. Little wonder then that for the jump Markov processes it has crawled out to haunt us.

Let us return then to the starting point, to the Chapman-Kolmogorov equation. We commence with the forward one

P (x,t + Δt)|(x0,t0) = Ω(x)P (x,t + Δt)|(x x,t) P (x x,t)|(x 0,t0) dx. (80)

What this equation says is as follows. We have a system that transits from (x0,t0) to (x,t + Δt). We posit that at time t the system must pass through some x x on its way to x, and that therefore its trajectory is

from(x0,t0)through(x x,t)to(x,t + Δt). (81)

For any given x x, the probability of reaching (x,t + Δt) through (x x,t) from (x0,t0) is

P (x,t + Δt)|(x x,t) P (x x,t)|(x 0,t0) . (82)

But as x may run all over Ω(x), that is, the process may transfer through x x1 Ω or through x x2 Ω or through x x3 Ω and so on, on its way to x, the or logic operators translate into the addition of the related probabilities, so we need to sum equation (82) over all possible xs, which is how we arrive at the integral in equation (80).

This is quite similar to how quantum mechanics is constructed by the means of Feynman path integrals, but here we sum over probabilities, whereas in the Feynman method we sum over probability amplitudes. Otherwise the resulting mathematics and methodology are really similar. Why it is the probability amplitudes that we sum in quantum physics instead of probabilities themselves, as the conventional logic might dictate, is the central, unexplained puzzle of the quantum world.

Looking at the forward Chapman-Kolmogorov equation we may think that it implies a certain continuity of the system’s evolution on its way from x0 at t0 to x at t + Δt and we may ask how this is compatible with the idea of a jump Markov process. But jump Markov processes are still described by equation (80). If a jump is to occur from x0 to x x1 then we may find that most P (x x,t)|(x 0,t0) are zero, with the exception of P (x x1,t)|(x 0,t0), where P itself spikes in the form of the Dirac delta. P may assume the form of multiple spikes scattered over Ω(x) with different weights for each location x xi. The Chapman-Kolmogorov equation then says that the probability of the system transiting from x0 at t0 to x at t + Δt is a sum of probabilities that correspond to jumps from x0 to x xi times probabilities of jumps from x xi to x:

P (x,t + Δt)|(x0,t0) = i=1nP (x,t + Δt)|(x x i,t) P (x x i,t)|(x 0,t0) . (83)

But what, we may ask next, if the system is such that it stays put between t and t0 and doesn’t jump at all, and then only it jumps directly to x at t + Δt? In this case, the probability P (x x,t)|(x 0,t0) would be δ(x). The point is that the system must be somewhere at t, whether the jump has occurred or not. The formalism of Markov processes assumes a continuous existence of the system in time, whereas it may jump in space or move in some other way amenable to stochastic analysis.

Let us get back to the Chapman-Kolmogorov equation (80). We can reformulate it also as follows

P (x,t)|(x0,t0) = Ω(x)P (x,t)|(x0 + x,t 0 + Δt0) P (x0 + x,t 0 + Δt0)|(x0,t0) dx. (84)

This equation says something similar to equation (80). It says that on its way from x0 at t0 to x at t the system passes through some x0 + x at t0 + Δt0, where x can be any x in Ω. The probability of getting from x0 to x should therefore be a sum of probabilities of passing through x0 + x at t0 + Δt0 on the way, which are

P (x,t)|(x0 + x,t 0 + Δt0) P (x0 + x,t 0 + Δt0)|(x0,t0) . (85)

Now we are going to shrink Δt and Δt0 in both equations to dt and dt0 and make use of the fact that in this limit the relevant probabilities also shrink to the propagator density, namely

Π x|dt|(x,t) = P (x + x,t + dt)|(x,t) . (86)

We apply this to equation (80) first replacing the P term with Δt under the integral:

P x,t + Δt)|(x x,t) Δt dtΠ x|dt|(x x,t) . (87)

Similarly we replace the Δt0 term in equation (84):

P (x0 + x,t 0 + Δt0)|(x0,t0) Δt0 dt0Π x|dt 0|(x0,t0) . (88)

This yields two equivalent though differently expressed equations

P (x,t + dt)|(x0,t0) = Ω(x)Π x|dt|(x x,t) P (x x,t)|(x 0,t0) dx (89)

and

P (x,t)|(x0,t0) = Ω(x)P (x,t)|(x0 + x,t 0 + dt0) Π x|dt 0|(x0,t0) dx. (90)

Now, we do not intend to subtract equation (90) from equation (89) to form the time derivative, because one has dt in it whereas the other one has dt0, so this wouldn’t work. Instead we’ll use equation (89) to form the forward master equation and we’ll use equation (90) later to form the backward master equation. We’ll do this by substituting the equation for the jump propagator in place of Π in both,

Π x|dt|(x,t) = a(x,t)dtw x|(x,t) + 1 a(x,t)dtδ(x). (91)

The substitution converts equation (89) to

P (x,t + dt)|(x0,t0) = Ω(x) a(x x,t)dtw x|(x x,t) + 1 a(x x,t)dtδ(x) P (x x,t)|(x 0,t0) dx = Ω(x)a(x x,t)dtw x|(x x,t) P (x x,t)|(x 0,t0) dx + 1 a(x,t)dtP (x,t)|(x0,t0) , (92)

because the delta in the second summand has killed the x.

Let us observe that there is a stand-alone P (x,t)|(x0,t0) in equation (92). So, we simply subtract P (x,t)|(x0,t0) from both sides of equation (92), which reduces the second summand to aP term only. Then we divide both sides by dt and obtain the forward master equation for the jump Markov process,

tP (x,t)|(x0,t0) = Ω(x)a(x x,t)w x|(x x,t) P (x x,t)|(x 0,t0) dx a(x,t)P (x,t)|(x0,t0) . (93)

Remembering that a(x,t)w x|(x,t) is the consolidated characterizing function of the jump Markov process, W x|(x,t), and that

Ω(x)w x|(x,t) dx = 1, (94)

(it also holds for + x, but in this case it is customary to use x instead) lets us rewrite equation (93) as

tP (x,t)|(x0,t0) = Ω(x) W x|(x x,t) P (x x,t)|(x 0,t0) W x|(x,t) P (x,t)|(x 0,t0) dx. (95)

Now we return to equation (90) and substitute equation (91) in place of Π:

P (x,t)|(x0,t0) = Ω(x)P (x,t)|(x0 + x,t 0 + dt0) Π x|dt 0|(x0,t0) dx = Ω(x)P (x,t)|(x0 + x,t 0 + dt0) a(x0,t0)dt0w x|(x 0,t0) + 1 a(x0,t0)dt0 δ xdx = dt0Ω(x)P (x,t)|(x0 + x,t 0 + dt0) a(x0,t0)w x|(x 0,t0) dx + 1 a(x0,t0)dt0 P (x,t)|(x0,t0 + dt0) . (96)

We notice there is a pure P (x,t)|(x0,t0 + dt0) term in the second summand. We move this term to the left side of the equation, divide both sides by dt0 and take the limit dt0 0, which yields the backward master equation for the jump Markov process,

t0P (x,t)|(x0,t0) = Ω(x)a(x0,t0)w x|(x 0,t0) P (x,t)|(x0 + x,t 0) dx a(x0,t0)P (x,t)|(x0,t0) . (97)

Again, making use of the normalization condition given by equation (94), but this time with + x, we can rewrite the above equation in terms of the consolidated characterizing function of the jump Markov process as follows:

t0P (x,t)|(x0,t0) = Ω(x)W x|(x 0,t0) P (x,t)|(x0 + x,t 0) P (x,t)|(x0,t0) dx. (98)

6 The Evolution of the Moments

Because the evolution equations for jump Markov processes, namely, the Kramers-Moyal equations and the master equation, do not in general simplify, the evolutions of the moments remain as we discussed in module m44014 with one improvement. Let us recall that for the jump processes Πn(x,t) = Wn(x,t), see equation (50), where

Wn(x,t) =xna(x,t)w x|(x,t) dx =xnW x|(x,t) dx. (99)

The reason for this was that the no-jump term in the propagator integrated to zero when calculating the moments on account of the Dirac delta function of x. Therefore we can substitute Wn in every place in the moments evolution equations where Πn appears.

And so, the general equation for the evolution of the nth moment of a jump Markov process is

d dt xn(t) = k=1nn k xnk(t)W k x(t),t. (100)

Specifically, the equations for the evolution of the mean, variance and covariance are

evolution of the mean

d dt x(t) = W1 x(t),t, x(t0) = x0, (101)
evolution of the variance

d dt var x(t) = 2 x(t)W1(x(t),t) 2 x(t) W1(x(t),t) + W2(x(t),t) , var(x(t0)) = 0, (102)
evolution of the covariance

d dt2 cov x(t1),x(t2) = x(t1)W1(x(t2),t2) x(t1) W1(x(t2),t2) , cov x(t1),x(t2 = t1) = var x(t1) . (103)

The integral of the Markov process is not in itself a Markov process. It is a stochastic process that remembers its past over which the integral is accumulated. But on account of its close relationship to the Markov process, of which it is the integral, the process is tractable. We discussed such processes and their general properties in module m44376 The observation that Πn = Wn still applies, therefore we can write down the relevant equations as follows.

The equation for the evolution of the moments of the integral process

s(t) =t0tx(t)dt (104)

is

d dt sn(t) = n sn1(t)x(t) , sn(t 0) = 0, (105)

where the cross moments sn1(t)x(t) have to be evaluated by solving

d dt sn(t)xm(t) = n sn1(t)xm+1(t) + k=1mm k sn(t)xmk(t)W k(x(t),t) , sn(t 0)xm(t 0) = 0. (106)

We specify these for the two lowest moments, which yields

integral: evolution of the mean

s(t) =t0t x(t) dt. (107)
integral: evolution of the variance

var s(t) = 2t0t cov s(t),x(t) dt, (108)

where cov s(t),x(t) must be found by solving

d dt cov s(t),x(t) = var x(t) + s(t)W1 x(t),t s(t) W1 x(t),t) , cov s(t0),x(t0) = 0. (109)
integral: evolution of the covariance

cov s(t1),s(t2) =t1t2 cov s(t1),x(t) dt + var s(t 1) , (110)

where cov s(t1),x(t) must be found by solving

d dt2 cov s(t1),x(t2) = s(t1)W1 x(t2),t2 s(t1) W1 x(t2),t2 , cov s(t1),x(t2 = t1) = cov s(t1),x(t1) . (111)

7 The Next Jump Density Function

The next jump density function provides us with another way to characterize Markov jump processes. Whereas

Π Δx|dt|(x,t) dΔx (112)

is the probability of an excursion from x at t by Δx ±1 2dΔx after the passing of dt, the next jump function

𝒥 (Δx, Δt)|(x,t) dΔxdΔt (113)

is the probability that the next jump will actually occur at Δt ±1 2dΔt past t and the system will jump by Δx ±1 2dΔx away from x.

We can express 𝒥 in terms of the two characteristic functions of the jump process, a and w. Let us recall their physical meaning:

  1. a(x,t)dt is the probability that the jump will occur within dt past t. In particular, we have found that the probability of a jump occurring within a finite Δt past t was
    1 exp 0Δta(x,t + t)dt, (114)

    see equation (26). Whereas

    exp 0Δta(x,t + t)dt (115)

    was the probability that the jump would not occur within Δt past t, see equation (25)

  2. w Δx(x,t) dΔx is the probability that the jump, once it has happened at t, would take the system away by Δx ±1 2dΔx from x.

The probability that the jump will happen at Δt ±1 2dΔt after t and that it will take the system by Δx ±1 2dΔx away from x, in other words, 𝒥 (Δx, Δt)|(x,t) dΔxdΔt, is equal to

  1. the probability that the system will not jump between t and t + Δt, which is
    exp 0Δta(x,t + t)dt, (116)

    times

  2. the probability that the system will jump within dΔt after t + Δt, which is
    a(x,t + Δt)dΔt, (117)

    times

  3. the probability that the system will land, on having performed the jump, by Δx ±1 2dΔx away from x, which is
    w Δx|(x,t + Δt) dΔx. (118)

In summary

𝒥 (Δx, Δt)|(x,t) dΔxdΔt = exp 0Δta(x,t + t)dta(x,t + Δt)dΔtw Δx|(x,t + Δt) dΔx, (119)

which upon the division of both sides by dΔtdΔx and reordering of multiplicants on the right side of the equation yields

𝒥 (Δx, Δt)|(x,t) = a(x,t + Δt) exp 0Δta(x,t + t)dtw Δx|(x,t + Δt) . (120)

As can be seen from above, function 𝒥 naturally splits into

𝒥t Δt|(x,t) = a(x,t + Δt) exp 0Δta(x,t + t)dt (121)

and

𝒥x Δx|Δt|(x,t) = w Δx|(x,t + Δt) , (122)

so that

𝒥 (Δx, Δt)|(x,t) = 𝒥t Δt|(x,t) 𝒥x Δx|Δt|(x,t) (123)

Because w Δx|(x,t + Δt) integrates to 1 over Δx, it is easy to see that

𝒥t Δt|(x,t) =Ω(Δx)𝒥 (Δx, Δt)|(x,t) dΔx. (124)

Then

𝒥x Δx|Δt|(x,t) = 𝒥 (Δx, Δt)|(x,t) Ω(Δx)𝒥 (Δx, Δt)|(x,t) dΔx. (125)

8 Homogeneous Jump Markov Processes

It is useful to observe that for temporally homogeneous processes the exp term becomes ea(x)Δt and the expressions for 𝒥t and 𝒥x simplify to

𝒥t Δt|(x,t) = a(x)ea(x)Δt (126)

and

𝒥x Δx|Δt|(x,t) = w Δx|x. (127)

We can rewrite equation (126) as follows

𝒥t Δt|(x,t) = 1 τ(x)eΔtτ(x), (128)

where τ(x) = 1a(x) is Pausing time.the average pausing time in x. Indeed, let us observe that equation (126) is an exponential distribution with a standard deviation of 1a(x), which avails us of the interpretation Interpretation of aof a, in the case of temporally homogeneous jump processes, as the inverse of the average pausing time at x. We should also observe that in this case, the probability density of the jump displacement from x, represented by 𝒥x, is independent of the average pausing time τ.

For the remainder Completely homogeneous jump Markov processesof this section we shall focus on completely homogeneous jump Markov processes. In this case the two characteristic functions, a(x,t) and w Δx|(x,t) reduce to a constant, a (or 1τ), and a function of Δx, namely, w(Δx). We observe the important difference when contrasted with completely homogeneous continuous Markov processes, in which case, the Wiener process is fully characterized by two constants, the drift coefficient μ and the diffusion coefficient D.

The probability Interpretation of w(Δx)density of the jump displacement from x,

𝒥x(Δx) = w(Δx) (129)

is in principle arbitrary, although it must satisfy the smoothness conditions discussed in Section 3. But it is both interesting and highly applicable to consider three special cases

exponential

w(Δx) = 1 σeΔxσ, (130)
Gaussian

w(Δx) = 1 2πσ2e(Δx)2(2σ2), (131)
Cauchy-Lorentz
w(Δx) = 1 π σ (Δx)2 + σ2. (132)

First, however, Kramers-Moyal and master equationswe are going to specify the time evolution equations for the completely homogeneous case.

We recall that because of the complete homogeneity of the system, the probability density of transition from (x0,t0) to (x,t) depends on differences x x0 and t t0 only,

P((x,t)|(x0,t0)) = P((x x0,t t0)|(0, 0)). (133)

Looking at equation (50) we find that the propagator density function moments Πn are constants,

Πn = awn = Wn, (134)

where

wn =Δxnw ΔxdΔx. (135)

The forward Kramers-Moyal equation (51) therefore is

tP (x,t)|(x0,t0) = a n=1(1)n n! wn n xnP (x,t)|(x0,t0) (136)

and the forward master equation (93) becomes

tP (x,t)|(x0,t0) = aw(Δx)P (x Δx,t)|(x 0,t0) dΔx aP (x,t)|(x0,t0) . (137)

The backward equations are not in this case independent and can be obtained from the forward ones above by the following substitutions

P t0 = P t , P x0 = P x , P (x,t)|(x0 + Δx,t) = P (x Δx,t)|(x0,t0) . (138)

Neither Quadrature solution for Pequation (136) nor (137) are in general tractable, but let us recall another way to find P (x,t)|(x0,t0) that happens to work for completely homogeneous processes. In this case we have the formula

P (x,t)|(x0,t0) = 1 2πeik(xx0) Π ̂ k|dt|(,) ndk, (139)

where n is the number of time slices that divide the interval [t0,t] such that

ndt = t t0 (140)

and

Π ̂ k|dt|(,) =Π Δx|dt|(,) eikΔxdΔx (141)

is the propagator density Fourier transform.

We normally take a limit n which converts a polynomial into the exp function so that the final expression is tractable. And so it is in this case.

Given that for the completely homogeneous jump Markov process

Π Δx|dt|(,) = aw(Δx)dt + (1 adt)δ(Δx) (142)

we evaluate Π ̂ as follows

Π ̂ k|dt|(,) = aw(Δx)dt + (1 adt)δ(Δx) eikΔxdΔx = adtw(Δx)eikΔxdΔx + 1 adt = 1 + (ŵ(k) 1)adt, (143)

where

ŵ(k) =w(Δx)eikΔxdΔx (144)

is the Fourier transform of w(Δx).

The next step is to evaluate Π ̂n. Here we make use of the compound interest formula by Bernoulli

ex = lim n1 + x nn. (145)

Looking at equation (143) we can see that it can be rewritten in this form, namely

Π ̂ k|dt|(,) = 1 + (ŵ(k) 1)adt = 1 + (ŵ(k) 1)a(ndt) n = 1 + (ŵ(k) 1)a(t t0) n (146)

wherefrom

lim nΠ ̂ k|dt|(,) n = exp (ŵ(k) 1)a(t t 0) , (147)

and on substitution to equation (139)

P (x,t)|(x0,t0) = 1 2πe(ŵ(k)1)a(tt0)eik(xx0)dk. (148)

To progress further we must have an explicit expression for w(Δx). For most functions of interest the integral in equation (148) is far from trivial, but it is tractable numerically.

Finally, Evolution of the momentslet us write down equations for the evolution of the moments and sums for the completely homogeneous jump Markov processes.

And so, equation (100) becomes

d dt xn(t) = a k=1nw kn k xnk(t) . (149)

evolution of the mean

Equation (101) solves to
x(t) = x0 + (t t0)aw1. (150)
evolution of the variance

Equation (102) solves to
var x(t) = (t t0)aw2. (151)
evolution of the covariance

And equation (103) solves to
cov x(t1),x(t2) = (t1 t0)aw2. (152)

These results are trivial.

Equations for the integral of the Markov process are more complicated. Equation (105) remains unchanged. Equation (106) for the cross-moments becomes

d dt sn(t)xm(t) = n sn1(t)xm+1(t) + a k=1mw km k sn(t)xmk(t) . (153)

integral: evolution of the mean

The evolution of the mean of the integral is given by equation (107), which in this case becomes
s(t) =t0t x(t) ,dt =t0t x 0 + (t t 0)aw1 dt = x0(t t0) + aw1 2 t t0 2. (154)
integral: evolution of the variance

To find the evolution of the variance we must first solve the differential equation for the covariance of the cross-moment, equation (109), which here becomes
d dt cov s(t),x(t) = var x(t) = (t t0)aw2 (155)

which solves to We will need this intermediate result to evaluate the evolution of the covariance too. See equation (159).

cov s(t),x(t) = aw2 2 t t0 2 (156)

with the initial condition set to zero. Now, equation (108) solves to

var s(t) = 2t0t cov s(t),x(t) dt = 2t0taw2 2 t t 0 2dt = aw2 3 t t0 3. (157)
integral: evolution of the covariance

Finally, to find the expression for cov s(t1),s(t2) we must solve equation (111). But in this case, it is
d dt2 cov s(t1),x(t2) = 0 (158)

because W1 = aw1 is a constant. Therefore cov s(t1),s(t2) is defined by its initial condition

cov s(t1),s(t2) = cov s(t1),x(t1) = aw2 2 t1 t0 2 (159)

from equation (156) above. In this way we find, following equation (110), that

cov s(t1),s(t2) =t1t2 cov s(t1),x(t) dt + var s(t 1) =t1t2 aw2 2 t1 t0 2dt + aw2 3 t1 t0 3 = aw2 2 t1 t0 2 t 2 t1 + aw2 3 t1 t0 3 = aw2 t1 t0 2 1 2 t2 t1 + 1 3 t1 t0 . (160)

As we now turn to specific examples, our methodology will be as follows. Every formula in this section, so far, has been expressed in terms of wn, the moments of w(Δx). But w(Δx) is a probability density function, and the functions that are of interest to us here have all been discussed in module m43336 and the corresponding moments evaluated. So, there won’t be so much computation involved.

8.1 Exponential jump displacement

The jump displacement probability density function w(Δx) is given by

w(Δx) = 1 σeΔxσif Δx 0 0 if Δx < 0. (161)

Now we look up what we learned about this distribution in module m43336. There we found that

wn = Δxn w =Δxnw(Δx)dΔx = n!σn. (162)

8.1.1 Evolution of Moments

This lets us rewrite equations for the evolution of the moments right away. Let us note that we will require the first two moments only:

w1 = σ w2 = 2σ2. (163)

And so we find that equations (150)-(152) translate to

x(t) = x0 + t t0 aσ, var x(t) = t t0 2aσ2, cov x(t1),x(t2) = t1 t0 2aσ2, (164)

and equations (154), (157) and (160) for the integral of the jump Markov process translate to

s(t) = t t0 x0 + aσ 2 t t0 2, var s(t) = 2aσ2 3 t t0 3, cov s(t1),s(t2) = 2aσ2 t 1 t0 2 1 2 t2 t1 + 1 3 t1 t0 , (165)

8.1.2 Kramers-Moyal Equation

The forward Kramers-Moyal equation becomes

tP (x,t)|(x0,t0) = a n=1σn n xnP (x,t)|(x0,t0) (166)

because 1n! in equation (136) cancels with n! in equation (162).

Now we refer to Connexions module m44258 in which we discussed continuous Markov processes and, in particular, the Wiener process, which is a completely homogeneous continuous Markov process. Let us recall that in this case we had

x(t) = x0 + (t t0)μ, var x(t) = (t t0)D, cov x(t1),x(t2) = t1 t0 D, (167)

where μ (drift) and D (diffusion) are constants. Comparing equation (164) with equation (167) we notice the similarity upon having identified

aσ = μ, 2aσ2 = D. (168)

It is easy to see, comparing with the corresponding equations for the Wiener process in Connexions module m44376 that the same substitutions will also work for s(t), var s(t) and cov s(t1),s(t2). The reason for this is easy to see. Let us write down the first two terms of the forward Kramers-Moyal equation for the completely homogeneous jump Markov process with the exponential displacement density function:

tP (x,t)|(x0,t0) = aσ xP (x,t)|(x0,t0) + 2aσ2 2 2 x2P (x,t)|(x0,t0) + (169)

and compare this with the forward Fokker-Planck equation for the Wiener process

tP (x,t)|(x0,t0) = μ xP (x,t)|(x0,t0) + D 2 2 x2P (x,t)|(x0,t0) . (170)

The two equations become identical on substitution given by equation (168) and assuming that we can truncate higher terms in equation (169). This would be the case if σ is sufficiently small, so small that σ2 σ3. Because σ is the average length of a jump,

Δx = σ, (171)

we infer that for exponentially distributed jumps with a very short average displacement, the completely homogeneous jump Markov processes look like Wiener processes.

It is interesting to invert equation (168). This yields

1 τ = a = 2μ2 D σ = D 2μ. (172)

Let us recall now that a is the inverse of τ, the average pausing time in x, but here, because a does not depend on x it is simply the inverse of the average pausing time. And σ, as we stated above, is the average length of the jump. So, for σ to be small we must have small diffusion or large drift or both. This then implies that a must be large, but this then means that the average pausing time τ = σμ is even smaller than σ. In summary, for a completely homogeneous jump Markov system to look like a continuous system, we must have very short jumps that happen very quickly–the pausing time must be small.

Of course, for the time being, these observations pertain to the jump displacement density function being exponential.

8.1.3 Master Equation

The forward master equation (137) becomes

tP (x,t)|(x0,t0) = a σ0eΔxσP (x Δx,t)|(x 0,t0) dΔx aP (x,t)|(x0,t0) , (173)

because w(Δx) is 0 for negative Δx. It provides us with a numerical procedure for finding P (x,t + Δt)|(x0,t0) assuming that P (x,t)|(x0,t0) is a known function of x at t.

8.1.4 Quadrature Formula

The quadrature solution for P (x,t)|(x0,t0) is given by equation (148), which we recall here for convenience

P (x,t)|(x0,t0) = 1 2π exp a(ŵ(k) 1)(t t 0) eik(xx0)dk. (174)

As we pointed out before, to progress, we must evaluate

ŵ(k) =w(Δx)eikΔxdΔx, (175)

in this case

1 σ0eΔxσeikΔxdΔx = 1 σ0 exp Δx 1 σ + ikdΔx. (176)

We change Δx to Δz = Δx 1 σ + ik which yields

1 σ 1 1 σ + ik0eΔzdΔz = 1 σ 1 1 σ + ik = 1 1 + ikσ, (177)

and so

ŵ(k) 1 = 1 1 + ikσ 1 + ikσ 1 + ikσ = ikσ 1 + ikσ. (178)

It is usually desirable to get rid of the imaginary unit from the denominator. We do this by multiplying ω̂(k) 1 by 1 = (1 ikσ)(1 ikσ):

ŵ(k) 1 = ikσ 1 + ikσ 1 ikσ 1 ikσ = ikσ + k2σ2 1 + k2σ2 . (179)

We substitute equation (179) into the exp term in the integral of equation (174), which yields

P (x,t)|(x0,t0) = 1 2π exp a(t t 0)ikσ + k2σ2 1 + k2σ2 exp ik(x x0) dk = 1 2π exp ik (x x 0) aσ(t t0) 1 + k2σ2 exp ak2σ2(t t 0) 1 + k2σ2 dk. (180)

Now, let us recall that

eiϕ = cos ϕ + i sin ϕ. (181)

Because of the asymmetry of sin, the corresponding integral vanishes, which lives us with the real part only, the cos one. Furthermore, because cos is symmetric, we don’t need to integrate from to . We can integrate from 0 to and multiply by two, which leaves us with

P (x,t)|(x0,t0) = 1 π0 cos k (x x 0) aσ(t t0) 1 + k2σ2 exp ak2σ2(t t 0) 1 + k2σ2 dk. (182)

This does not look at a glance like an undoable integral. It is somewhat similar to

0ea2k2 cos(bk)dk = π 2a eb2(4a2) (183)

so there may be a way to crack it. We could expect P (x,t)|(x0,t0) to contain e(xx0)2(tt0), perhaps in a limit and with some other complications.

8.2 Gaussian jump displacement

Following our previous example, the first step is to retrieve the relevant formulae from module m43336.

This time our jump displacement density function is

w(Δx) = 1 2πσ2e(Δx)2(2σ2). (184)

Incidentally, the reason we write 12πσ2 instead of 1(σ2π) is to draw the eye to the obvious variable substitution in the integral

Δx2σ2 Δz. (185)

The corresponding moments of the jump displacement density function are

wn = (Δx)n w =(Δx)nw(Δx)dΔx = n! (n2)!σ2 2 nfor n = 0, 2, 4, 0 for n = 1, 3, 5,. (186)

8.2.1 Evolution of Moments

The evolution of the mean, the variance and the covariance require the first two moments only. From the above

w1 = 0, w2 = σ2. (187)

Therefore equations (150)-(152) translate to

x(t) = x0, var x(t) = (t t0)aσ2, cov x(t1),x(t2) = (t1 t0)aσ2. (188)

That x(t) is just x0, that is, constant, may be inferred from the symmetry of w(Δx). In this case the system may just as easily to the left as to the right of x0.

Equations (154), (157) and (160) for the integral of the jump Markov process yield

s(t) = (t t0)x0, var s(t) = aσ2 3 (t t0)3, cov s(t1),s(t2) = aσ2(t 1 t0)2 1 2(t2 t1) + 1 3(t1 t0) . (189)

8.2.2 Kramers-Moyal Equation

The right side of the Kramers-Moyal equation

a n=1(1)n n! wn n xnP (x,t)|(x0,t0) (190)

is different from zero for even n only. Therefore we replace n with 2n to the right of the sum and run n itself from 1 through . This yields

tP (x,t)|(x0,t0) = a n=1 1 (2n)! (2n)! n! σ2 2 n 2n x2nP (x,t)|(x0,t0) = a n=1 σ2n 2nn! 2n x2nP (x,t)|(x0,t0) = aσ2 2 2 x2P (x,t)|(x0,t0) + (191)

In the limit of small sigma, we can introduce approximate truncation of higher terms and then this is a diffusion equation with the diffusion coefficient of aσ22. Or, if we were to compare to the Fokker-Planck equation, we’d have D = aσ2.

How come there is no drift? This is because wn are non-zero for even values of n only, so there is no Px term.

8.2.3 Master Equation

Looking at the master equation (93) we replace w Δx|(x Δx,t) with the exponential distribution in Δx (this is where the 12πσ2 comes from) and a(x Δx,t) is a constant. So we obtain

tP (x,t)|(x0,t0) = a 2πσ2eΔx22σ2 P (x Δx,t)|(x0,t0) dΔx aP (x,t)|(x0,t0) . (192)

8.2.4 Quadrature Formula

The quadrature formula (148) calls for evaluation of

ŵ(k) =w(Δx)eikΔxdΔx, (193)

which here becomes

1 2πσ2e(Δx)2(2σ2)eikΔxdΔx. (194)

In Connexions module m43336, we demonstrated that

eikxeax2 dx = π aek2(4a). (195)

This result is directly applicable here upon the following substitutions:

x Δx, a 1 2σ2. (196)

Therefore

4a 2 σ2, π a 2πσ2. (197)

The latter cancels with 12πσ2 in front of the integral in equation (194), which yields

ŵ(k) = e(kσ)22. (198)

Thus the quadrature formula becomes

P (x,t)|(x0,t0) = 1 2πe(ŵ(k)1)a(tt0)eik(xx0)dk = ea(tt0) 2π ea(tt0) exp((kσ)22)eik(xx0)dk = ea(tt0) π 0ea(tt0) exp((kσ)22) cos(k(x x 0))dk, (199)

where we have made use of the nonsymmetry of sin(k(x x0)) and of the symmetry of cos(k(x x0)), with respect to k, thus integrating from 0 to only and multiplying the result by 2.

8.3 Cauchy-Lorentz jump displacement

8.3.1 Evolution of Moments

8.3.2 Kramers-Moyal Equation

8.3.3 Master Equation

8.3.4 Quadrature Formula

Index

Brownian motion, 1

compound interest formula, 2

Gillespie, Daniel T., 3

Hardy, Lucien, 4

infinitesimality of dt, 5

Markov process
    Chapman-Kolmogorov equations, 6, 7, 8
    continuous, 9
        Fokker-Planck equation, 10, 11
    integral of, 12
        covariance, 13
        cross moments, 14
        mean, 15
        moments, 16
        variance, 17
    jump, 18
        characterizing functions, 19, 20
        consolidated characterizing function, 21
        covariance, 22
        homogeneous, 23
        Kramers-Moyal equations, 24, 25, 26
        master equations, 27
        mean, 28
        next jump density function, 29, 30
        propagator density, 31, 32, 33, 34
        tractability, 35
        variance, 36

Proportional Function Lemma, 37

quadrature, 38

Schrödinger equation, 39