Piecewise-deterministic Markov processes (PDMPs) represent a class of stochastic models, that have recently received extensive research attention. They were introduced by Davis [7] as a general and various practical systems in which randomness is limited to jumps. These processes are governed by deterministic semiflows and mentioned jumps which change the state as well the semiflows which will determine evolution of the system. PDMPs have proven to be effective mathematical models for many phenomena, such as gene expression [14] and population dynamics [1]. Results concerning ergodicity and asymptotic stability of these systems are already quite extensively researched [1, 2, 3, 8, 9, 5, 16].
Having the asymptotic stability of such processes, natural questions arise regarding the limit theorems. In [11], we focused on demonstrating a law of the iterated logarithm for some PDMPs in which the intensity of jumps depends on the state of the system. In this paper we prove the central limit theorem (CLT) for this class of processes.
In [4], a criterion for the CLT was introduced, and an example of a model with a constant intensity of jumps was given. The difference between our model and the one considered in [4] lies in the assumption that the intensity of jumps follows an exponential distribution with a constant intensity parameter. This assumption seems to be too restrictive concerning biological models, such as gene expression, in which the relation between the intensity of jumps and the state of the system is visible. It turns out that the criterion for the CLT from [4] is also applicable in the case we are considering.
The paper contains three sections. In Section 2, basic definitions and notations are introduced. In Section 3, we start with an intuitive description of the considered model, and then we formulate strict definitions. We also provide conditions that were used in [6] to show exponential ergodicity. The last section contains the proof of the CLT for the model described earlier, and it is strongly influenced by the methods used in [4].
Let ℝ be the set of all real numbers and let ℝ+ = [0, ∞). Let (E, ρ) be a Polish space with the σ-field ℬ(E) of all Borel subsets of E. By B(E) we denote the space of all bounded, Borel measurable functions f : E → ℝ, equipped with the supremum norm and we list two of its subsets: C(E) and Lip(E) consisting of all continuous functions and Lipschitz-continuous functions, respectively. A continuous function V : E → ℝ+ is called Lyapunov function if it is bounded on bounded sets and for some x0 ∈ E
Let ℳs(E) be the set of all finite, countably additive functions on ℬ(E). By ℳ+(E) and ℳ1(E) we denote the subsets of ℳs(E) consisting of all non-negative measures and all probability measures, respectively. We write
The set ℳs(E) will be considered with the Fortet-Mourier norm ([12, 13]), given by
As usual, by B(x, r) we denote the open ball in E centered at x and radius r > 0. For the fixed set A ⊂ E we define the indicator function 𝟙A : E → {0, 1} as 𝟙A(x) = 1 for x ∈ A and 𝟙A(x) = 0 otherwise.
A function 𝒦: E × ℬ(E) → [0, 1] is called a (sub)stochastic kernel if for each A ∈ ℬ(E), x ↦ 𝒦(x, A) is a measurable map on E, and for each x ∈ E, A ↦ 𝒦(x, A) is a (sub)probability Borel measure on ℬ(E).
An operator P : ℳ+(E) → ℳ+(E) is called a Markov operator if:
P (α1μ1 + α2μ2) = α1Pμ1 + α2Pμ2 for α1, α2 ∈ ℝ+, μ1, μ2 ∈ ℳ+(E),
Pμ(E) = μ(E) for μ ∈ ℳ+(E).
A Markov operator P is called regular if there is a linear operator U : B(E) → B(E), called the dual operator to P, such that
Let us note that the operator U can be extended to a linear operator defined on the space of all bounded below Borel functions B̅(E) so that (1) holds for all f ∈ B̅(E).
A regular Markov operator P is called Feller if U f ∈ C(E) for every f ∈ C(E).
We say that μ∗ ∈ ℳ+(E) is invariant with respect to P if Pμ∗ = μ∗.
If there exists an invariant measure μ∗ ∈ ℳ1(E) and a constant β ∈ [0, 1) such that, for every
It is well known that for every stochastic kernel 𝒦 and any fixed measure μ ∈ ℳ1(E), we can always define on suitable probability space, say (Ω, ℱ, Probμ), a discrete-time homogeneous Markov chain {Xn}n∈ℕ0 for which
We say that a time-homogeneous Markov chain evolving on the space E2 (endowed with the product topology) is a Markovian coupling of some stochastic kernel 𝒦: E × ℬ(E) → [0, 1] whenever its stochastic kernel B : E2 × ℬ(E2) → [0, 1] satisfies
Let {Xn}n∈ℕ0 be a Markov chain evolving in E with transition kernel 𝒦 and initial distribution μ. Suppose μ∗∈ ℳ1(E) is the unique invariant measure of the Markov operator P corresponding to the kernel given by (2). For any n ∈ ℕ and any Borel function g : E → ℝ define
Let g : E → ℝ be a Borel function such that 〈g2, μ∗〉 < ∞. By definition, {g(Xn)}n∈ℕ0 satisfies the CLT condition, if σ2(ḡ) < ∞ and Φsn(ḡ) converges weakly to 𝒩(0; σ2(ḡ)), as n → ∞.
Proving the CLT we will use the following theorem proved in [4].
Let {Xn}n∈ℕ0 be a time-homogeneous Markov chain with values in E and let 𝒦: E × ℬ(E) → [0, 1] be the transition law of {Xn}n∈ℕ0 which satisfies the following conditions:
- (B0)
The Markov operator P corresponding to 𝒦 has the Feller property.
- (B1)
There exist a Lyapunov function V : E → ℝ+ and constants a ∈ (0, 1) and b ∈ (0, ∞) such that
In addition, assume that there is a substochastic kernel Q : E2 × ℬ(E2) → [0, 1] satisfying\matrix{{U{V^2}\left(x \right) \le {{(aV\left(x \right) + b)}^2}} \hfill & {for\,every\,x \in E} \hfill \cr}. for x, y ∈ E, A ∈ ℬ(E).\matrix{{Q\left({x,y,A \times E} \right) \le {\cal K}\left({x,A} \right)} \hfill & {{\rm{and}}} \hfill & {Q\left({x,y,E \times A} \right) \le {\cal K}\left({y,A} \right)} \hfill \cr} - (B2)
There exist F ⊂ E2 and q ∈ (0, 1) such that
for (x, y) ∈ F.\matrix{{{\rm{supp}}\,Q\left({x,y, \cdot} \right) \subset F} \hfill & {and} \hfill & {\int_{{E^2}} {\rho \left({u,v} \right)Q\left({x,y,du \times dv} \right) \le q\rho \left({x,y} \right)}} \hfill \cr} - (B3)
For G(r) = {(x, y) ∈ F : ρ(x, y) ≤ r}, r > 0, we have
\mathop {\inf}\limits_{(x,y) \in F} Q\left({x,y,G\left({q\rho \left({x,y} \right)} \right)} \right) > 0. - (B4)
There exist constants ν ∈ (0, 1] and l > 0 such that
\matrix{{Q\left({x,y,{E^2}} \right) \ge 1 - l\rho {{(x,y)}^\nu}} \hfill & {for\,every\,(x,y) \in F} \hfill \cr}. - (B5)
There is a coupling
of 𝒦 with transition law B ≥ Q such that for some R > 0 and{\{\left({X_n^{\left(1 \right)},X_n^{\left(2 \right)}} \right)\}_{n \in {{\mathbb{N}}_0}}} we can find ζ ∈ (0, 1) and C̅ > 0 satisfyingK: = \{\left({x,y} \right) \in F:V\left(x \right) + V\left(y \right) < R\} \matrix{{{{\mathbb{E}}_{\left({x,y} \right)}}\left({{\zeta^{- {\sigma_K}}}} \right) \le \bar C\,\,\,\,\,whenever\,\,V\left(x \right) + V\left(y \right) < 4b{{(1 - a)}^{- 1}},} \cr {where\,\,{\sigma_K} = \inf \left\{{n \in {\mathbb{N}}:{X_n} \in K} \right\}.} \cr}
Let
Let (Y, ρ) be a Polish space and let I = {1, ..., N} for a fixed positive integer N. We define X : = Y × I and consider the space (X, ρc), where
Our considerations are conducted for a discrete-time dynamical system related to the stochastic process {(Y (t), ξ(t))}t∈ℝ+, which evolves through random jumps in the space X. We proceed to provide a description of this process.
Assume that we have a finite collection of maps, called semiflows, Πi : ℝ+ × Y → Y, i ∈ I, which are continuous with respect to each variable and satisfy for every i ∈ I and each y ∈ Y, the following conditions:
The process {Y (t)}t∈ℝ+ evolves between jumps according to one of the transformation Πi, whose index i is determined by {ξ(t)}t∈ℝ+.
Let πij : Y → [0, 1], i, j ∈ I be a matrix of continuous functions such that
Just after every jump, the semiflow is changed from Πi to Πj in accordance with πij and at the moment of a jump the process {Y (t)}t∈ℝ+ shifts to the new state by a function ω(θ, ·): Y → Y, which is randomly chosen from a given set {ω(θ, ·) : θ ∈ Θ}. We assume that Y × Θ ∋ (y, θ) ↦ ω(θ, y) ∈ Y is continuous and that the probability of choosing ω(θ, ·) is related with density functions Θ ∋ θ ↦ p(θ, y), y ∈ Y, such that (θ, y) ↦ p(θ, y) is continuous. Moreover, we assume that the intensity of jumps is given by a Lipschitz-continuous function λ : Y → (0, ∞), which satisfies the following conditions:
The evolution of {(Y (t), ξ(t))}t∈ℝ+ can be described as follows. Assume that the process starts at some point (y0, i0) ∈ Y × I. Then
Assuming that t0 = 0, we get
Let us emphasize that in this work we study only the sequence of random variables given by the post-jump locations of the process {(Y (t), ξ(t))}t∈ℝ+, namely the process {(Yn, ξn)n∈ℕ0, where Yn = Y (τn), ξn = ξ(τn) for n ∈ ℕ0 and τn is a random variable describing the jump time tn.
We can now give the formal description of the model. Let us set a probability space (Ω, ℱ, Probμ) and define {(Yn, ξn)}n∈ℕ0 as follows. Let (Y0, ξ0): Ω → X be a random variable with arbitrary and fixed distribution μ ∈ ℳ1(X). Further, let us define by induction the sequences of random variables {τn}n∈ℕ0, {ξn}n∈ℕ, {ηn}n∈ℕ and {Yn}n∈ℕ, which describe the sequences {tn}n∈ℕ, {in}n∈ℕ, {θn}n∈ℕ and {yn}n∈ℕ respectively, such that the following conditions are fulfilled:
The sequence τn : Ω → [0, ∞), n ∈ ℕ0, where τ0 = 0, is strictly increasing such that τn → ∞ a.e., and the increments Δτn = τn − τn−1 are mutually independent and their conditional distributions are given by
where y ∈ Y, i ∈ I and L is given by\matrix{{{\rm{Pro}}{{\rm{b}}_\mu}(\Delta {\tau_{n + 1}} \le t|{Y_n} = y\,\,{\rm{and}}\,\,{\xi_n} = i) = 1 - {e^{- L\left({t,y,i} \right)}}} \hfill & {{\rm{for}}\,t \in {{\mathbb{R}}_ +}} \hfill \cr}, (6) L\left({t,y,i} \right) = \mathop \int \nolimits_0^t \lambda \left({{\Pi_i}\left({s,y} \right)} \right)ds. The sequence ξn : Ω → I, n ∈ ℕ satisfies the following condition
\matrix{{{\rm{Pro}}{{\rm{b}}_\mu}({\xi_n} = j|{Y_n} = y,\,\,{\xi_{n - 1}} = i) = {\pi_{ij}}\left(y \right)} \hfill & {{\rm{for}}\,i,j \in I,y \in Y} \hfill \cr}. ηn : Ω → Θ, n ∈ ℕ, is defined as follows
for all D ∈ ℬ(Θ) and y ∈ Y.{\rm{Pro}}{{\rm{b}}_\mu}({\eta_{n + 1}} \in D|{\Pi_{{\xi_n}}}\left({\Delta {\tau_{n + 1}},{Y_n}} \right) = y) = \int_D {p\left({\theta,y} \right)d\theta} Yn : Ω → Y, n ∈ ℕ, are given in following way
\matrix{{{Y_{n + 1}} = \omega \left({{\eta_{n + 1}},{\Pi_{{\xi_n}}}\left({\Delta {\tau_{n + 1}},{Y_n}} \right)} \right)} \hfill & {{\rm{for}}\,n \in {{\mathbb{N}}_{\rm{0}}}} \hfill \cr}.
We can easily check that the process {(Yn, ξn)}n∈ℕ0 is a time-homogeneous Markov chain with phase space X such that the evolution of the distributions μn(·) := Probμ((Yn, ξn) ∈ ·) can be described by the Markov operator P : ℳ+(E) → ℳ+(E) given by
We apply the following assumptions:
- (A1)
There is y∗ ∈ Y such that
\matrix{{\mathop {\sup}\limits_{y \in Y} \int \nolimits_0^\infty {e^{\underline \lambda t}}\int_\Theta {\rho (\omega (\theta,{\Pi_i}(t,{y_*})),{y_*})p\left({\theta,{\Pi_i}\left({t,y} \right)} \right)d\theta dt < \infty}} \hfill & {{\rm{for}}\,i\, \in \,I.} \hfill \cr} - (A2)
There exist constants γ ∈ ℝ, L > 0, and a bounded on bounded subsets of Y function ℒ: Y → ℝ+ such that the following inequality is satisfied
for t ∈ ℝ+, y1, y2 ∈ Y, i, j ∈ I, where Ψ(i, j) is given by (4).\rho \left({{\Pi_i}\left({t,{y_1}} \right),{\Pi_j}\left({t,{y_2}} \right)} \right) \le L{e^{\gamma t}}\rho \left({{y_1},{y_2}} \right) + t{\cal L}\left({{y_2}} \right)\Psi \left({i,j} \right) - (A3)
There exists a constant M > 0 such that
\int_\Theta {\matrix{{\rho \left({\omega \left({\theta,{y_1}} \right),\omega \left({\theta,{y_2}} \right)} \right)p\left({\theta,{y_1}} \right)d\theta \le M\rho \left({{y_1},{y_2}} \right)} \hfill & {{\rm{for}}\,\,{y_1},{y_2} \in Y} \hfill \cr}.} - (A4)
There exists S > 0 such that
\matrix{{\left| {\lambda \left({{y_1}} \right) - \lambda \left({{y_2}} \right)} \right| \le S\rho ({y_1},{y_2})} \hfill & {{\rm{for}}\,{y_1},{y_2} \in Y} \hfill \cr}. - (A5)
There exists T > 0 and W > 0 such that
\matrix{{\sum\limits_{J \in I} {\left| {{\pi_{ij}}\left({{y_1}} \right) - {\pi_{ij}}\left({{y_2}} \right)} \right| \le T\rho \left({{y_1},{y_2}} \right)}} \hfill & {{\rm{for}}\,\,{y_1},{y_2} \in Y,\,\,i \in I,} \hfill \cr {\int_\Theta {\left| {p\left({\theta,{y_1}} \right) - p\left({\theta,{y_2}} \right)} \right|d\theta \le W\rho \left({{y_1},{y_2}} \right)}} \hfill & {{\rm{for}}\,\,{y_1},{y_2} \in Y.} \hfill \cr} - (A6)
There exists ɛπ > 0 and ɛp > 0 such that
where Θ(y1, y2) = {θ ∈ Θ : ρ(ω(θ, y1), ω(θ, y2)) ≤ ρ(y1, y2)}.\matrix{{\sum\limits_{J \in I} {\min \{{\pi_{{i_1}j}}\left({{y_1}} \right),{\pi_{{i_2}j}}\left({{y_2}} \right)\} \ge {\varepsilon_\pi}}} \hfill & {{\rm{for}}\,\,{y_1},{y_2} \in Y,\,\,{i_1},{i_2} \in I,} \hfill \cr {\int_{\Theta ({y_1},{y_2})} {\min \{p\left({\theta,{y_1}} \right),p\left({\theta,{y_2}} \right)\} d\theta \ge {\varepsilon_p}}} \hfill & {{\rm{for}}\,\,{y_1},{y_2} \in Y,} \hfill \cr}
As it was written at the beginning of this section, the constant c appearing in (3) is needed to be sufficiently large and depends on constants appearing in conditions (A1)–(A4). For more details, we refer to [6].
In [6] it is shown, that if the conditions (A1)–(A6) hold and the constants occurring in them satisfy the inequality
We would like to justify the conditions (A1)–(A6) stated above. The condition (A2) is met by a quite large class of semiflows defined on reflexive Banach spaces which can be generated by some differential equations engaging dissipative operators [10]. It happens often [5] then that the condition (A1) is a consequence of the conditions (A2) and (A3). The conditions (A3)–(A6) concerning assumptions such as contractivity are quite natural and in our setting commonly used regarding ergodic properties. One may consult [13, 15] for details.
In this section we prove the CLT for model {(Yn, ξn)}n∈ℕ0 described in the previous section. Our proof is strongly based on the proof of Theorem 4.1 in [4]. The proof will require extensions of the conditions (A1) and (A3) of the model, namely:
- (A1)′
There exists y͂ > 0 such that
\matrix{{\mathop {\sup}\limits_{y \in Y} \mathop \int \nolimits_0^\infty {e^{\underline \lambda t}}\int_\Theta {{{[\rho (\omega (\theta,{\Pi_i}(t,\tilde y)),\tilde y)]}^2}p\left({\theta,{\Pi_i}\left({t,y} \right)} \right)d\theta dt < \infty}} \hfill & {{\rm{for}}\,i\, \in \,I.} \hfill \cr} - (A3)′
There exists M′ > 0 such that
\matrix{{\int_\Theta {{{[\rho (\omega (\theta,{y_1}),\,\omega (\theta,{y_2}))]}^2}p\left({\theta,{y_1}} \right)d\theta \le M'{{[\rho ({y_1},{y_2})]}^2}}} \hfill & {{\rm{for}}\,{y_1},{y_2}\, \in \,Y.} \hfill \cr}
Let us also define a function V : X → ℝ+ by
Let {(Yn, ξn)}n∈ℕ0 be the Markov chain with transition law corresponding to the Markov operator given by (7) and let (A1)′, (A2), (A3)′, (A4)–(A6) hold with constants satisfying
Let g ∈ Lip(X). If the initial distribution μ of the chain {g(Yn, ξn)}n∈ℕ0 belongs to
Before we proceed to justifying that the conditions (B0)–(B5) of Theorem 1 are met, we check that the inequality (10) implies (8) with
From (10) we see that
Given that conditions (A1)′ and (A3)′ imply (A1) and (A3), respectively, studying the proof of [6, Theorem 3.1], we conclude that the assumptions (A1)′, (A2), (A3)′, (A4)–(A6) guarantee that all conditions (B0)–(B5) beyond (B1) are fulfilled. It remains to show that (B1) is also met.
We start from describing the left-hand side of this condition.
We obtain that
Obviously I2 is finite because of the condition (A1)′. Finally, setting the constant to be