getdoca0bb. 265KB Jun 04 2011 12:04:46 AM

(1)

El e c t ro n ic

Jo ur

n a l o

f P

r o

b a b il i t y Vol. 15 (2010), Paper no. , pages 1743–1771.

Journal URL

http://www.math.washington.edu/~ejpecp/

A class of

F

-doubly stochastic Markov chains

Jacek Jakubowski

Institute of Mathematics, University of Warsaw Banacha 2, 02-097 Warszawa, Poland

and

Faculty of Mathematics and Information Science Warsaw University of Technology Plac Politechniki 1, 00-661 Warszawa, Poland

E-mail:jakub@mimuw.edu.pl

Mariusz Niew˛

egłowski

Faculty of Mathematics and Information Science, Warsaw University of Technology Plac Politechniki 1, 00-661 Warszawa, Poland

E-mail:M.Nieweglowski@mini.pw.edu.pl

Abstract

We define a new class of processes, very useful in applications, F-doubly stochastic Markov

chains which contains among others Markov chains. This class is fully characterized by some martingale properties, and one of them is new even in the case of Markov chains. Moreover a predictable representation theorem holds and doubly stochastic property is preserved under natural change of measure.

Key words: F-doubly stochastic Markov chain, intensity, Kolmogorov equations, martingale

characterization, sojourn time, predictable representation theorem.

AMS 2000 Subject Classification:Primary 60G99; Secondary: 60G55, 60G44, 60G17, 60K99. Submitted to EJP on March 29, 2010, final version accepted September 28, 2010.

Research supported in part by Polish KBN Grant P03A 034 29 “Stochastic evolution equations driven by Lévy noise”


(2)

1

Introduction

Our goal is to find a class of processes with good properties which can be used for modeling

differ-ent kind of phenomena. So, in Section 2 we introduce a class of processes, which we callF-doubly

stochastic Markov chains. The reason for the name is that there are two sources of uncertainty in their definition, so in analogy to Cox processes, called doubly stochastic Poisson processes, we chose

the name “F-doubly stochastic Markov chains”. This class contains Markov chains, compound

Pois-son processes with jumps inZ, Cox processes, the process of rating migration constructed by Lando

[16]and also the process of rating migration obtained by the canonical construction in Bielecki and

Rutkowski [3]. We stress that the class of F-doubly stochastic Markov chains contains processes

that are not Markov. In the following we use the shorthand “F–DS Markov chain” for the “F-doubly

stochastic Markov chain”. Note that anF-doubly stochastic Markov chain is a different object than a

doubly stochastic Markov chain which is a Markov chain with a doubly stochastic transition matrix.

On the end of this section we give examples of F-doubly stochastic Markov chains. Section 3 is

devoted to investigation of basic properties ofF-doubly stochastic Markov chains. In the first part

we prove that anF–DS Markov chainC is a conditional Markov chain and that anyF-martingale is

aFFC-martingale. This means that the immersion property for(F,FFC), so called hypothesis

H, holds. Moreover, the family of transition matrices satisfies the Chapman-Kolmogorov equations.

In the second part and until the end of the paper we restricted ourselves to a class ofF–DS Markov

chains with values in a finite setK ={1, . . . ,K}. We introduce the notion of intensity of anF–DS

Markov chain and formulate conditions which ensure its existence. In section 4 we prove that an

F–DS Markov chain C with intensity is completely characterized by the martingale property of the

compensated process describing the position ofC as well as by the martingale property of the

com-pensated processes counting the number of jumps of C from one state to another (Theorem 4.1).

The equivalence between points iii) and iv) in Theorem 4.1 in a context of Markov chains has not yet been known according to the best of our knowledge. In a view of the above characterizations,

theF–DS Markov chains can be described as the class of processes that behave like time

inhomoge-neous Markov chains conditioned onF. Moreover these equivalences and the fact that the class of

F-DS Markov chains contains the most ofFconditionalF∨FC Markov chains used for modeling in

finance, indicate that the class ofF-DS Markov chains is a natural, and very useful in applications,

subspace ofFconditionalFFC Markov chains. Next, anF–DS Markov chain with a given

inten-sity is constructed. Section 5 is devoted to investigation of properties of distribution ofC and the

distribution of sojourn time in fixed state junder assumption thatC does not stay in jforever. We

find some conditional distributions which among others allows to find a conditional probability of transition from one state to another given that transition occurs at known time. In Section 6 a kind of predictable representation theorem is given. Such theorems are very important for applications,

for example in finance and backward stochastic differential equations (see Pardoux and Peng[18]

and El Karoui, Peng and Quenez[7]). By the way we prove thatF–DS Markov chain with intensity

and arbitraryFadapted process do not have jumps at the same time. Our results allows to describe

and investigate a credit risk for a single firm. In such case the predictable representation theorem

(Theorem 6.5) generalize the Kusuoka theorem[14]. In the last section we study how replacing the

probability measure by an equivalent one affects the properties of anF–DS Markov chain.

Summing up, the class ofF–DS Markov chains is a class with very good and desirable properties in


(3)

used for modeling a credit rating migration process and which contains processes usually taken for this purpose. This allows us to include rating migration in the process of valuation of defaultable claims and generalize the case where only two states are considered: default and non-default (for

such generalizations see Jakubowski and Niew˛egłowski[12]). These processes can also be applied

in other fields where system evolves in a way depending on random environment, e.g., in insurance.

2

Definition and examples

In this section we introduce and investigate a new class of processes, which will be calledF-doubly

stochastic Markov chains. This class contains among others Markov chains and Cox processes. We

assume that all processes are defined on a complete probability space (Ω,F,P). We also fix a

filtrationFsatisfying usual conditions, which plays the role of a reference filtration.

Definition 2.1. A càdlàg process C is called anF–doubly stochastic Markov chain with state space

K ⊂Z={. . . ,−1, 0, 1, 2, . . .}if there exists a family of stochastic matricesP(s,t) = (pi,j(s,t))i,j∈K for 0≤st such that

1) the matrixP(s,t)isFt–measurable, andP(s,·)isFprogressively measurable,

2) for anyts≥0 and everyi,j∈ K we have

P(Ct= j| F∨ FsC)✶{C

s=i}=✶{Cs=i}pi,j(s,t). (2.1)

The processPwill be called the conditional transition probability process ofC.

The equality (2.1) implies that P(t,t) = I a.s. for all t 0. Definition 2.1 extends the notion

of Markov chain with continuous time (when F∞ is trivial). A process satisfying 1) and 2) is

called a doubly stochastic Markov chain by analogy with Cox processes (doubly stochastic Poisson processes). In both cases there are two sources of uncertainty. As mentioned in the Introduction, we

use the shorthand “F–DS Markov chain” for the “F–doubly stochastic Markov chain”. Now, we give

a few examples of processes which areF–DS Markov chains.

Example 1. (Compound Poisson process) LetX be a compound Poisson process with jumps inZ,

i.e., Xt = PNt

i=1Yi, where N is a Poisson process with intensity λ, Yi is a sequence of independent

identically distributed random variables with values inZand distribution ν, moreover(Yi)i≥1 and

N are independent. By straightforward calculations we see that:

a)X is anF–DS Markov chain withF=FN and

pi,j(s,t) =ν⊗(NtNs)(ji).

b)X is anF–DS Markov chain with respect to Fbeing the trivial filtration, and with deterministic

transition matrix given by the formula

pi,j(s,t) =

X

k=0

νk(ji)[λ(ts)]

k

k! e

−λ(ts).

From these examples we have seen that the conditional transition probability matrix depends on the choice of the reference filtration F, andP(s,t) can be either continuous with respect tos,t or discontinuous.


(4)

Example 2. (Cox process) The processC with càdlàg trajectories such that

P(CtCs=k| F∞∨ FsC) =e

−Rt

sλudu Rt

s λudu

k

k! (2.2)

for someF–adapted processλsuch thatλ≥0,Rt

0λsds<∞for allt ≥0 we call a Cox process. This

definition implies that

P(CtCs=k| F∞∨ FsC) =P(CtCs=k| F∞),

so the increments and the past (i.e. FC

s ) are conditionally independent givenF∞. Therefore for

ji,

P(Ct = j| F∞∨ FsC)✶{Cs=i}=✶{Cs=i}P(CtCs= ji| F∞∨ F

C s )

=✶{C

s=i}e

−Rt

s λudu Rt

s λudu

ji

(ji)! . Thus

pi,j(s,t) =

  

Rt sλudu

ji

(ji)! e

−Rstλudu for ji,

0 for j<i,

satisfy conditions 1) and 2) of Definition 2.1. A Cox processC is therefore an F–DS Markov chain

with K = N. Usually in a definition of Cox process there is one more assumption on intensity,

namelyR∞

0 λsds=∞a.s. Under this assumptionC has properties similar to Poisson process and is

called conditional Poisson process (or doubly stochastic Poisson process). So our definition of Cox process is a slight generalization of a classical one.

Example 3. (Time changed discrete Markov chain) Assume that ¯C is a discrete time Markov chain with values inK ={1, . . . ,K}, N is a Cox process and the processes(C¯k)k0 and(Nt)t0 are

inde-pendent and conditionally indeinde-pendent givenF. Then the processCt := C¯N

t is anF–DS Markov

chain (see Jakubowski and Niew˛egłowski[11, Theorem 7 and 9]).

Example 4. (Truncated Cox process) Simple calculations give us that the processCt:=minNt,K ,

whereN is a Cox process andK∈N, is anF–DS Markov chain with state spaceK ={0, . . . ,K}.

3

Basic properties

3.1

General case

In this subsection we consider the case of an arbitrary countable state space K. We study basic

properties of transition matrices and martingale invariance property of F with respect to F∨FC.

Moreover we prove that the class of F-DS Markov chains is a subclass of F conditional FFC

Markov chains.

For the rest of the paper we assume thatC0=i0for somei0∈ K. We start the investigation ofF–DS

Markov chains from the very useful lemma describing conditional finite–dimensional distributions


(5)

Lemma 3.1. If C is anF–DS Markov chain, then

P(Cu1=i1, . . .Cun=in| F∞∨ FuC0)✶

¦

Cu0=i0

© (3.1)

=✶¦C

u0=i0

©p

i0,i1(u0,u1)

n−1

Y

k=1

pi

k,ik+1(uk,uk+1) for arbitrary0≤u0≤. . .≤unand(i0, . . . ,in)∈ Kn+1.

Proof. The proof is by induction onn. For n=1 the above formula obviously holds. Assume that it holds forn, arbitrary 0≤u0 ≤. . . ≤un and(i0, . . . ,in)∈ Kn+1. We will prove it for n+1 and arbitrary 0≤u0≤. . .≤unun+1,(i0, . . . ,in,in+1)∈ Kn+2. Because

E(✶n

Cu1=i1,...Cun+1=in+1

o| F

∞∨ FuC0)✶

¦

Cu0=i0

© =E E ✶n

Cu2=i2,...,Cun+1=in+1

o|F

∞∨FuC1

✶¦

Cu1=i1

©|F

∞∨FuC0

✶¦

Cu0=i0

©,

by the induction assumption applied tou1≤. . .≤un+1 and(i1, . . . ,in+1)∈ Kn+1we know that the

left hand side of (3.1) is equal to

E ✶¦C u1=i1

©p

i1,i2(u1,u2)

n

Y

k=2

pik,ik+1(uk,uk+1)| F∞∨ FuC0

!

✶¦

Cu0=i0

©=I.

UsingF∞–measurability of family of transition probabilities(P(s,t))0≤st<∞, and the definition of

F–DS Markov chain, we obtain

I=E 

✶¦

Cu1=i1

©|F

∞∨ FuC0

‹

✶¦

Cu0=i0

© p

i1,i2(u1,u2)

n

Y

k=2

pi

k,ik+1(uk,uk+1)

!

=✶¦

Cu0=i0

©p

i0,i1(u0,u1) pi1,i2(u1,u2)

n

Y

k=2

pik,ik+1(uk,uk+1)

!

=✶¦

Cu0=i0

©p

i0,i1(u0,u1)

n

Y

k=1

pik,ik+1(uk,uk+1)

and this completes the proof.

Remark 3.2. Of course, if (3.1) holds, then condition 2) of Definition 2.1 ofF–DS Markov chain is

satisfied. Therefore(3.1)can be viewed as an alternative to equality(2.1).

As a consequence of our assumption thatC0=i0 and Lemma 3.1 we obtain

Proposition 3.3. If C is an F–DS Markov chain, then for arbitrary 0u1 ≤ . . . ≤ unt and

(i1, . . . ,in)∈ Kn we have

P(Cu1=i1, . . .Cu

n=in| F∞) =pi0,i1(0,u1)

n−1

Y

k=0

pi

k,ik+1(uk,uk+1), (3.2) and


(6)

The following hypothesis is standard for many applications e.g., in credit risk theory.

HYPOTHESIS H: For every bounded F–measurable random variable Y and for each t ≥ 0 we have

E(Y | Ft∨ FtC) =E(Y | Ft).

It is well known that hypothesis H for the filtrations F and FC is equivalent to the martingale

invariance property of the filtrationFwith respect toF∨FC (see Brémaud and Yor[4]or Bielecki

and Rutkowski[3, Lemma 6.1.1, page 167]) i.e. anyFmartingale is anF∨FC martingale. From

results in[4]one can deduce that hypothesis H implies (3.3). So, by Proposition 3.3, we can expect

that for a class ofF-DS Markov chains hypothesis H is satisfied. Indeed,

Proposition 3.4. If C is anF–DS Markov chain then hypothesis H holds.

Proof. According to[11, Lemma 2] we know that hypothesis H is equivalent to (3.3). This and Proposition 3.3 complete the proof.

Now, we will show that eachF–DS Markov chain is a conditional Markov chain (see[3, page 340]

for a precise definition). For an example of a process which is anFconditionalFFC Markov chain

and is not anF–DS Markov chain we refer to Section 3 of Becherer and Schweizer[2].

Proposition 3.5. Assume that C is anF–DS Markov chain. Then C is anFconditionalFFCMarkov

chain.

Proof. We have to check that forst,

P(Ct=i| Fs∨ FsC) =P(Ct=i| Fsσ(Cs)).

By the definition of anF–DS Markov chain,

P(Ct=i| Fs∨ FsC) =E(E(✶{Ct=i} | F∞∨ F

C

s )| Fs∨ FsC)

=E X

j∈K

{Cs=j}pj,i(s,t)|Fs∨ F

C s

= X

j∈K

{Cs=j}E €

pj,i(s,t)|Fs∨ FsCŠ=I. But

I= X

j∈K

{Cs=j}E €

pj,i(s,t)| FsŠ since hypothesis H holds (Proposition 3.4), and this ends the proof.

Now we define processesHi, which play a crucial role in our characterization of the class ofF–DS

Markov chains:

Hti:=✶{C

t=i} (3.4)

fori∈ K. The process Hti tells us whether at time t the processC is in statei or not. Let Ht := (t)⊤α∈K, where⊤denotes transposition.

We can express condition (2.1) in the definition of anF–DS Markov chain, fortu, in the form


(7)

or equivalently

E(Huj | F∨ FtC) = X

i∈K

Htipi,j(t,u)

and so (2.1) is equivalent to

E€Hu| F∞∨ FtC

Š

=P(t,u)⊤Ht. (3.5)

The next theorem states that the family of matricesP(s,t) = [pi,j(s,t)]i,j∈K satisfies the Chapman-Kolmogorov equations.

Theorem 3.6. Let C be anF–DS Markov chain with transition matrices P(s,t). Then for any uts we have

P(s,u) =P(s,t)P(t,u) a.s., (3.6)

so on the setCs=i we have

pi,j(s,u) = X

k∈K

pi,k(s,t)pk,j(t,u).

Proof. It is enough to prove that (3.6) holds on each setCs=i ,i∈ K. So we have to prove that

HsP(s,u) =HsP(s,t)P(t,u).

By the chain rule for conditional expectation, equality (3.5) and the fact that P(t,u) is F

measurable it follows that forstu,

P(s,u)⊤Hs =E

€

Hu| F∞∨ FsC

Š

=E€E€Hu| F∞∨ FtC

Š

| F∨ FC s

Š

=E€P(t,u)⊤Ht| F∞∨ FsC

Š

=P(t,u)⊤E€Ht | F∞∨ FsC

Š

=P(t,u)⊤P(s,t)⊤Hs= (P(s,t)P(t,u))⊤Hs,

and this completes the proof.

3.2

The case of a finite state space

From this subsection, until the end of the paper we restrict ourselves to a finite setK, i.e. K =

{1, . . . ,K}, withK<∞. It is enough for the most of applications, e.g., in finance to model markets with rating migrations we use processes with values in a finite set. In this caseHt:= (H1t . . . ,HtK)⊤. We recall the standing assumption thatC0=i0for somei0∈ K.

The crucial concept in this subsection and in study of properties of F–DS Markov chains is the

concept of intensity, analogous to that for continuous time Markov chains.

Definition 3.7. We say that an F–DS Markov chainC has an intensity if there exists anF–adapted

matrix-valued processΛ = (Λ(s))s≥0= (λi,j(s))s≥0 such that:

1) Λis locally integrable, i.e. for anyT >0

Z

]]0,T]]

X

i∈K


(8)

2) Λsatisfies the conditions:

λi,j(s)≥0 ∀i,j∈ K,i6= j, λi,i(s) =−X

j6=i

λi,j(s) ∀i∈ K, (3.8)

the Kolmogorov backward equation: for allvt,

P(v,t)−I=

Z t v

Λ(u)P(u,t)du, (3.9)

the Kolmogorov forward equation: for allvt,

P(v,t)−I= Z t

v

P(v,u)Λ(u)du. (3.10)

A processΛsatisfying the above conditions is called an intensity of theF–DS Markov chainC.

It is not obvious that if we have a solution to the Kolmogorov backward equation then it also solves the Kolmogorov forward equation. This fact follows from the theory of differential equa-tions, namely we have

Theorem 3.8. Assume thatΛis locally integrable. Then the random ODE’s

d X(t) =−Λ(t)X(t)d t, X(0) =I, (3.11)

d Y(t) =Y(t)Λ(t)d t, Y(0) =I, (3.12)

have unique solutions, and in addition X(t) = Y−1(t). Moreover, Z(s,t) := X(s)Y(t) is a unique solution to the Kolmogorov forward equation(3.10)and to the Kolmogorov backward equation(3.9). Proof. The existence and uniqueness of solutions of the ODE’s (3.11) and (3.12) follows by standard arguments. To deduce thatX(t) =Y−1(t)we apply integration by parts to the productX(t)Y(t)of finite variation continuous processes and get

d(Y(t)X(t)) = Y(t)d X(t) + (d Y(t))X(t)

= Y(t) (−Λ(t)X(t)d t) +Y(t)Λ(t)X(t)d t=0.

From Y(0) =X(0) =I we haveY(t)X(t) =I, which means thatX(t) is a right inverse matrix of

Y(t). It is also the left inverse, since we are dealing with square matrices.

Now we check that Z(s,t) are solutions to the Kolmogorov backward equation and also the

Kol-mogorov forward equation. Indeed,

dsZ(s,t) = (d X(s))Y(t) =−Λ(s)X(s)Y(t)ds=−Λ(s)Z(s,t)ds,

and

dtZ(s,t) =X(s)d Y(t) =X(s)Y(t)Λ(t)d t=Z(s,t)Λ(t)d t. This ends the proof sinceX(t) =Y−1(t)implies thatZ(t,t) =Ifor everyt≥0.


(9)

Corollary 3.9. If anF–DS–Markov chain C has intensity, then the conditional transition probability

process P(s,t)is jointly continuous at(s,t)for st. Proof. This follows immediately from Theorem 3.8, since

P(s,t) =X(s)Y(t)

and both factors are continuous insandt, respectively.

Theorem 3.8 gives us existence and uniqueness of solutions to (3.9) and (3.10). Next proposition provides the form of these solutions.

Proposition 3.10. Let a matrix process(Λ(s))s0satisfies conditions(3.7)and(3.8). Then the solution to(3.9)is given by the formula

P(v,t) =I+

X

n=1

Z t v

Z t v1

. . .

Z t vn−1

Λ(v1). . .Λ(vn)d vn. . .d v1,

and the solution to(3.10)is given by P(v,t) =I+

X

n=1

Z t v

Z v1

v

. . .

Z vn−1

v

Λ(vn). . .Λ(v1)d vn. . .d v1.

Proof. It is a special case of Theorem 5 in Gill and Johansen [8], see also Rolski et al. [20, § 8.4.1].

Proposition 3.11. Let P = (P(s,t)), 0 ≤ st, be a family of stochastic matrices such that the matrix P(s,t)is Ft–measurable, and P(s,·) isF–progressively measurable. Let Λ = (Λ(s))s0 be an F–adapted matrix-valued locally integrable process such that the Kolmogorov backward equation(3.9)

and Kolmogorov forward equation(3.10)hold. Then

i) For each s∈[0,t]there exists an inverse matrix of P(s,t)denoted by Q(s,t).

ii) There exists a version of Q(·,t) such that the process Q(·,t) is a unique solution to the integral (backward) equation

dQ(s,t) =Q(s,t)Λ(s)ds, Q(t,t) =I. (3.13)

This unique solution is given by the following series: Q(s,t) =I+

X

k=1 (−1)k

Z t s

Z t u1

. . .

Z t uk−1

Λ(uk). . .Λ(u1)duk. . .du1. (3.14)

iii) There exists a version of Q(s,·) such that the process Q(s,·) is a unique solution to the integral (forward) equation

dQ(s,t) =−Λ(t)Q(s,t)d t, Q(s,s) =I. (3.15)

This unique solution is given by the following series: Q(s,t) =I+

X

k=1 (−1)k

Z t s

Z u1

s

. . .

Z uk−1

s


(10)

Proof. i) From Theorem 3.8 it follows that P(s,t) = X(s)Y(t), where X, Y are solutions to the

random ODE’s (3.11), (3.12) and moreoverY =X−1. Therefore the matrix P(s,t)is invertible and

its inverseQ(s,t)is given byQ(s,t) =X(t)Y(s).

ii) We differentiateQ(s,t)with respect to the first argument and obtain

dsQ(s,t) =X(t)d Y(s) =X(t)Y(s)Λ(s)ds=Q(s,t)Λ(s)ds.

MoreoverQ(t,t) =X(t)Y(t) =I. SoQ(·,t)solves (3.13). Uniqueness of solutions to (3.13) follows by standard arguments based on Gronwall’s lemma. Formula (3.14) is derived analogously to a similar formula forP(s,t)in § 8.4.1, page 348 of Rolski et al.[20].

iii) The proof of iii) is analogous to that of ii).

In the next theorem we prove that under some conditions imposed on the conditional transition

probability process P, an F–DS Markov chain C has intensity. Using this intensities we can

con-struct martingale intensities for different counting processes building in natural way from F–DS

Markov chains (see Theorem 4.1). Therefore, Theorem 3.12 is in a spirit of approaches of Del-lacherie (Meyer’s Laplacian see DeDel-lacherie[6], Guo and Zeng[9]) and of Aven[1]. Theorem 3.12 generalizes forF–DS Markov chains results from[1].

Theorem 3.12(Existence of Intensity). Let C be anF–DS–Markov chain with conditional transition

probability process P. Assume that

1) P as a matrix–valued mapping is measurable, i.e.

P:(R+×R+×Ω,B(R+×R+)⊗ F)→(RK×K,B(RK×K)).

2) There exists a version of P which is continuous in s and in t. 3) For every t≥0the following limit exists almost surely

Λ(t):=lim

h↓0

P(t,t+h)−I

h , (3.17)

and is locally integrable. ThenΛis the intensity of C.

Proof. By assumption 3 the processΛis well defined and by 1) it is(R+×Ω,B(R+)⊗F)measurable.

By assumption 3, Λ(t) is Ft+–measurable, but F satisfies the usual conditions, so Λ(t) is Ft

measurable. It is easy to see that (3.8) holds.

It remains to prove that equations (3.9) and (3.10) are satisfied. Fixt. From the assumptions and

the Chapman–Kolmogorov equations it follows that forvv+ht,

P(v+h,t)−P(v,t) = P(v+h,t)−P(v,v+h)P(v+h,t) = −(P(v,v+h)−I)P(v+h,t),

so

P(v+h,t)−P(v,t)

h =−

(P(v,v+h)−I)


(11)

Therefore +vP(v,t) exists for a.e. v and is(R+×R+×,B(R+×R+)⊗ F) measurable. Using

assumptions 2 and 3 we finally have

+

∂vP(v,t) =−Λ(v)P(v,t), P(t,t) =I. (3.18)

Since elements of P(u,t) are bounded by 1, andΛis integrable over [v,t](by assumption 3), we see that +

uP(u,t)is Lebesgue integrable on[v,t], so (see Walker[21])

I−P(v,t) =

Z t v

+

∂uP(u,t)du.

Hence, by (3.18), we have

P(v,t)−I= Z t

v

Λ(u)P(u,t)du, and this is exactly the Kolmogorov backward equation (3.9).

Similar arguments apply to the case of right derivatives ofP(v,t)with respect to the second variable. Since forvtt+h,

P(v,t+h)−P(v,t) =P(v,t)(P(t,t+h)−I),

we obtain

+

∂tP(v,t) =P(v,t)Λ(t), P(v,v) =I,

which gives (3.10),

P(v,t)−I= Z t

v

P(v,u)Λ(u)du.

Now, we find the intensity for the processes described in Examples 3 and 4.

Example 5. IfCt =min

Nt,K , whereN is a Cox process with càdlàg intensity process ˜λ, thenC

has the intensity process of the form

λi,j(t) =

  

λ˜(t) fori= j∈ {0, . . .K−1}; ˜

λ(t) for j=i+1 withi∈ {0, . . .K−1};

0 otherwise.

Example 6. If anF–DS Markov chain C is defined as in Example 3 with a discrete time Markov

chain ¯C with a transition matrix Pand a Cox processN with càdlàg intensity process ˜λ, then

P(s,t) =e(PI)

Rt ˜(u)du

(see Theorem 9 in[11]), so the intensity of C is given by


(12)

4

Martingale properties

We prove that under natural assumptions, belonging of a process X to the class of F-DS Markov

chains is fully characterized by the martingale property of some processes strictly connected withX.

These nice martingale characterizations allow to check by using martingales whenever process is an

F-DS Markov chain. To do this, we introduce a filtrationGb= (Gbt)t0, where b

Gt:=F∨ FtC. (4.1)

Theorem 4.1. Let(Ct)t0 be aK–valued stochastic process and(Λ(t))t0be a matrix valued process satisfying(3.7)and(3.8). The following conditions are equivalent:

i) The process C is anF–DS Markov chain with intensity processΛ.

ii) The processes

Mti:=Hti

Z

]]0,t]]

λCu,i(u)du, i∈ K, (4.2)

areGb local martingales.

iii) The processes Mi,j defined by Mti,j:=Hit,j

Z

]]0,t]]

Hsi,j(s)ds, i,j∈ K, i6= j, (4.3)

where

Hti,j:=

Z

]]0,t]]

Huid Huj, (4.4)

areGb local martingales.

iv) The process L defined by

Lt:=Q(0,t)⊤Ht, (4.5)

where Q(0,t)is a unique solution to the random integral equation

dQ(0,t) =−Λ(t)Q(0,t)d t, Q(0, 0) =I, (4.6)

is aGb local martingale.

Proof. Denoting byM the vector valued process with coordinatesMi, we can writeM as follows

Mt:=Ht

Z

]]0,t]]

Λ⊤(u)Hudu. (4.7)

"i)⇒ii)" Assume that C is anF–DS Markov chain with intensity process (Λ(t))t0. Fix t 0 and

set

Ns:=P(s,t)⊤Hs for 0≤st. (4.8)

The processC satisfies (3.5), which is equivalent to N being a Gb martingale for 0st. Using

integration by parts and the Kolmogorov backward equation (3.9) we find that

d Ns= (d P(s,t))⊤Hs+P⊤(s,t)d Hs (4.9)


(13)

Hence, using Q(s,t) (the inverse of P(s,t); we know that it exists, see Proposition 3.11iii)), we conclude that

MsM0=

Z

]]0,s]]

Q⊤(u,t)P⊤(u,t)d Mu=

Z

]]0,s]]

Q⊤(u,t)d Nu. Therefore, by theGb martingale property ofN, we see thatM is aGb local martingale.

"ii)⇒i)" Assume that the processM associated withC andΛis aGb martingale. Fixt0. To prove

thatCis anF–DS Markov chain it is enough to show that for some process(P(s,t))0stthe process

N defined by (4.8) is aGb martingale on [0,t]. Let P(s,t):= X(s)Y(t) withX, Y being solutions

to the random ODE’s (3.11) and (3.12). We know that P(·,t) satisfies the integral equation (3.9)

(see Theorem 3.8). We also know thatP(s,t) isFt–measurable (Remark 3.10) and continuous in

t, henceFprogressively measurable. Using the same arguments as before, we find that (4.9) holds.

So, using the martingale property of M we see that N is a local martingale. The definition of N

implies thatN is bounded (sinceH andPare bounded, see Last and Brandt[17, §7.4]). Therefore

N has an integrable supremum, so it is a Gb martingale, which implies that C is anF–DS Markov

chain with transition matrixP. From Theorem 3.8 it follows thatΛis the intensity matrix process

ofC.

"ii)⇔ iii)" and "iii)⇔ i v)" These equivalences follows from Lemmas 4.3 and 4.4 below, respec-tively, withA=Gb given by (4.1).

The proof is complete.

Remark 4.2. The equivalence between i) and ii) in the above proposition corresponds to a well known martingale characterization of a Markov chain with a finite state space. In a view of this character-ization of F–DS Markov chains with intensities, they can be described as the class of processes that

conditioned onF∞behave as time inhomogeneous Markov chains with intensities. Hence, we can also

see that the nameF–doubly stochastic Markov chain well describe this class of processes. The

equiva-lence between iii) and iv) in a context of Markov chains has not yet been known according to the best of our knowledge.

The equivalence between points ii), iii) and iv) in Theorem 4.1 is a simple consequence of slightly more general results, which we formulate in two separate lemmas. It is worth to note that equiva-lences in lemmas below follow from general stochastic integration theory and in the proofs we do not use the doubly stochastic property.

Lemma 4.3. LetAbe a filtration such thatΛand H are adapted toA. Assume thatΛsatisfies(3.7)

and(3.8). The processes Mi, defined by(4.2)for i ∈ K, areAlocal martingales if and only if for all

i,j∈ K, i6= j, the processes Mi,jdefined by(4.3)areAlocal martingales.

Proof. ⇒Fixi6= j, i,j∈ K. Using the definition ofMti,j andMti we have

Mti,j=

Z

]]0,t]]

Huid Huj

Z

]]0,t]]

Hui,j(u)du=

Z

]]0,t]]

Huid Huj

Z

]]0,t]]

HuC

u,j(u)du

=

Z

]]0,t]]

Huid Huj

Z

]]0,t]]

HuiλCu,j(u)du=

Z

]]0,t]]


(14)

HenceMi,jis anAlocal martingale, sinceMj is one andHi

u−is a bounded process.

⇐Assume that theMti,j areAlocal martingales for alli6= j, i,j∈ K. First notice that Hi can be obtained fromHj,i by the formula

Hti=H0i +X

j6=i

Htj,iHti,j

.

Here and in what follows we use the notationPj6=i =Pj∈K \{i}. Indeed, from (4.4) it follows that

X

j6=i

(Htj,iHit,j) =

Z

]]0,t]]

  

X

j6=i

Huj

  d Hui +

Z

]]0,t]]

Huid

  −X

j6=i

Huj

  

=

Z

]]0,t]]

(1−Hui)d Hui +

Z

]]0,t]]

Huid Hui =HtiH0i.

Next, by (4.3),

Hti=H0i +X

j6=i

(Mtj,iMti,j) +X

j6=i

Z

]]0,t]]

Hsj,i(s)−Hsi,j(s)ds

!

=H0i +X

j6=i

Mtj,iMti,j

+

Z

]]0,t]]

K

X

j=1

Hsj,i(s)ds

=H0i +X

j6=i

Mtj,iMti,j

+

Z

]]0,t]]

λCs,i(s)ds,

which implies that

Mti=Hit

Z

]]0,t]]

λCs,i(s)ds=H0i+X

j6=i

Mtj,iMti,j

,

and thereforeMi is anAlocal martingale for eachi∈ K as a finite sum ofAlocal martingales.

The processHi,jdefined by (4.4) counts the number of jumps from stateito jover the time interval

]]0,t]]. Indeed, it is easy to see that

Hit,j= X 0<ut

HuiHuj.

Lemma 4.4. LetAbe a filtration such thatΛand H are adapted toA. Assume thatΛsatisfies(3.7)

and (3.8). The processes Mi, i ∈ K, are Alocal martingales if and only if the process L defined by

(4.5)is anAlocal martingale.

Proof. Integration by parts formula gives


(15)

indeed

d Lt=Q(0,t)⊤d Ht+d(Q(0,t)⊤)Ht =Q(0,t)⊤d HtQ(0,t)⊤Λ(t)⊤Htd t=Q(0,t)⊤d Mt.

We know thatQ isApredictable and locally bounded. So, if M is anAlocal martingale, then L is

also anAlocal martingale.

On the other hands, takingP(0,t)as an unique solution to the integral equation

d P(0,t) =P(0,t)Λ(t)d t, P(0, 0) =I

and noting thatP(0,t)is the inverse ofQ(0,t)(see Proposition 3.11) we have

P(0,t)⊤d Lt =P(0,t)⊤Q(0,t)⊤d Mt=d Mt.

Therefore, if Lis anAlocal martingale, thenM is also anAlocal martingale.

Corollary 4.5. If C is anF–DS Markov chain, then Mi areGlocal martingales withGt=Ft∨ FC t .

Proof. This follows from the fact that theMi are adapted toG, andGis a subfiltration ofGb.

Remark 4.6. The process C obtained by the canonical construction in[3]is an F–DS Markov chain.

This is a consequence of Theorem 4.1, becauseΛin the canonical construction is bounded and calcula-tions analogous to those in[3, Lemma 11.3.2 page 347]show that Mi areGb martingales. In a similar

way one can check that if C is a process of rating migration given by Lando[16], then Mi areGb local

martingales, so C is anF–DS Markov chain.

The martingaleM is orthogonal to all square integrableFmartingales.

Proposition 4.7. Let C be an F–DS Markov chain. The martingale M given by (4.7) is strongly

orthogonal to all square integrableFmartingales.

Proof. Denote by N an arbitrary Rd–valued, square integrableF martingale. Since C is an F–DS

Markov chain we have, by Proposition 3.5, thatH hypothesis holds and thereforeN is alsoF∨FC

martingale. SinceM is anF∨FC martingale, we need to show that the processNjMi is anF∨FC

martingale for everyi∈ K and j∈ {1, . . . ,d}. Fix arbitraryt≥0 and anyst, then

E NtjMtiFs∨ FsC=ENtjE€MtiF∨ FsCŠ Fs∨ FsC

=E

NtjMsiFs∨ FsC=MsiE

NtjFs∨ FsC=NsjMsi

and hence the result follows.

Now we construct an F–DS Markov chain with intensity given by an arbitraryF adapted

matrix-valued locally bounded stochastic process which satisfies condition (3.8).

Theorem 4.8. Let(Λ(t))t0be an arbitraryFadapted matrix-valued stochastic process which satisfies


(16)

Proof. We assume that on a probability space(Ω,F,P)with a filtrationFwe have a family of Cox

processesNi,j fori,j∈ K with intensities(λi,j(t))such that theNi,j are conditionally independent

givenF∞(otherwise we enlarge the probability space). We construct on(Ω,F,P)anF–DS Markov

chainC with intensity(Λ(t))t≥0 and given initial state i0. It is a pathwise construction inspired by

Lando[16]. First, we define a sequence(τn)n≥1of jump times of C and a sequence(C¯n)n≥0 which

describes the states of rating after change. We define these sequences by induction. We put ¯

C0=i0, τ1:= min

j∈K \C¯0

infnt>0 :∆NC¯0,j

t >0

o

Ifτ1=∞, thenC never jumps so ¯C1 :=i0. Ifτ1 <∞, then we put ¯C1:= j, where jis the element

ofK \C¯0 for which the above minimum is attained. By conditional independence of Ni,j given

F, the processesNi,jhave no common jumps, so ¯C1 is uniquely determined. We now assume that

τ1, . . . ,τk, ¯C1, . . . , ¯Ck are defined,τk<∞and we constructτk+1 as the first jump time of the Cox

processes afterτk, i.e.

τk+1:= min

j∈K \C¯k

inf

n

t> τk:∆N

¯

Ck,j

t >0

o

.

Ifτk+1 =∞then ¯Ck+1 =C¯k, and ifτk+1 <∞we put ¯Ck+1:= j, where j is the element ofK \C¯k

for which the above minimum is attained. Arguing as before, we see thatτk+1 and ¯Ck+1 are well

defined.

Having the sequences(τn)n≥0 and(C¯n)n≥0 we define a processC by the formula

Ct:=

X

k=0

✶[τk,τk+1)(t)C¯k. (4.11)

This processCis càdlàg and adapted to the filtrationA= (At)t0, whereAt:=FtW

i6=jF Ni,j

t

, and hence it is also adapted to the larger filtrationAe = (Aft)t0,Aft:=F∨Wi6=jFtNi,j. Notice thatHi,j defined by (4.4) is equal to

Hit,j=

Z

]]0,t]]

✶{i}(Cs−)d Nsi,j a.s.

The processesNti,j−R

]]0,t]]λi,j(s)dsareAelocal martingales (since they are compensated Cox’s

pro-cesses, see e.g.[3]). Likewise, eachMi,j defined by (4.3) is anAe local martingale, since

Mti,j=Hit,j

Z

]]0,t]]

Hsi,j(s)ds=Hti,j

Z

]]0,t]]

Hsiλi,j(s)ds

=

Z

]]0,t]]

✶{i}(Cs−)d(Nsi,jλi,j(s)ds).

Recall thatGbt=F∨ FC

t , soGb ⊆Ae. Therefore eachMi,jis also aGb local martingale, sinceMi,j is

b


(17)

Remark 4.9. Suppose that in the above constructionτk<a.s. and Cτ

k =i. Then

a)τk+1<a.s. provided

j∈ K \ {i}

Z ∞

0

λi,j(s)ds=∞ a.s.,

b)τk+1=∞a.s. provided

j∈ K \ {i}

Z ∞ τk

λi,j(s)ds=0 a.s.

5

Distribution of

C

and sojourn times

In this section we investigate properties of distribution ofC and the distribution of sojourn time in

a fixed state j under assumption thatC does not stay in j forever. These properties give, among

others, some interpretation to

λj(t):= X

l∈K \{j}

λj,l(t) =−λj,j(t) (5.1)

and to the ratios λλi,j

i . The first result says that the sojourn time of a given state has an exponential

distribution in an appropriate time scale.

Proposition 5.1. Let C be anF–DS Markov chain with intensityΛ,τ0=0, and

τk=inf¦t> τk1:Ct 6=Cτ

k−1,Ct−=Cτk−1

©

. (5.2)

Ifτk<a.s., then the random variable

Ek:=

Z τk

τk−1

λCτk1(u)du (5.3)

is independent of Gbτk1 and Ek is exponentially distributed with parameter equal to 1. Moreover, (Ei)1ikis a sequence of independent random variables.

Proof. For arbitrary j∈ K define the process that counts number of jumps from j Ntj:= X

l∈K \{j}

Htj,l.

Let

e

Ytj:=Ntj

Z t

0

Hujλj(u)du.

e

Yj is aGb local martingale by Theorem 4.1. Moreover, a sequence of stopping times defined by

σn:=inf

¨

t≥0 :Ntjn

Z t

0

Hujλj(u)dun


(18)

reducesYej. In what follows we denote

Ytn:=Yetj∧σ

n.

The definition ofσnimplies thatYnis a boundedGb martingale for everyn. Using the Itô lemma on

the interval]τk1,τk]we get

eiuYτnk =eiuY

n

τk1+

Z

]]τk−1,τk]]

iueiuYsnd Yn

s +

X

τk−1<s≤τk €

eiuYsneiuY n

s−−iueiuY n s−∆Yn

s

Š

=eiuYτnk1+

Z

]]τk−1,τk]]

iueiuYsnd Yn

s +

X

τk−1<s≤τk

eiuYsn−€eiu1iuŠNj

s.

For everynthe boundedness of(Yn)and(eiuYn)imply that the process

Z

]]0,t]]

iueiuYsnd Yn

s

!

t≥0

is a uniformly integrableGb martingale. Therefore, by the Doob optional sampling theorem we have

E   eiu

YτnkYτnk1

− X

τk−1<s≤τk

eiu

Ysn−−Yτnk1

€

eiu−1−iuŠ∆Nsj∧σ

n bGτk−1

 

=1. (5.4)

On the set¦Cτk−1= j

©

we have

Yτn

kY

n

τk−1=✶{τk≤σn} 1− Z τk

τk−1

λj(s)ds

!

−✶

k−1≤σn<τk} Z σn

τk−1

λj(s)ds

!

and

Yτn

k−−Y

n

τk−1=−

Z τk∧σn

τk−1∧σn

λj(u)du.

Hence using (5.4) and facts that∆Nτj

k∧σn =✶{τk≤σn} and∆N

j

s∧σn =0 forτk−1 <s< τk we infer

that on the set¦Cτ

k−1= j

©

:

E

k≤σn}e

iu



1−Rτk

τk−1λj(s)ds

‹

+✶

n<τk}e

iuRσn

τk1σnλj(s)ds

−✶{τ

k≤σn}e

iuRττk

k−1λj(s)ds€eiu−1−iuŠ bG

τk−1

= 1, and therefore

E

{τk≤σn}e

iuRτk

τk−1λj(s)ds(1+iu) +✶

{σn<τk}e

iuRσn

τk−1∧σnλj(s)ds bGτ

k−1

=1.

Applying Lebesgue’s dominated convergence theorem and the fact thatσn↑+∞yields

E

eiu

Rτk

τk1λj(u)du bGτ

k−1

✶n

k1=j

o= 1

1+iu.

Which implies the first statement of theorem. The second statement follows immediately from the above considerations.


(19)

As a consequence of the Proposition 5.1 we have the following result which states that the exit times are in some sense exponentially distributed. This is a generalization of the well known property of Markov chains.

Corollary 5.2. Let C be an F–DS Markov chain with intensity Λ and let τk <a.s., where τk is defined by(5.2). Then

P€τkτk−1>t|Gbτk−1

Š

=e

Rτk1+t

τk1 λCτk1(u)du. (5.5)

Moreover, on the set¦Cτk1=i©we have

Z ∞ τk−1

λi(u)du=∞. (5.6)

Proof. Because

τkτk−1=inf

(

t≥0 :

Z τk−1+t

τk−1

λCτ

k−1(u)duEk

)

,

we obtain (5.5). Indeed

P€τkτk1t|Gbτ

k−1

Š

=P

Z τk−1+t

τk−1

λC

τk1(u)duEk

bGτk−1

!

=e

Rτk1+t

τk1 λCτk1(u)du.

(5.5) yields (5.6) by letting t→ ∞.

The next proposition describes the conditional distribution of the vector(Cτ

k,τkτk−1)givenGbτk−1.

Proposition 5.3. Let C be an F-DS Markov chain with an intensity Λ and τk be given by (5.2). If

τk<a.s., then we have

P€Cτk = j,τkτk−1≤t

bGτk1Š=

Z t

0

e

Ru

0λi(v+τk−1)d vλ

i,j(τk−1+u)du (5.7)

on the set¦Cτ

k−1=i

©

.

Proof. First we note that RHS of (5.7) is well defined. Fix arbitrary j ∈ K. Let Nj be the process that counts number of jumps ofC to the state j, i.e.

Ntj:= X

i∈K \{j}

Hti,j,

and

Ytj:=Ntj

Z t

0

X

i∈K \{j}


(20)

Then Yj is a Gb local martingale by Theorem 4.1. Therefore there exist a sequence ofGb stopping

times σn ↑ +∞ such that (Ytj∧σ

n)t≥0 is an uniformly integrable martingale. By Doob’s optional

sampling theorem we have on the set¦Cτ

k−1=i

©

EN(jt+τ

k−1)∧τk∧σnN

j

τk−1∧σn bGτ

k−1

=E

Z (t+τk−1)∧τk∧σn

τk−1∧σn

λi,j(u)du

bGτk−1

!

. (5.8)

Since on the set¦Cτ

k−1=i

©

N(jt+τ

k−1)∧τk∧σnN

j

τk−1∧σn= ¨

1 ifXτk = j,τkτk−1≤t,τkσn,

0 otherwise ,

we can pass to the limit on the LHS of (5.8) and obtain lim

n→∞E

N(jt+τ

k−1)∧τk∧σnN

j

τk−1∧σn bGτ

k−1

=P(Xτk=j,τkτk−1≤t|Gbτk−1).

The RHS of (5.8) we can write as a sum

E

Z (t+τk−1)∧τk∧σn

τk−1∧σn

λi,j(u)du

bGτk−1

!

=I1(n) +I2(n), where

I1(n):=E{τ k≤σn}

Z (τk−τk−1)∧t

0

λi,j(s+τk1)ds

bGτk−1

!

,

I2(n):=E{τk−1≤σn<τk}

Z (σn−τk−1)∧t

0

λi,j(s+τk−1)ds

bGτk−1

!

.

By a monotone convergence theorem and fact thatσn↑+∞we obtain

lim

n→∞I1(n) =E

Z (τk−τk−1)∧t

0

λi,j(s+τk−1)ds

bGτk−1

!

. On the other hand

lim

n→∞I2(n) =0, a.s.

Summing up, we have proved the following equality holds on the set¦Cτ

k−1=i

©

P€Xτk = j,τkτk−1≤t bGτ

k−1

Š

=E

Z (τk−τk−1)∧t

0

λi,j(s+τk1)ds|Gbτ

k−1

!

. (5.9)

To finish the proof it is enough to transform the RHS of (5.9). Corollary 5.2 and Fubbini theorem yield

E

Z (τk−τk−1)∧t

0

λi,j(s+τk1)ds

bGτk−1


(1)

Therefore, we need to derive predictable representation for an arbitrary random variable inM. Fix

X ∈ M. Without loss of generality we can assume thatt1≤. . .≤tn. Let Zbe a martingale defined byZt:=E(XFtN∨ FtC), so it has the form

Zt=E Y n Y k=1

✶n Ctk=ik

oFN

t ∨ F

C t

!

.

Fixm∈ {1, . . . ,n}. ForKdimensional vectorseik with 1 onik-th component and zero otherwise, and

t∈(tm−1,tm]we have

Zt= m−1

Y k=1

✶n Ctk=ik

o !

E Y n Y k=m

✶n Ctk=ik

oFN

t ∨ F

C t

!

=

mY−1

k=1 ✶n

Ctk=ik

o !

X i∈K

HitE Y Pi,im(t,tm)

n Y k=m+1

Pik−1,ik(tk−1,tk)

FtN

!

=

mY−1

k=1 ✶n

Ctk=ik

o !

HtE P(t,tm)eimY

n Y k=m+1

Pik−1,ik(tk−1,tk)

FtN

!

=

mY−1

k=1 ✶n

Ctk=ik

o !

HtQ(0,t)E P(0,tm)ei

mY

n Y k=m+1

Pi

k−1,ik(tk−1,tk)

FtN

!

,

where we have used conditional Chapman-Kolmogorov equation, i.e., P(0,tm) = P(0,t)P(t,tm),

together with invertibility ofP(0,t). Define a bounded martingaleNmby

Ntm=E P(0,tm)eimY

n Y k=m+1

Pik−1,ik(tk−1,tk)

FtN

!

.

Then fort∈(tm−1,tm], usingLdefined by (4.5), we have

Zt=

mY−1

k=1 ✶n

Ctk=ik

o !

(Ntm)⊤Lt. (6.3)

By assumption,Nmadmits the following predictable decomposition:

NTm=N0m+

Z

]]0,T]]

(Rmt)⊤d Nt

for someF-predictable stochastic processRm. Hence, using (4.10) and Proposition 6.1, we have

d((Ntm)⊤Lt) = (Ntm)⊤d Lt+d(Ntm)⊤Lt−= (Ntm−) ⊤d L

t+ (Ltd N

m t )

= (Ntm)⊤Q(0,t)⊤d Mt+HtQ(0,t)d Ntm

= (Q(0,t)Ntm)⊤d Mt+ (HtQ(0,t))d N

m t


(2)

Therefore,

(Ntm

m)

L

tm=(N

m tm−1)

L

tm−1+ Z

]]tm−1,tm]]

(Q(0,u)Num)⊤d Mu (6.4)

+

Z

]]tm−1,tm]]

€

RmuQ(0,u)⊤HuŠ⊤d Nu.

Moreover,

(Ntm

m)

L

tm= (N

m tm)

Q(0,t

m)⊤Htm = (Q(0,tm)N

m tm)

H

tm

=ei

mHtmE Y

n Y k=m+1

Pik−

1,ik(tk−1,tk)

FtN

m

!

and

(Ntm+1

m )

L

tm = (N

m+1

tm )

Q(0,t

m)⊤Htm= (Q(0,tm)N

m+1

tm )

H

tm

= E Q(0,tm)P(0,tm+1)ei

m+1Y n Y k=m+2

Pik−

1,ik(tk−1,tk)

FtN

m

!!⊤

Ht

m

= E P(tm,tm+1)eim+1Y n Y k=m+2

Pik−1,ik(tk−1,tk)

FtN

m

!!⊤

Htm.

These imply that

Zt

m=

m Y k=1

✶n Ctk=ik

o !

(Ntm+1

m )

L

tm, (6.5)

By (6.4), (6.3), (6.5) form−1 and properties of stochastic integral we have

Ztm=Ztm−1+ Z

]]tm−1,tm]]

m−1

Y k=1

✶n Ctk=ik

o !

(Q(0,u)Num)⊤d Mu

+

Z

]]tm−1,tm]]

m−1

Y k=1

✶n Ctk=ik

o !

€

RmuQ(0,u)⊤Hu

Š⊤

d Nu.

Thus defining processesR,S by

Rt:= n X m=1

✶]]tm−1,tm]](t)

mY−1

k=1 ✶n

Ctk=ik

o !

RmuQ(0,u)⊤Hu

St:=

n X m=1

✶]]tm−1,tm]](t)

mY−1

k=1 ✶n

Ctk=ik

o !

Q(0,t)Ntm


(3)

7

Change of probability and doubly stochastic property

Now, we investigate how changing the probability measure to an equivalent one affects the proper-ties of anF–DS Markov chain. We start from a lemma

Lemma 7.1. LetQ,Pbe equivalent probability measures with density factorizing as dQ

dP

F∞∨FTC

:=η1η2, (7.1)

whereη1 is anF–measurable strictly positive random variable andη2 is anF∨ FTC–measurable

strictly positive random variable integrable underP. Let(η2(t))t[0,T]be defined by the formula

η2(t):=EP(η2| F∞∨ FtC), η2(0) =1. (7.2)

Then (N(t))t[0,T] is a Gb martingale (resp. local martingale) under Q if and only if (N(t)η2(t))t∈[0,T]is aGb martingale (resp. local martingale) underP.

Proof. ⇒ By the abstract Bayes rule and the fact that η1 is F∞ measurable and hence also Gu

measurable for allu≥0, we obtain, fors<t,

N(s) = EQ€N(t)|GbsŠ=EP

€

N(t)η1η2|GbsŠ EP

€

η1η2|Gbs

Š =EP N(t)EP

€

η1η2|GbtŠ EP

€

η1η2|Gbs Š

bGs

!

= EP N(t)EP

€

η2|GbtŠ EP€η2|GbsŠ

bGs

!

=EP

N(t)η2(t) η2(s)

bGs

=EP

€

N(t)η2(t)|GbsŠ η2(s) . ⇐The proof is similar.

The next lemma is a standard one, describing the change of compensator of a pure jump martingale under a change of probability measure. Although it can be proved using Girsanov-Meyer theorem (see e.g. [10],[15],[19]), we give a short self-contained proof.

Lemma 7.2. Let C be anF–DS Markov chain underPwith intensity(λi,j)and suppose thatη2defined

by(7.2)satisfies

2(t) =η2(t−)

 X

k,l∈K:k6=l

κk,l(u)d Muk,l

 (7.3)

with someGpredictable stochastic processesκi,j, i,j∈ K, such thatκi,j>−1. Then

e

Mti,j=Hti,j

Z

]]0,t]]

Hui(1+κi,j(u))λi,j(u)du, (7.4)


(4)

Proof. By Lemma 7.1 it is enough to prove thatMei,2is aGb local martingale underP. Integration

by parts yields

d(Meti,2(t)) =Meti,jdη2(t) +η2(t−)dMeti,j+ ∆Meti,jη2(t) =:I. Since

e

Mti,j=Mti,j

Z

]]0,t]]

Huiκi,j(u)λi,j(u)du,

we have

dMeti,j=d Mti,jHti,j(t)λi,j(t)d t

and

Meti,jη2(t) = ∆Mti,2(t−)

 X

k,l∈K:k6=l

κk,l(t)∆Mtk,l

=η2(t−)κi,j(t)

Mti,j 2

=η2(t−)κi,j(t)∆Hti,j 2

=η2(t−)κi,j(t)∆Hti,j. Hence

I = Meti,2(t−)

 X

k,l∈K:k6=l

κk,l(t)d Mtk,l

+η2(t−)d Mti,j+ η2(t−)κi,j(t)(∆Hti,jHti,j(t)d t)

= Meti,2(t−)

 X

k,l∈K:k6=l

κk,l(t)d Mtk,l

+η2(t−)(1+κi,j(t))d Mti,j,

which completes the proof.

Hence and from Theorem 4.1 we deduce that the doubly stochastic property is preserved by a wide class of equivalent changes of probability measures.

Theorem 7.3. Let C be anF–DS Markov chain underPwith intensity(λi,j), andQbe an equivalent probability measure with density given by(7.1)andη2 satisfying(7.3)with an Fpredictable

matrix-valued processκ. Then C is anF–DS Markov chain underQwith intensity((1+κi,j)λi,j).

Now, we exhibit a broad class of equivalent probability measures such that the factorization (7.1) in Lemma 7.1 holds.

Example 7. LetF=FW be the filtration generated by some Brownian motionW underP, and letC

be anF–DS Markov chain with intensity matrix processΛ. LetQbe a probability measure equivalent

toPwith Radon-Nikodym density process given as a solution to the SDE

dηt =ηt

γtdWt+ X k,l∈K:k6=l

κk,l(u)d Muk,l

, η0=1,

withFpredictable stochastic processesγandκ. It is easy to see that this density can be written as

a product of the following two Doleans-Dade exponentials:


(5)

and

2(t) =η2(t−)

 X

k,l∈K:k6=l

κk,l(u)d Muk,l

, η2(0) =1.

Therefore a factorization

η(t) =η1(t)η2(t)

as in Lemma 7.1 holds, sinceη1 isFmeasurable. As an immediate consequence we find that C

is anF–DS Markov chain underQwith intensity[λQ]i,j= ((1+κi,j)λi,j)and moreover the process defined byWt∗:=Wt

Rt

0γuduis a Brownian motion underQ.

AcknowledgmentsThe authors thank to anonymous referee for her/his bibliographical comments.

References

[1] T. Aven. A Theorem for Determining the Compensator of a Counting Process. Scand. J. Statist.

12(1985), No. 1, 69–72. MR0804226

[2] D. Becherer, M. Schweizer. Classical solutions to reaction-diffusion systems for hedging prob-lems with interacting Itô and point processes.Ann. Appl. Probab.15(2005), No. 2, 1111–1144. MR2134099

[3] T. Bielecki and M. Rutkowski. Credit Risk: Modeling, Valuation and Hedging. Springer Finance. Springer-Verlag Berlin Heidelberg, New York, 2001.

[4] P. Brémaud and M. Yor. Changes of Filtrations and of Probability Measures. Z. Wahrsch. Verw. Gebiete45(1978), no. 4, 269–295. MR0511775

[5] R. Cont and P. Tankov Financial modelling with jump processes . Chapman & Hall, 2004. MR2042661

[6] C. Dellacherie. Capacités et processus stochastiques. Springer, Berlin. (1972).

[7] N. El Karoui, S. Peng, and M.C. Quenez. Backward Stochastic Differential Equations in Finance. Math. Finance7(1997), No. 1, 1–71. MR1434407

[8] R.D. Gill and S. Johansen A survey of product-integration with a view toward application in survival analysis. Ann. Statist.18(1990), No. 4, 1501–1555. MR1074422

[9] X. Guo and Y. Zeng Intensity Process And Compensator: A New Filtration Expansion Approach And The Jeulin-Yor Theorem. Ann. Appl. Probab. 18(2008), No. 1, 120–142. MR2380894

[10] Jacod, Jean and Shiryaev, Albert N. Limit theorems for stochastic processes., Grundlehren der Mathematischen Wissenschaften, 288. Springer-Verlag, Berlin, 1987. MR0959133

[11] J. Jakubowski and M. Niew˛egłowski. Pricing bonds and CDS in the model with rating migra-tion induced by Cox process. Advances in Mathematics of Finance, ed. L. Stettner, Banach Centre Publication.83(2008), 159–182. MR2509232


(6)

[12] J. Jakubowski and M. Niew˛egłowski. Hedging strategies of payment streams and application to defaultable rating-sensitive claims. To appear in AMaMeF volume

[13] F.C. Klebaner. Introduction to stochastic calculus with applications. Second edition. Imperial College Press, London, 2005. MR2160228

[14] S. Kusuoka. A remark on default risk models. Adv. Math. Econ.1(1999), 69–82. MR1722700

[15] H. Kunita. Itô’s stochastic calculus: Its surprising power for applications Stoch. Proc. Appl.

120(2010), No. 5, 622–652. MR2603057

[16] D. Lando. On Cox processes and credit risky securities. Rev. Deriv. Res.2(1998), 99–120.

[17] G. Last and A. Brandt. Marked Point Processes on the Real Line: The Dynamic Approach. Springer-Verlag Berlin Heidelberg, New York, 1995. MR1353912

[18] E. Pardoux, and S. Peng. Adapted solution of a backward stochastic differential equation.

Systems Control Lett.14(1990), No. 1, 55–61. MR1037747

[19] P. Protter. Stochastic Integration and Differential Equations. Springer-Verlag Berlin Heidelberg, New York, 2004. MR2020294

[20] T. Rolski, H. Schmidli, V. Schmidt and J. Teugels. Stochastic Processes for Insurance and Finance. Wiley Series in Probability and Statistics, 1999. MR1680267

[21] P. L. Walker. On Lebesgue Integrable Derivatives. Am. Math. Mon.84(1977), No. 4, 287–288. MR0432835


Dokumen yang terkait

AN ALIS IS YU RID IS PUT USAN BE B AS DAL AM P E RKAR A TIND AK P IDA NA P E NY E RTA AN M E L AK U K A N P R AK T IK K E DO K T E RA N YA NG M E N G A K IB ATK AN M ATINYA P AS IE N ( PUT USA N N O MOR: 9 0/PID.B /2011/ PN.MD O)

0 82 16

ANALISIS FAKTOR YANGMEMPENGARUHI FERTILITAS PASANGAN USIA SUBUR DI DESA SEMBORO KECAMATAN SEMBORO KABUPATEN JEMBER TAHUN 2011

2 53 20

EFEKTIVITAS PENDIDIKAN KESEHATAN TENTANG PERTOLONGAN PERTAMA PADA KECELAKAAN (P3K) TERHADAP SIKAP MASYARAKAT DALAM PENANGANAN KORBAN KECELAKAAN LALU LINTAS (Studi Di Wilayah RT 05 RW 04 Kelurahan Sukun Kota Malang)

45 393 31

FAKTOR – FAKTOR YANG MEMPENGARUHI PENYERAPAN TENAGA KERJA INDUSTRI PENGOLAHAN BESAR DAN MENENGAH PADA TINGKAT KABUPATEN / KOTA DI JAWA TIMUR TAHUN 2006 - 2011

1 35 26

A DISCOURSE ANALYSIS ON “SPA: REGAIN BALANCE OF YOUR INNER AND OUTER BEAUTY” IN THE JAKARTA POST ON 4 MARCH 2011

9 161 13

Pengaruh kualitas aktiva produktif dan non performing financing terhadap return on asset perbankan syariah (Studi Pada 3 Bank Umum Syariah Tahun 2011 – 2014)

6 101 0

Pengaruh pemahaman fiqh muamalat mahasiswa terhadap keputusan membeli produk fashion palsu (study pada mahasiswa angkatan 2011 & 2012 prodi muamalat fakultas syariah dan hukum UIN Syarif Hidayatullah Jakarta)

0 22 0

Pendidikan Agama Islam Untuk Kelas 3 SD Kelas 3 Suyanto Suyoto 2011

4 108 178

ANALISIS NOTA KESEPAHAMAN ANTARA BANK INDONESIA, POLRI, DAN KEJAKSAAN REPUBLIK INDONESIA TAHUN 2011 SEBAGAI MEKANISME PERCEPATAN PENANGANAN TINDAK PIDANA PERBANKAN KHUSUSNYA BANK INDONESIA SEBAGAI PIHAK PELAPOR

1 17 40

KOORDINASI OTORITAS JASA KEUANGAN (OJK) DENGAN LEMBAGA PENJAMIN SIMPANAN (LPS) DAN BANK INDONESIA (BI) DALAM UPAYA PENANGANAN BANK BERMASALAH BERDASARKAN UNDANG-UNDANG RI NOMOR 21 TAHUN 2011 TENTANG OTORITAS JASA KEUANGAN

3 32 52