An Infinite Series Expression for the Joint Probability Density Function of OccupationTimes in a Three-State Markov Chain

Abstract

We obtain an infinite series expression for the joint probability density function of occupation times in a Three-State Markov chain. By extending the identity given by Darroch, we demonstrate that Pedler’s technique of employing moment generating functions can be applied to the three-state case as well.

Share and Cite:

Evans, J. and Korzeniowski, A. (2025) An Infinite Series Expression for the Joint Probability Density Function of OccupationTimes in a Three-State Markov Chain. Advances in Pure Mathematics, 15, 718-725. doi: 10.4236/apm.2025.1511038.

1. Introduction

The occupation time probability density function was the subject of several research papers in the 60s and 70s as seen in Darroch [1], Pedler [2], Good [3], and Hsia [4]. More recently, there has been renewed interest in the subject as seen in Sericola [5] as well as applications in Fang [6] and Guo [7]. In this paper, we extend Darroch’s identity to any finite state Markov chain and then derive an explicit three-state joint probability density function in a manner similar to Pedler’s original two-state derivation.

2. Occupation Time for a Three-State Markov Chain

Consider an irreducible continuous time Three-State Markov chain M( t ) with rate matrix R=( λ ij ), λ ij >0 if ij ranging over states E 1 , E 2 , E 3 with initial distribution Π 3 =[ π 1 , π 2 , π 3 ] where 0 π i 1 denotes the probability of starting in E i at time 0. Then for 0ut , the occupation times of E 1 , E 2 , E 3 by time t are

X( t )= 0 t 1 E 1 du ,Y( t )= 0 t 1 E 2 du ,Z( t )= 0 t 1 E 3 du

where X( t )+Y( t )+Z( t )=t and π 1 + π 2 + π 3 =1 .

Theorem 1 Given X( t )=x,Y( t )=y,t>0 , the joint probability density function of the occupation time for a Three-State Markov chain up to time t is given by

f( x,y,t )= e x( λ 12 + λ 13 λ 31 λ 32 )y( λ 21 + λ 23 λ 31 λ 32 )t( λ 31 + λ 32 ) g( x,y,tyx )

where

g( x,y,tyx )= π 1 m=1 l=1 m k=0 ml j=0 mlk b ( m,l,k,j ) x ml y mk1 ( txy ) mj1 ( ml )!( mk1 )!( mj1 )!

+ π 1 m=2 k=1 m1 j=0 mk b ( m,0,k,j ) x m y mk1 ( txy ) mj1 ( m )!( mk1 )!( mj1 )!

+ π 1 m=1 j=0 m1 b ( m,0,0,j ) x m y m1 ( txy ) mj1 ( m )!( m1 )!( mj1 )!

+ π 1 δ y m=1 b ( m,0,m,0 ) x m ( txy ) m1 ( m )!( m1 )!

+ π 1 δ z m=1 b ( m,0,0,m ) x m y m1 ( m )!( m1 )!

+ π 1 δ y δ z

+ π 2 m=2 l=1 m1 k=0 ml j=0 mlk b ( m,l,k,j ) x ml1 y mk ( txy ) mj1 ( ml1 )!( mk )!( mj1 )!

+ π 2 m=1 k=1 m j=0 mk b ( m,0,k,j ) x m1 y mk ( txy ) mj1 ( m1 )!( mk )!( mj1 )!

+ π 2 m=1 j=0 m1 b ( m,0,0,j ) x m1 y m ( txy ) mj1 ( m1 )!( m )!( mj1 )!

+ π 2 δ x m=1 b ( m,m,0,0 ) y m ( txy ) m1 ( m )!( m1 )!

+ π 2 δ z m=1 b ( m,0,0,m ) x m1 y m ( m1 )!( m )!

+ π 2 δ x δ z

+ π 3 m=1 l=0 m1 k=0 ml1 j=0 mlk b ( m,l,k,j ) x ml1 y mk1 (txy) mj ( ml1 )!( mk1 )!( mj )!

+ π 3 m=2 l=1 m1 b ( m,l,ml,0 ) x ml1 y l1 ( txy ) m ( ml1 )!( l1 )!( m )!

+ π 3 δ x m=1 b ( m,m,0,0 ) y m1 ( txy ) m ( m1 )!( m )!

+ π 3 δ y m=1 b ( m,0,m,0 ) x m1 ( txy ) m ( m1 )!( m )!

+ π 3 δ x δ y

+ c 1 m=1 l=0 m1 k=0 ml j=0 mlk b ( m,l,k,j ) x ml1 y mk ( txy ) mj ( ml1 )!( mk )!( mj )!

+ c 1 δ x m=1 b ( m,m,0,0 ) y m ( txy ) m ( m )!( m )!

+ c 1 δ x

+ c 2 m=1 l=1 m k=0 ml j=0 mlk b ( m,l,k,j ) x ml y mk1 ( txy ) mj ( ml )!( mk1 )!( mj )!

+ c 2 m=1 k=0 m1 j=0 mk b ( m,0,k,j ) x m y mk1 ( txy ) mj ( m )!( mk1 )!( mj )! (1)

+ c 2 δ y m=1 b ( m,0,m,0 ) x m ( txy ) m ( m )!( m )!

+ c 2 δ y

+ c 3 m=1 l=1 m k=0 ml j=0 mlk b ( m,l,k,j ) x ml y mk ( txy ) mj1 ( ml )!( mk )!( mj1 )!

+ c 3 m=1 k=1 m j=0 mk b ( m,0,k,j ) x m y mk ( txy ) mj1 ( m )!( mk )!( mj1 )!

+ c 3 m=1 j=0 m1 b ( m,0,0,j ) x m y m ( txy ) mj1 ( m )!( m )!( mj1 )!

+ c 3 δ z m=1 b ( m,0,0,m ) x m y m ( m )!( m )!

+ c 3 δ z

+ c 4 m=1 l=0 m k=0 ml j=0 mlk b ( m,l,k,j ) x ml y mk ( txy ) mj ( ml )!( mk )!( mj )!

+ c 4

c 1 = π 2 λ 23 + π 3 λ 32 c 2 = π 1 λ 13 + π 3 λ 31 c 3 = π 1 λ 12 + π 2 λ 21 c 4 = π 1 [ λ 12 λ 13 +( λ 13 λ 23 )( λ 32 λ 12 ) ]+ π 2 [ λ 21 λ 23 +( λ 31 λ 21 )( λ 23 λ 13 ) ] + π 3 [ λ 31 λ 32 +( λ 12 λ 32 )( λ 31 λ 21 ) ]

b( m,l,k,j ) =( m l )( ml k )( mlk j ) ( λ 23 λ 32 ) l ( λ 13 λ 31 ) k ( λ 12 λ 21 ) j ( λ 12 λ 23 λ 31 + λ 13 λ 21 λ 32 ) mlkj

and δ( x )= δ x ,δ( y )= δ y ,δ( txy )= δ z are Dirac delta functionals satisfying the property that α β δ( uk )f( u )du = α β δ( ku )f( u )du =f( k ) if k[ α,β ] and 0 otherwise.

3. Derivation of Three-State Occupation Time Probability Density Function

The proof of Theorem 1 is based on the following Lemmas:

Lemma 2

E[ e θ n1 X n1 ( t ) ]= Π n e t( RH ) 1 n (2)

where θ n1 X n1 ( t ) is the dot product of θ n1 =[ k 1 θ 1 , k 2 θ 2 ,, k n1 θ n1 ] and X n1 ( t )= [ X 1 ( t ), X 2 ( t ),, X n1 ( t ) ] T , Π n =[ π 1 , π 2 ,, π n ] , 1 n the column vector of 1s, X i ( t )= 0 t 1 E i du the occupation time of state i by time t>0 , and

H=[ k 1 θ 1 0 0 0 k n1 θ n1 0 0 0 0 ]

Proof of Lemma 2.

From Darroch [1], one may write the moment generating function of the weighted sum of states as

E[ e μY( t ) ]= Π n e t( Rμ D 1 ) 1 n

with D 1 =[ D 0 0 0 ] , D the matrix with elements δ ij d i , Y( t )= i=1 s d i X i ( t ) where d i are arbitrary weights, δ ij the Kronecker delta, and 1s<n

Then setting d i = θ i k i μ , with k i arbitrary, the moment generating function is extended to a multivariate moment generating function and we obtain (2) as desired.

Lemma 3

The inverse Laplace transform of the expression below reads

1 [ π 1 vw m=0 l=0 m k=0 ml j=0 mlk b ( m,l,k,j ) u m+l1 v m+k1 w m+j1 ]( x,y,tyx ) = π 1 m=1 l=1 m k=0 ml j=0 mlk b ( m,l,k,j ) x ml y mk1 ( txy ) mj1 ( ml )!( mk1 )!( mj1 )! + π 1 m=2 k=1 m1 j=0 mk b ( m,0,k,j ) x m y mk1 ( txy ) mj1 ( m )!( mk1 )!( mj1 )! + π 1 m=1 j=0 m1 b ( m,0,0,j ) x m y m1 ( txy ) mj1 ( m )!( m1 )!( mj1 )! + π 1 δ y m=1 b ( m,0,m,0 ) x m ( txy ) m1 ( m )!( m1 )! + π 1 δ z m=1 b ( m,0,0,m ) x m y m1 ( m )!( m1 )! + π 1 δ y δ z (3)

Proof of Lemma 3.

Recall that 1 [ 1 u n ]( x )=δ( x ) if n=0 and 1 [ 1 u n ]( x )= x n1 ( n1 )! if n is a positive integer where the latter identity may be obtained by repeated integration by parts. Splitting

π 1 vw m=0 l=0 m k=0 ml j=0 mlk b ( m,l,k,j ) u m+l1 v m+k1 w m+j1 (4)

into cases where the exponent(s) of v,w are zero or non-zero, we may ascertain that (4) is equivalent to

π 1 u + π 1 m=1 l=1 m k=0 ml j=0 mlk b ( m,l,k,j ) u m+l1 v m+k w m+j + π 1 m=2 k=1 m1 j=0 mk b ( m,0,k,j ) u m1 v m+k w m+j + π 1 m=1 j=0 m1 b ( m,0,0,j ) u m1 v m w m+j + π 1 m=1 b ( m,0,m,0 ) u m1 w m + π 1 m=1 b ( m,0,0,m ) u m1 v m (5)

Taking the inverse Laplace transform of (5) gives (3) as desired.

Proof of Theorem 1.

Let the joint probability density function of X( t ),Y( t ),Z( t ) be f( x,y,t ) and denote the bivariate moment generating function of f( x,y,t ) by E[ e θ 1 x θ 2 y ] . Then taking the Laplace transform of E[ e θ 1 x θ 2 y ] and applying Lemma 2 for n=3 and k i =1 , we obtain

{ f( x,y,t ) }( θ 1 , θ 2 ,ϕ ) = 0 0 0 f ( x,y,t ) e θ 1 x θ 2 yϕt dydxdt = 0 E[ e θ 1 x θ 2 y ] e ϕt dt = 0 Π 3 e t( RH ) e ϕt 1 3 dt = Π 3 [ RϕIH ] 1 1 3 = π 1 vw+ π 2 uw+ π 3 uv+ c 1 u+ c 2 v+ c 3 w+ c 4 uvwu λ 23 λ 32 v λ 13 λ 31 w λ 12 λ 21 λ 12 λ 23 λ 31 λ 13 λ 21 λ 32 (6)

where

[ RϕIH ]=[ u λ 12 λ 13 λ 21 v λ 23 λ 31 λ 32 w ], Π 3 =[ π 1 , π 2 , π 3 ], 1 3 =[ 1 1 1 ]

u= θ 1 +ϕ+ λ 12 + λ 13 ,v= θ 2 +ϕ+ λ 21 + λ 23 ,w=ϕ+ λ 31 + λ 32

Denoting the numerator of (6) as N and factoring ( uvw ) 1 , (6) may be written as

[ N 1 u λ 23 λ 32 +v λ 13 λ 31 +w λ 12 λ 21 + λ 12 λ 23 λ 31 + λ 13 λ 21 λ 32 uvw ] 1 ( uvw ) 1 (7)

Then for sufficiently large θ 1 , θ 2 ,ϕ we may expand (7) as a geometric series and rewrite it as

N m=0 l=0 m k=0 ml j=0 mlk b ( m,l,k,j ) u m+l1 v m+k1 w m+j1 ={ g( x,y,t ) }( u,v,w ) ={ e x( λ 12 + λ 13 λ 31 λ 32 )y( λ 21 + λ 23 λ 31 λ 32 )t( λ 31 + λ 32 ) g( x,y,tyx ) }( θ 1 , θ 2 ,ϕ ) (8)

From the equality of (8) and (6), we have

f( x,y,t )= e x( λ 12 + λ 13 λ 31 λ 32 )y( λ 21 + λ 23 λ 31 λ 32 )t( λ 31 + λ 32 ) g( x,y,tyx )

Finally, by taking the inverse Laplace transform of (8) in a manner akin to the proof of Lemma 3 for all terms gives (1) as desired.

4. Illustrative Example

We present a table with values provided by a Wolfram Mathematica implementation of (1) as well as a simulation of 109 paths in C++. The Mathematica implementation of (1) involves only a summation of index m up to 40 terms instead of the infinite summation of (1), and for the path simulation the proportion of paths satisfying x 1 x< x 2 , y 1 y< y 2 was used to calculate the simulated probabilities. The relevant parameters are as follows.

R=[ 9 3 6 4 6 2 1 3 4 ], Π 3 =[ 1,0,0 ]

x 1 x 2 y 1 y 2 f ( x,y,5 )dydx

0x<1

1x<2

2x<3

3x<4

4x<5

5x<6

0y<1

0.036051100

0.055836600

0.002708140

0.000005337

0.000000001

0

1y<2

0.290716000

0.384080000

0.009010030

0.000003378

0

0

2y<3

0.123652000

0.093462900

0.000392612

0

0

0

3y<4

0.003429840

0.000652630

0

0

0

0

4y<5

0.000003147

0

0

0

0

0

5y<6

0

0

0

0

0

0

Path Simulation

0x<1

1x<2

2x<3

3x<4

4x<5

5x<6

0y<1

0.036047441

0.055842521

0.002709707

0.000005406

0

0

1y<2

0.290723689

0.384056728

0.009008985

0.000003493

0

0

2y<3

0.123652673

0.093470497

0.000393724

0

0

0

3y<4

0.003433051

0.000649946

0

0

0

0

4y<5

0.000002139

0

0

0

0

0

5y<6

0

0

0

0

0

0

Comparison of the simulated values with the exact values in the table above shows good agreement, whereby offering Monte Carlo method as an alternative to exact pdf formula which are expected to be very lengthy for n larger than 3.

Without a closed form expression, considerations must be made when determining an upper bound for index m in (1). An appropriate upper bound for m often depends on t and R where sufficiently large values of t can result in the truncated version of (1) performing poorly. An informal heuristic for gauging the accuracy of the truncated pdf is by confirming that its integral over its domain is sufficiently close to 1. The authors decision to pick the rate matrix and initial distribution above is based off of this observation.

5. Discussions

We would like to note that in principle, one may derive the occupation time probability density function for any finite state Markov chain by utilizing the relevant extension of Darroch’s identity as given by Lemma 2. However, as seen in (1), even for a Three-State Markov chain, the formula is rather lengthy and its derivation is more suited to a computer algebra system. To the best of the authors’ knowledge, (1) is the first explicit example of the occupation time probability density function for a Three-State Markov chain, and is also the first demonstration that Darroch’s identity and Pedler’s derivation may be extended beyond two states. As far as we are aware, Sericola’s approach appears to be the only other method that may be used to derive the occupation time probability density function in a finite state Markov chain; however, we did not find any explicit results in the literature. Concerning Hsia’ result, the probability density function found in Hsia [4] is expressed in terms of random variables with unknown distributions and cannot be directly integrated to calculate numerical values.

On a final note, we would like the reader to be aware that there are several published results for the Two-State occupation time probability density function that contain typographical errors. Among incorrectly typed formulas, the omission of the Dirac delta functional is particularly common and the resulting probability density function fails to integrate to 1 over its domain. Regarding the Two-State occupation time probability density function, Pedler’s formula as written was verified to be correct. As for future work, it would be desirable to have a closed form for various parts of (1) if it exists. Alternatively, new methods for deriving the probability density function may yield more tractable representations of (1).

Acknowledgements

The authors would like to thank the reviewers for their valuable suggestions that considerably improved the presentation of our findings in the revised version of the manuscript.

Conflicts of Interest

The authors declare no conflicts of interest regarding the publication of this paper.

References

[1] Darroch, J.N. and Morris, K.W. (1968) Passage-Time Generating Functions for Continuous-Time Finite Markov Chains. Journal of Applied Probability, 5, 414-426.[CrossRef
[2] Pedler, P.J. (1971) Occupation Times for Two State Markov Chains. Journal of Applied Probability, 8, 381-390.[CrossRef
[3] Good, I.J. (1961) The Frequency Count of a Markov Chain and the Transition to Continuous Time. The Annals of Mathematical Statistics, 32, 41-48.[CrossRef
[4] Hsia, W.S. (1976) The Joint Probability Density Function of the Occupation Time of a Three-State Problem. Journal of Applied Probability, 13, 57-64.[CrossRef
[5] Sericola, B. (2000) Occupation Times in Markov Processes. Communications in Statistics. Stochastic Models, 16, 479-510.[CrossRef
[6] Fang, C.Y., Liu, Y., Shi, Z.Y. and Chen, C. (2023) Closed-Form Expression of Geometric Brownian Motion with Regime-Switching and Its Applications to European Option Pricing. Symmetry, 15, 575.[CrossRef
[7] Guo, X. (1999) Inside Information and Stock Fluctuations.
https://www.researchgate.net/publication/2499237_Inside_Information_And_Stock_Fluctuations

Copyright © 2025 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.