Analysis of a POMDP Model for an Optimal Maintenance Problem with Multiple Imperfect Repairs

Abstract

I consider a system whose deterioration follows a discrete-time and discrete-state Markov chain with an absorbing state. When the system is put into practice, I may select operation (wait), imperfect repair, or replacement at each discrete-time point. The true state of the system is not known when it is operated. Instead, the system is monitored after operation and some incomplete information concerned with the deterioration is obtained for decision making. Since there are multiple imperfect repairs, I can select one option from them when the imperfect repair is preferable to operation and replacement. To express this situation, I propose a POMDP model and theoretically investigate the structure of an optimal maintenance policy minimizing a total expected discounted cost for an unbounded horizon. Then two stochastic orders are used for the analysis of our problem.

Share and Cite:

Tamura, N. (2023) Analysis of a POMDP Model for an Optimal Maintenance Problem with Multiple Imperfect Repairs. American Journal of Operations Research, 13, 133-146. doi: 10.4236/ajor.2023.136008.

1. Introduction

With recent advances in science and technology, many systems have become more complex and more massive. Accordingly, the breakdown or deterioration of such systems could frequently result in economic and social damage in our communities. To address those problems mathematically, preventive maintenance models for stochastically deteriorating systems have been widely investigated in the literature. Osaki [1] summarized the mathematical approach for reliability and maintenance. And the papers by Cho and Parlar [2] , Dekker [3] , and Wang [4] are excellent reviews of the subject. Since various techniques for detecting the deterioration level of systems have been developed and available without expensive cost, one can adopt condition-based maintenance (CBM) rather than time-based maintenance (TBM) in many cases. The above three reviews include studies on CBM as well as TBM. Jardine et al. [5] summarized and reviewed the recent research and developments in diagnostics and prognostics of mechanical systems implementing CBM with algorithms and technologies for data processing. Also, Olde Keizer et al. [6] and Alaswad and Xiang [7] reviewed stochastic models for addressing issues on CBM policies.

For constructing a stochastic model for CBM, I need first express a deterioration of a system as some kind of stochastic process. If one can assume that the deterioration levels of a system correspond to finite non-negative integers, then the deterioration of the system could be described as a Markov chain with an absorbing state which expresses a failure of the system. This model is called Markovian deteriorating system and Markov decision processes (MDP) are used for addressing maintenance problems. Derman [8] , Douer and Yechiali [9] , Hopp [10] , Kurt and Kharoufeh [11] [12] , and Tamura [13] studied the discrete-time Markovian deteriorating system and derived the sufficient conditions that some monotone properties including the control-limit policy hold. Also, Lam and Yeh [14] [15] [16] , and Yeh [17] [18] studied the continuous-time Markovian deteriorating system. These studies have some barriers from realistic viewpoints because it is too costly or too technically difficult to detect the true state of a system in practical situations. In these cases, a system is monitored and some incomplete information on the true state of the system is obtained to select an optimal action. When I observe alternative characteristics instead of a true deterioration level of a system at each inspection, the result has some kind of relationship with the true deterioration level although it does not express the true deterioration level. Then, partially observable Markov decision process (POMDP) models are available to express the behavior of such systems. POMDP models use a so-called state vector whose element means the probability of a true state. If an element of a state vector is always 1 in a POMDP model, it coincides with an MDP model. Ross [19] incorporated the concept of incomplete information into the Markovian system with two states. However, it was a little different from the POMDP model. Afterwards, Ohnishi et al. [20] , Grosfeld-Nir [21] [22] , and Chen et al. [23] studied the POMDP model for the maintenance problem where only replacement can be selected as a maintenance action. Srinivasan and Parlikad [24] studied the extended POMDP model which is called a POSMDP (Partially Observable Semi-Markov Decision Process) for optimizing maintenance decisions. Oosterom [25] used the new stochastic order to analyze the POMDP model. And Kivanc et al. [26] applied the POMDP model to real problems. On the other hand, Ohnishi et al. [27] proposed the POMDP model with minimal repair as well as replacement and showed that the monotone properties hold under several assumptions. Pham [28] reviewed the studies on imperfect maintenance. Then, Ivy and Pollock [29] introduced the concept of repair, which is different from minimal repair, in addition to replacement for the POMDP model and derived the marginal monotonicity for the two-state model. Tamura et al. [30] and Fan et al. [31] also studied the POMDP model with repair and replacement. Then, these models assumed one option for repair. In practical situations, however, I can choose some options for repair of which the effects are different.

In this paper, I construct a POMDP model where one of H imperfect repairs can be selected as a maintenance action as well as replacement. For the model, I theoretically investigate the structure of an optimal maintenance policy minimizing the total expected discounted cost for unbounded horizon via the theory of stochastic order relations. In particular, I derive sufficient conditions for monotone properties of the optimal maintenance policy.

2. Model Description

Now I consider a system whose deterioration follows a discrete-state and discrete-time Markov chain with an absorbing state. The system can be classified into one of N + 2 states, and then, state 0 represents the process before any deterioration takes place that is, it is an initial new state, whereas, N + 1 is a failure state of the system. The intermediates state 1 ∙∙∙ N are ordered to reflect the relative degree of deterioration in ascending order. Thus, I let p i j denote the transition probability of the Markov chain and let P denote the transition probability matrix. S = { 0,1, , N + 1 } expresses that state space of the Markov chain. And I let x i denote the probability that the true state of the system is i S . Also, e i means the N + 2 dimensional vector that the ith element is 1 and other elements are all zero. The model supposes that an action is selected based on the state vector x instead of the true state i. Thus, I introduce γ i θ which means the probability that an outcome θ is obtained through monitor given that the true state of the system is i, where ( Γ ) i θ = γ i θ and θ M = { 1, , M } . To express that the result of repair is uncertain as well as imperfect, I let q i j ( h ) denote the repair probability for the system when repair h is selected, where ( Q ( h ) ) i j = q i j ( h ) and h H = { 1, , H } . That is, H is a set of options for imperfect repair.

Under these settings, at each time epoch, I can select wait, repair h ( 1 h H ), or replacement as follows:

Wait: The system is operated for one period and it isn’t directly inspected after operation. Thus, the system in state i moves to state j with probability p i j after one period with operation. Instead of inspection, I monitor the system and obtain an outcome θ with probability γ j θ .

Imperfect repair: I decided on an option among the H possible imperfect repairs. If repair h is selected as an optimal option, then the true state of the system is identified through inspection before repair and the system in state i is repaired to state j with probability q i j ( h ) .

Imperfect repair: The system is replaced by a new and identical one. Thus, after replacement, the system in state i is certainly returned to state 0.

In the next, I assume some costs and probabilities for the model. Hereafter, I consider that increasing means non-decreasing and decreasing means non-increasing, respectively. And c Τ means a transpose of a vector c . Before explanation, two stochastic order relations are defined below.

Definition 1. For a non-negative matrix B on M × M , B T P 2 means B has a property of totally positive of order 2, that is, b i j b k l b k j b i l holds for i < k M and j < l M . Similarly, B S I means B has a property of usual stochastic order, that is, for any h, j = h M b i j j = h M b k j holds when i k . Furthermore, for non-negative vectors a = ( a 1 , , a K ) Τ and b = ( b 1 , , b K ) Τ , if a matrix ( a Τ b Τ ) T P 2 then I write a T b . And if j = h K a j j = h K b j , then I write a s t b .

For the above stochastic order relations, I provide some important results as a lemma.

Lemma 1.

(i) If a T b then a s t b .

(ii) For any increasing function g i on i where g = ( g 1 , , g K ) Τ , if a s t b then a Τ g b Τ g .

(iii) If B T P 2 then B S I .

The details of the stochastic order relations including T P 2 and SI are explained by Kijima [32] and Shaked and Shanthikumar [33] . Thus, I omit the proof of Lemma 1.

I introduce the following assumptions for the probabilities P , Q ( h ) , and Γ .

Assumption 1. P T P 2 and Γ T P 2 .

Assumption 2. For any h H , Q ( h ) T P 2 .

In Assumption 1, the assumption on the transition probability matrix means that as the system deteriorates, it is more likely to move to worse states, and the assumption on Γ means that the higher deterioration of the system gives higher outcomes from the monitor. These two assumptions in Assumption 1 have been used in the concerned previous studies to derive the monotone properties. Assumption 2 means that as the system deteriorates, it is less likely to be returned to better states through repair.

When each action is selected, a cost is incurred as follows:

u i : operating cost for the system in state i S .

r i ( h ) : repair cost for the system in state i S under repair h H .

c i : replacement cost for the system in state i S .

And I define u = ( u 0 , , u N + 1 ) Τ , r ( h ) = ( r 0 ( h ) , , r N + 1 ( h ) ) Τ , c = ( c 0 , , c N + 1 ) Τ , F N + 1 = { f R N + 1 | f i 1 f i ( 1 i N + 1 ) } , where R N + 1 means a set of N + 1 dimensional vectors whose elements are real numbers. For these costs, I introduce the following assumptions.

Assumption 3. For any h H , u F N + 2 , r ( h ) F N + 2 , and c F N + 2 .

Assumption 4. For any h H , u r ( h ) F N + 2 , r ( h ) c F N + 2 , and u c F N + 2 .

Assumption 3 indicates that as the system deteriorates, it is more costly to operate, repair, or replace the system. Also, the first relation of Assumption 4 means that, as the system deteriorates, the merit of repair becomes bigger than that of operation. Similarly, in terms of the latter two relations, the merit of replacement becomes bigger than those of operation and repair with deterioration, respectively. The previous studies have also assumed these relations for the analysis of the POMDP models.

3. Formulation

This section provides the formulation of our problem as a partially observable Markov process model and derives a total expected discounted cost for an unbounded horizon. For the formulation, I define x = ( x 0 , x 1 , , x N + 1 ) Τ where x is the probability that the true state of the system is i as a state vector. And I let σ j ( x , θ ) denote the probability that the true state is j when the current state vector is x and the outcome is obtained as θ through monitor after operating the system. And β is a discount factor. Then, by using Bayes’ formula, I have

σ j ( x , θ ) = i x i p i j γ j θ i j x i p i j γ j θ , (1)

and I define the jth element of σ ( x , θ ) is σ j ( x , θ ) . Also, I let σ ( θ | x ) denote the probability that the outcome is θ through monitor given that the current state vector is x and I have

σ ( θ | x ) = i x i p i j γ j θ . (2)

Hence I let W ( x ) , R h ( x ) , and R t ( x ) denote the optimal expected discounted cost when wait, repair h, and replacement are selected at state vector x , respectively. When W ( x ) is expressed as

W ( x ) = u Τ x + β θ M σ ( θ | x ) V ( σ ( x , θ ) ) (3)

Also, I have

R h ( x ) = r ( h ) Τ x + β j ( x Q ( h ) ) j V ( e j ) , (4)

R t ( x ) = c Τ x + β V ( e 0 ) , (5)

where ( x Τ Q ( h ) ) j = i x i q i j ( h ) .

Now I suppose that V ( x ) is the optimal expected discounted cost when an optimal action is selected at state vector x and the optimal maintenance policy is adopted afterwards. Then I have

V ( x ) = min [ W ( x ) , min 1 h H R h ( x ) , R t ( x ) ] . (6)

This problem can be treated as a Markov decision process with a state space X where X = { x : x i 0, i = 0 N + 1 x i = 1 ( i S ) } . Thus, by solving Equation (6), I can find the optimal maintenance policy which minimizes the total expected discounted cost for an unbounded horizon. Hereafter, for simplicity, when a function h ( x ) satisfies h ( x ) h ( x ˜ ) for x T x ˜ , I describe this relation as h ( x ) is increasing x with respect to T P 2 .

4. Structure of the Optimal Maintenance Policy

In this section, I provide some results on the structure of the optimal maintenance policy. Before proceeding discussion, I summarize several results as the following lemm in preparation.

Lemma 2.

(i) If x T x ˜ then x Τ P T x ˜ Τ P and x Τ Q ( h ) T x ˜ Τ Q ( h ) .

(ii) If x T x ˜ then x Τ P G T x ˜ Τ P G .

(iii) If x T x ˜ then σ ( x , θ ) T σ ( x ˜ , θ ) .

(iv) If θ < θ ˜ then σ ( x , θ ) T σ ( x , θ ˜ ) .

(v) If a real valued function f ( x , θ ) is increasing in θ for any x X and is increasing in x with respect to T P 2 , then θ σ ( θ | x ) f ( x , θ ) θ σ ( θ | x ) f ( x ˜ , θ ) . The results (i)-(v) are proved by Ohinishi et al. [20] .

4.1. Basic Results

Next, some lemmas which are necessary to derive the main results of this study are developed. In previous studies, similar lemmas were presented and used to derive the monotone properties.

Theorem 1. V ( x ) is an increasing function of x in terms of T P 2 and is concave in x .

Proof. First, I show that the characteristics of W ( x ) , R h ( x ) , and R t ( x ) , respectively. V n ( x ) is assumed to be the expected cost when an optimal action is selected for the n-period version of our problem and I use the mathematical induction method, where V 0 ( x ) = 0 . Then, I have V n ( x ) = min [ W n ( x ) , min 1 h H R h n ( x ) , R t n ( x ) ] , W n ( x ) = u Τ x + β θ M σ ( θ | x ) V n 1 ( σ ( x , θ ) ) , R h n ( x ) = r ( h ) Τ x + β j ( x Q ( h ) ) j V n 1 ( e j ) , R t n ( x ) = c Τ x + β V n 1 ( e 0 ) . When n = 1 , W 1 ( x ) , R h 1 ( x ) , and R t 1 ( x ) are increasing in x from Assumption 3. Thus, V 1 ( x ) is also increasing in x . Thus, I suppose that W n ( x ) , R h n ( x ) , R t n ( x ) are increasing in x . The first terms of these three functions are increasing in x from Assumption 3. Since the second term of W ( x ) is increasing in x from Lemma 2 (v) and the assumption on V n ( x ) , W n + 1 ( x ) is increasing in x . The second term of R h n ( x ) is a linear function of x and j ( x Q ( h ) ) j V n 1 ( e j ) j ( x ˜ Q ( h ) ) j V n 1 ( e j ) for x T x ˜ from e i T e j for i j and Lemma 1 (i)-(ii) and Lemma 2 (i). Thus, R h n + 1 ( x ) is increasing in x . And R t n + 1 ( x ) is increasing in x since the second term is constant, that is, it does not depend upon x .

Second, I consider V n + 1 ( x ) through the results on W n + 1 ( x ) , R h n + 1 ( x ) , and R t n + 1 ( x ) . When I use the argument similar to those of Ohnishi et al. [20] , the second term of W n + 1 ( x ) is increasing and concave in x . Since R h n + 1 ( x ) and R t n + 1 ( x ) are increasing and linear in x , V n + 1 ( x ) is increasing and concave in x . Hence the proof of this theorem completes. □

In addition, I have a result on the difference of the costs as follows.

Lemma 3. For any h, W ( x ) R h ( x ) is concave in x . W ( x ) R t ( x ) and R h ( x ) R t ( x ) are increasing in x with respect to T P 2 .

Proof. Since

W ( x ) R h ( x ) = ( u r ( h ) ) x + β { θ M σ ( θ | x ) V ( σ ( x , θ ) ) j ( x Q ( h ) ) j V ( e j ) } , (7)

the first and the third terms are linear, the second term is concave in x from the proof of Theorem 1. As a result, the difference between the second and the third term is concave in x . Thus. Equation (7) is concave in x for any h. And since

W ( x ) R t ( x ) = ( u c ) x + β { σ ( θ | x ) V ( σ ( x , θ ) ) V ( e 0 ) } , (8)

the first term is increasing in x from Assumption 4, the second term is also increasing in x from the proof of Theorem 1, and the third term is constant for x . Thus, Equation (8) is increasing in x . And finally, since

R h ( x ) R t ( x ) = ( r ( h ) c ) x + β { j ( x Q ( h ) ) j V ( e j ) V ( e 0 ) } , (9)

the first term is increasing in x from Assumption 4, the second term is also increasing in x from the proof of Theorem 1, and the third term is constant for x . Thus, Equation (9) is increasing in x for any h. Hence, the proof of this lemma completes. □

4.2. Monotone Properties

Next, I provide some structural properties of the optimal maintenance policy. For the explanation, I let D ( x ) denote an optimal action selected at state vector x . And I define h arg min 1 h H [ R h ( x ) ] , R h min 1 h H [ R h ( x ) ] .

Then, I have the optimal maintenance policy has the structure expressed by Theorem 2.

Theorem 2. There exist real numbers α ˜ , α ^ , α ¯ such that D ( x α ) = ( W ( 0 α < α ˜ ) , R h ( α ˜ α < α ^ ) , W ( α ^ α < α ¯ ) , R T ( α ¯ α 1 ) , where x α = ( 1 α ) x + α x for x T x and 0 α ˜ α ^ α ¯ 1 .

Proof. I use Theorem 1 and Lemma 3 for the proof. Since W ( x ) R h ( x ) is concave, W ( x ) and R h ( x ) intersect twice at most. Also, since R h ( x ) R t ( x ) and W ( x ) R t ( x ) are increasing in x , as a result, each pair of R h ( x ) and R t ( x ) and W ( x ) and R t ( x ) intersect once at most. Hence the proof of Theorem 2 completes.

Theorem 2 means that the optimal maintenance policy may be classified into 4 regions at most. In particular, since repair includes inspection proposed by Ohnishi et al. [20] if I suppose q i i ( h ) = 1 for any i S and any h H , the structure of the optimal maintenance policy is also similar to that of Ohnishi et al. [20] . Except for Ohnishi et al. [20] , some studies have previously derived monotone properties by introducing several assumptions on costs and probabilities. However, those do not consider that an optimal repair is selected from multiple imperfect repairs whose respective effects are different.

Also, Theorem 2 gives us the following interpretation. In this section, hereafter, I use x instead of x α . When a state vector is less than x ˜ or between x ^ and x ¯ with respect to T P 2 order, I should operate the system, that is, maintenance actions are not necessary. Focused on maintenance actions, if a state vector stays between x ˜ and x ^ then an optimal imperfect repair is selected and finally, the system should be replaced whenever a state vector exceeds x ¯ .

Theorem 2 is useful for numerically finding an optimal maintenance policy. However, if the sequence of optimal imperfect repairs has some characteristics when D ( x ) = R h , then it would be possible to solve the optimal maintenance problem further efficiently. Thus, by introducing some additional assumptions, I obtain structural properties focused on imperfect repair.

Assumption 5. For any h , h such that h < h , I suppose that,

(i) r ( h ) r ( h ) F ,

(ii) Q ( h ) Q ( h ) S I ,

(iii) r ( h ) r ( h ) F ,

(iv) Q ( h ) Q ( h ) S I .

Assumption 5-(i) and 5-(iii), and Assumptions 5-(ii) and 5-(iv) do not hold together, respectively. Then, Assumptions 5-(i) indicates that it is less costly to select repair h compared with repair h' and Assumption 5-(ii) indicates that the system is more likely to be returned to better states through repair h compared with repair h', as the system deteriorates. Thus, Assumptions 5-(i) and 5-(ii) mean that the merit of repair h becomes bigger than that of repair h’ with deterioration of the system. Also, Assumptions 5-(iii) and 5-(iv) have the contrary interpretation.

Under some of the assumptions given by Assumption 5, I have a property on the relation of the optimal repairs.

Theorem 3. I suppose that D ( x l ) = R h l and D ( x u ) = R h u for x l T x u . If Assumptions 5-(i) and 5-(ii) hold then h l h u . And if Assumptions 5-(iii) and (iv) hold then h l h u .

Proof. I define that, for x 1 T x 2 and h 1 h 2 , Δ R h 1 , h 2 ( x ) R h 1 ( x ) R h 2 ( x ) = ( r ( h 1 ) r ( h 2 ) ) x + β j [ x { Q ( h 1 ) Q ( h 2 ) } ] j V ( e j ) , Δ 2 R h 1 , h 2 ( x 2 , x 1 ) Δ R h 1 , h 2 ( x 2 ) Δ R h 1 , h 2 ( x 1 ) = { r ( h 1 ) r ( h 2 ) } ( x 2 x 1 ) + β j [ ( x 2 x 1 ) { Q ( h 1 ) Q ( h 2 ) } ] j V ( e j ) . Also, I suppose that D ( x 2 ) = R h and D ( x 1 ) = R h . If Assumptions 5-(i) and 5-(ii) hold then Δ 2 R h 1 , h 2 ( x 2 , x 1 ) 0, and this means that there exists a vector x C 1 which satisfies Δ R h 1 , h 2 ( x ) > 0 for x T x C 1 , Δ R h 1 , h 2 ( x ) 0 for x T x C 1 . Hence, if D ( x ˜ 1 ) = R h 2 and D ( x ˜ 2 ) = R h 1 then x ˜ 1 T x ˜ 2 since I suppose that h 1 h 2 . And if Assumptions 5-(iii) and 5-(iv) hold then Δ 2 R h 1 , h 2 ( x 2 , x 1 ) 0. This indicates that there exists a vector x C 2 which satisfies Δ R h 1 , h 2 ( x ) 0 for x T x C 2 , Δ R h 1 , h 2 ( x ) > 0 for x T x C 2 . Thus, if D ( x ˜ 1 ) = R h 2 and D ( x ˜ 2 ) = R h 1 then x ˜ 1 T x ˜ 2 Hence, the proof of this theorem completes.

Theorem 3-(i) means that I should select larger numbered repairs in lower state vectors in terms of T P 2 and Theorem 3-(ii) has the contrary interpretation. This theorem is also useful for finding an optimal maintenance policy more efficiently. For example, I suppose that there are H options on imperfect repair and D ( x ) = R h for x [ x ˜ , x ^ ) in terms of T P 2 In this case, Theorem 3 tells us that, under Assumption 5, the number of options on imperfect repair gradually reduces from H as a state vector in [ x ˜ , x ^ ) increases, that is, the system deteriorates.

5. Discussion

In this section, I explain the significance of Theorems 2 and 3 to find an optimal maintenance policy. Generally, I know that monotone properties are available to find an optimal maintenance policy numerically.

Now this section supposes that the system has two states as good (0) and bad (1) and there are three options for imperfect repair ( H = { 1,2,3 } ). And the state vector x becomes a scalar x which means a probability that the system stays in a bad state. In this case, since I have to decide on an optimal action D ( x ) for x [ 0,1 ] , our problem has an infinite state space. However, if Theorem 2 holds, that is, I can know the problem satisfies Assumptions 1, 2, 3 and 4 in advance, then I may find 3 real numbers α ˜ , α ^ , α ¯ in Theorem 2. Then, I first compare W ( x ) and R h ( x ) for x [ 0,1 ] and find real numbers α ˜ and α ^ such that W ( x ) R h ( x ) for x [ 0 , α ˜ ) , W ( x ) R h ( x ) for x [ α ^ , 1 ) , W ( x ) > R h ( x ) for x [ α ˜ , α ^ ) . Second, similarly, I compare W ( x ) and R t ( x ) for x [ α ^ ,1 ] and find α ¯ such that W ( x ) > R t ( x ) for x [ α ¯ ,1 ] , W ( x ) R t ( x ) for x [ α ^ , α ¯ ) , since W ( x ) R h ( x ) for x [ α ^ ,1 ) . Thus, Theorem 2 can reduce the computational effort for finding an optimal maintenance policy numerically. This usefulness is conserved for more than two states.

To find the optimal maintenance policy in the above, I need obtain R h ( x ) , that is, the comparison of R h ( x ) for h H arises in this problem. Then Theorem 3 is useful to address the problem. Since I have 3 options for imperfect repair, if the costs and the probabilities satisfy Assumptions 5-(iii) and 5-(iv) then Theorem 3-(ii) holds. In this case, comparing R 1 ( x ) with W ( x ) , I find α ˜ 1 such that R 1 ( x ) W ( x ) for x [ 0,1 ] as a first step. In the next, I find α ˜ 2 that D ( x ) = R 2 , that is, R 2 ( x ) < R 1 ( x ) for x [ α ˜ 1 , α ^ ] through comparison of R 1 ( x ) and R 2 ( x ) . And finaly, I find α ˜ 3 that R 3 ( x ) < R 2 ( x ) for x [ α ˜ 2 ,1 ] through comparison of R 2 ( x ) and R 3 ( x ) . As a result of this procedure, I can find the real numbers α ˜ 1 , α ˜ 2 , α ˜ 3 , α ^ and α ¯ , and determine the optimal maintenance policy that satisfy α ˜ 1 α ˜ 2 α ˜ 3 α ^ α ¯ . On the other hand, if the costs and the probabilities satisfy Assumptions 5-(i) and 5-(ii) then I may use Theorem 3-(i) to find an optimal maintenance policy where α ˜ 3 α ˜ 2 α ˜ 1 α ^ α ¯ .

Even if the number of states is more than 2, I can similarly use Theorems 2 and 3. However, the computational effort becomes much harder with larger states.

6. Conclusions

In this paper, I considered a system whose deterioration follows a discrete-time and discrete-state Markov chain under incomplete information on the true state and constructed a POMDP model for an optimal maintenance problem. I have several options for repair which has different recovery effects on the system in addition to replacement as maintenance actions. Then, I derived sufficient conditions that the monotone property holds as per Theorem 2. As described in Section 4.2, since this theorem is a generalization of some previous results and is useful for efficiently solving the optimal maintenance problem since the optimal maintenance policy may be classified into 4 regions at most.

Furthermore, I established the sequence of optimal repairs under several assumptions in Theorem 3. Since the previous studies on POMDP models for optimal maintenance problems do not focus on the concept of multiple imperfect repairs that the repair probabilities are different, Theorem 3 is a significant property in this area even if the assumptions are so strict. In particular, the feature that the number of options on imperfect repair gradually reduces with the increase of a state vector in terms of T P 2 is so interesting.

I can check whether the costs for the respective actions and the transition and repair probabilities satisfy the assumptions for Theorems 2 and 3 before solving an optimal maintenance problem. Thus, as discussed in Section 5, the theorems obtained in this study are available to reduce the computational time since the number of regions of the optimal maintenance policy can be limited.

However, I do not provide quantitative analysis through numerical experiments. While there exist algorithms for POMDP models (see e.g. Monahan [34] , Guo and Lian [35] ), as stated by the previous studies, POMDP models with many states have an aspect of time-consuming computation. Thus, future works are to construct an efficient algorithm for solving our problem under the assumptions for the monotone properties and to obtain some characteristics of the optimal maintenance policy through numerical analysis. Also, I need to consider how to relax the sufficient conditions for the theorems. In particular, four conditions in Assumption 5, which are used for deriving Theorem 3 and are originally proposed in this study, are so strict that it would be difficult for them to be satisfied in real situations. Hence, it is important to construct a numerical method for addressing the optimal maintenance problems without Assumption 5.

Acknowledgements

This work was supported by JSPS KAKENHI Grant Number JP20K04989.

Conflicts of Interest

The authors declare no conflicts of interest regarding the publication of this paper.

References

[1] Osaki, S. (2001) Stochastic Models in Reliability and Maintenance. Springer, Berlin.
https://doi.org/10.1007/978-3-540-24808-8
[2] Cho, D.I. and Parlar, M. (1991) A Survey of Maintenance Models for Multi-Unit Systems. European Journal of Operational Research, 51, 1-23.
https://doi.org/10.1016/0377-2217(91)90141-H
[3] Dekker, R. (1996) Applications of Maintenance Optimization Models: A Review and Analysis. Reliability Engineering and System Safety, 51, 229-240.
https://doi.org/10.1016/0951-8320(95)00076-3
[4] Wang, H. (2002) A Survey of Maintenance Policies of Deteriorating Systems European. Journal of Operational Research, 139, 469-489.
https://doi.org/10.1016/S0377-2217(01)00197-7
[5] Jardine, A., Lin, D. and Banjevic, D. (2006) A Review on Machinery Diagnostics and Prognostics Implementing Condition-Based Maintenance. Mechanical Systems and Signal Processing, 20, 1483-1510.
https://doi.org/10.1016/j.ymssp.2005.09.012
[6] Olde Keizer, M.C.A., Flapper, S.D.P. and Teunter, R.H. (2017) Condition-Based Maintenance Policies for Systems with Multiple Dependent Components: A Review. European Journal of Operational Research, 261, 405-420.
https://doi.org/10.1016/j.ejor.2017.02.044
[7] Alaswad, S. and Xiang, Y. (2017) A Review on Condition-Based Maintenance Optimization Models for Stochastically Deteriorating System. Reliability Engineering and System Safety, 157, 54-63.
https://doi.org/10.1016/j.ress.2016.08.009
[8] Derman, C. (1963) On Optimal Replacement Rules When Changes of States Are Markovian. In: Bellman, R., Ed., Mathematical Optimization Techniques, University of California Press, Berkley, 201-210.
https://doi.org/10.1525/9780520319875-011
[9] Douer, N. and Yechiali, U. (1994) Optimal Repair and Replacement in Markovian Systems. Communications in Statistics. Stochastic Models, 10, 253-270.
https://doi.org/10.1080/15326349408807295
[10] Hopp, W. and Wu, A. (1990) Machine Maintenance with Multiple Maintenance Actions. IIE Transactions, 22, 226-233.
https://doi.org/10.1080/07408179008964177
[11] Kurt, M. and Kharoufeh, J.P. (2010) Monotone Optimal Replacement Policies for a Markovian Deteriorating System in Controllable Environment. Operations Research Letters, 38, 273-279.
https://doi.org/10.1016/j.orl.2010.03.001
[12] Kurt, M. and Kharoufeh, J.P. (2010) Optimally Maintaining a Markovian Deteriorating System with Limited Imperfect Repairs. European Journal of Operational Research, 205, 368-380.
https://doi.org/10.1016/j.ejor.2010.01.009
[13] Tamura, N. (2011) An Optimal Inspection Policy for a Markovian Deteriorating System with Stochastically Decreasing Lifetime. International Journal of Reliability Quality and Operations Management, 2, 13-28.
[14] Lam, C.T. and Yeh, R.H. (1994) Optimal Maintenance-Policies for Deteriorating Systems under Various Maintenance Strategies. IEEE Transactions on Reliability, 43, 423-430.
https://doi.org/10.1109/24.326439
[15] Lam, C.T. and Yeh, R.H. (1994) Optimal Replacement Policies for Multistate Deteriorating Systems. Naval Research Logistics, 41, 303-315.
https://doi.org/10.1002/1520-6750(199404)41:3<303::AID-NAV3220410302>3.0.CO;2-2
[16] Lam, C.T. and Yeh, R.H. (1994) Comparison of Sequential and Continuous Inspection Strategies for Deteriorating Systems. Advances in Applied Probability, 26, 423-435.
https://doi.org/10.2307/1427444
[17] Yeh, R.H. (1996) Optimal Inspection and Replacement Policies for Multi-State Deteriorating Systems. European Journal of Operational Research, 96, 248-259.
https://doi.org/10.1016/S0377-2217(96)00077-X
[18] Yeh, R.H. (1997) State-Age-Dependent Maintenance Policies for Deteriorating Systems with Erlang Sojourn Time Distributions. Reliability Engineering and System Safety, 58, 55-60.
https://doi.org/10.1016/S0951-8320(97)00049-5
[19] Ross, S.M. (1971) Quality Control under Markovian Deterioration. Management Science, 17, 587-596.
https://doi.org/10.1287/mnsc.17.9.587
[20] Ohnishi, M., Kawai, H. and Mine, H. (1986) An Optimal Inspection and Replacement Policy under Incomplete State Information. European Journal of Operational Research, 27, 117-128.
https://doi.org/10.1016/S0377-2217(86)80014-5
[21] Grosfeld-Nir, A. (1996) A Two-State Partially Observable Markov Decision Process with Uniformly Distributed Observation. Operations Research, 44, 458-463.
https://doi.org/10.1287/opre.44.3.458
[22] Grosfeld-Nir, A. (2007) Control Limits for Two-State Partially Observable Markov Decision Processes. European Journal of Operational Research, 182, 300-304.
https://doi.org/10.1016/j.ejor.2006.08.033
[23] Chen, M., Fan, H., Hu, C. and Zhou, D. (2014) Maintaining Partially Observed Systems with Imperfect Observation and Resource Constraint. IEEE Transactions on Reliability, 63, 881-890.
https://doi.org/10.1109/TR.2014.2337792
[24] Srinivasan, R. and Parlikad, A.K. (2014) Semi-Markov Decision Process with Partial Information for Maintenance Decisions. IEEE Transactions on Reliability, 63, 891-898.
https://doi.org/10.1109/TR.2014.2338811
[25] Oosterom, C., Peng, H. and Houtm, G. (2017) Maintenance Optimization for a Markovian Deteriorating System with Population Heterogeneity. IISE Transactions, 49, 96-109.
https://doi.org/10.1080/0740817X.2016.1205239
[26] Kivanc, I., Ozgur-Unluakin, D. and Bilgic, T. (2022) Maintenance Policy Analysis of the Regenerative Air Heater System Using Factored POMDPs. Reliability Engineering and System Safety, 219, Article ID: 108195.
https://doi.org/10.1016/j.ress.2021.108195
[27] Ohnishi, M., Morioka, T. and Ibaraki, T. (1994) Optimal Minimal-Repair and Replacement Problem of Discrete-Time Markovian Deterioration System under Incomplete State Information. Computers and Industrial Engineering, 27, 409-412.
https://doi.org/10.1016/0360-8352(94)90321-2
[28] Pham, H. and Wang, H. (1996) Imperfect Maintenance. European Journal of Operational Research, 94, 425-438.
https://doi.org/10.1016/S0377-2217(96)00099-9
[29] Ivy, J.S. and Pollock, S.M. (2005) Marginally Monotonic Maintenance Policies for a Multi-State Deteriorating Machine with Probabilistic Monitoring, and Silent Failures. IEEE Transactions on Reliability, 54, 489-497.
https://doi.org/10.1109/TR.2005.853443
[30] Tamura, N., Yuge, T. and Yanagi, S. (2011) Monotone Properties of Optimal Maintenance Policy for Two-State Partially Observable Markov Decision Process Model with Multiple Observations. International Journal of Operations Research, 7, 23-34.
[31] Fan, H., Xu, Z. and Chen, S.W. (2015) Optimally Maintaining a Multi-State System with Limited Imperfect Preventive Repairs. International Journal of Systems Science, 46, 1729-1740.
https://doi.org/10.1080/00207721.2013.828799
[32] Kijima, M. (1997) Markov Processes for Stochastic Modeling. Chapman & Hall, New York.
https://doi.org/10.1007/978-1-4899-3132-0
[33] Shaked, M. and Shanthikumar, J.G. (2007) Stochastic Orders. Springer, New York.
https://doi.org/10.1007/978-0-387-34675-5
[34] Monahan, G. (1982) A Survey of Partially Observable Markov Decision Processes: Theory, Models, and Algorithms. Management Science, 28, 1-111.
https://doi.org/10.1287/mnsc.28.1.1
[35] Guo, C. and Liang, Z. (2022) A Predictive Markov Decision Process for Optimizing Inspection and Maintenance Strategies of Partially Observable Multi-State Systems. Reliability Engineering and System Safety, 226, Article ID: 108683.
https://doi.org/10.1016/j.ress.2022.108683

Copyright © 2024 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.