^{1}

^{*}

^{2}

The main purpose of this paper is to obtain the inference of parameters of heterogeneous population represented by finite mixture of two Pareto (MTP) distributions of the second kind. The constant-partially accelerated life tests are applied based on progressively type-II censored samples. The maximum likelihood estimates (MLEs) for the considered parameters are obtained by solving the likelihood equations of the model parameters numerically. The Bayes estimators are obtained by using Markov chain Monte Carlo algorithm under the balanced squared error loss function. Based on Monte Carlo simulation, Bayes estimators are compared with their corresponding maximum likelihood estimators. The two-sample prediction technique is considered to derive Bayesian prediction bounds for future order statistics based on progressively type-II censored informative samples obtained from constant-partially accelerated life testing models. The informative and future samples are assumed to be obtained from the same population. The coverage probabilities and the average interval lengths of the confidence intervals are computed via a Monte Carlo simulation to investigate the procedure of the prediction intervals. Analysis of a simulated data set has also been presented for illustrative purposes. Finally, comparisons are made between Bayesian and maximum likelihood estimators via a Monte Carlo simulation study.

Accelerated life tests (ALTs) are used to obtain information quickly on the lifetime distribution of materials or products. The test units are run at higher-than- usual levels of stress to induce early failures. A model relating life length to stress is fitted to the accelerated failure times and then extrapolated to estimate the failure time distribution under the normal use condition. ALTs are preferred to be used in manufacturing industries to obtain enough failure data, in a short period of time, necessary to make inferences regarding its relationship with external stress variables.

According to [

The third is the step-stress ALT, in which the test condition changes at a given time or upon the occurrence of a specified number of failures, has been studied by several authors. [

When the acceleration factor cannot be assumed as a known value, the partially accelerated life test (PALT) will be a good choice to perform the life test. In ALTs, the units are tested only at accelerated conditions (see [

From the Bayesian viewpoint, few studies have been considered on PALT such as [

Pareto distribution of the second type (also known as the Lomax distribution) has been widely used in economic studies and to analyze business failure data. The Pareto distribution has been studied by several authors. According to [

[

A random variable T is said to have a Mixture of two Pareto distributions (MTPD) if its probability density function (PDF) is given by

f 1 Θ ( t ) = p 1 f 11 ( t , θ 1 ) + p 2 f 12 ( t , θ 2 ) , (1)

where Θ = ( θ 1 , θ 2 , p 1 , p 2 ) and for j = 1 , 2 ,

θ j = ( α j , β j ) ,

f 1 j ( t ; θ j ) = α j β j α j ( β j + t ) − ( α j + 1 ) , t > 0 , ( α j , β j > 0 ) , 0 ≤ p j ≤ 1 , p 1 + p 2 = 1. (2)

Also, the cumulative distribution function (CDF), the reliability function (RF) and the hazard rate function (HRF) take the forms.

F 1 j ( t ; θ j ) = 1 − β j α j ( β j + t ) − α j , (3)

R 1 j ( t ; θ j ) = β j α j ( β j + t ) − α j , (4)

H 1 j ( t ; θ j ) = α j ( β j + t ) − 1 , (5)

where H 1 j ( ⋅ ) = f 1 j ( ⋅ ) R 1 j ( ⋅ ) . (2) is a special form of Pearson type VI distribution. In

life-testing and reliability studies, the experimenter may not always obtain complete information on failure times for all experimental units. Data obtained from such experiments are called censored data. Saving the total time on test and the cost associated with it are some of the major reasons for censoring. A censoring scheme, which can balance between total time spent for the experiment, number of units used in the experiment and the efficiency of statistical inference based on the results of the experiment, is desirable. The most common censoring schemes are Type-I (time) censoring, and Type-II (item) censoring. The conventional Type-I and Type-II censoring schemes do not have the exorability of allowing removal of units at points other than the terminal point of the experiment. Because of that, a more general censoring scheme called progressive Type- II right censoring has been used in this article. Censored data are of progressively Type II right type when they are censored by the removal of a prospected number of survivors whenever an individual fails; this continues until a fixed number of failures has occurred, at which stage the remainder of the surviving individuals are also removed or censored. This scheme includes ordinary Type II censoring and complete scheme as special cases. A general account of theoretical developments and applications concerning progressive censoring is given in the book by [

An important problem that may face the experimenter in life testing experiments is the prediction of unknown observations that belong to a future sample, based on the current available sample, known in the literature as the informative sample. For example, the experimenters or the manufacturers would like to have the bounds for the life of their products so that their warranty limits could be plausibly set and customers purchasing manufactured products would like to know the bounds for the life of the product to be purchased. For different application areas, the reader can see [

The novelty of this paper is to consider the constant PALT applied to items whose life-times under design condition are assumed to follow MTPD under a progressive Type-II censoring and the main aim is to obtain the Bayes estimators (BEs) and prediction of the acceleration factor and the parameters under consi- deration using the method of MCMC. The rest of this paper is organized as follows. In Section 2, a description of the model is presented and the MLEs of the parameters are derived. In Section 3, Bayes estimates are obtained using the balanced square error loss (BSEL) function. Bayesian two-sample prediction is presented in Section 4. Monte Carlo simulation results are presented in Section 5. Finally, some concluding remarks are introduced in Section 6.

In a constant-PALT, n 1 items randomly chosen among n test items sampled are allocated to use condition and n 2 = n − n 1 remaining items are subjected to an accelerated condition progressive type-II censoring is performed as follows.

At the time of the first failure t s 1 : m s : n s R s , R s 1 items are randomly withdrawn from the remaining n s − 1 surviving items. At the second failure t s 2 : m s : n s R s , R s 2 items from the remaining n s − 2 − R s 1 items are randomly withdrawn. The test continues until the m s − th failure t s m s : m s : n s R s at which time, all remaining

R s m s = n s − m s − ∑ υ = 1 m s − 1 R s υ items are withdraws for s = 1 , 2 . In our study, R s i

are fixed prior and m s < n s .

If the failure times of the n s items originally in the test are from a continuous population with distribution function F j ( x ) and probability density function f j ( x ) , the joint probability density function for t s 1 : m s : n s R s < t s 2 : m s : n s R s < ⋯ < t s m s : m s : n s R s and s = 1 , 2 is given by

L ( θ ; t ) = ∏ s = 1 2 { A s ∏ i = 1 m s f s Θ ( t s i : m s : n s ) [ R s Θ ( t s i : m s : n s ) ] R s i } , (6)

where t = ( t 1 , t 2 ) and, for s = 1 , 2 , t s = ( t s 1 , ⋯ , t s m s ) and

A s = n s ( n s − 1 − R s 1 ) ( n s − 2 − R s 1 − R s 2 ) ⋯ ( n s − m s + 1 − R s 1 − R s 2 ⋯ − R s ( m s − 1 ) ) .

It is clear from (6) that the constant PALTs progressively Type-II censored scheme containing the following censoring schemes as special cases:

1) Type-II censored scheme when R = { 0 , 0 , ⋯ , 0 , n s − m s } .

2) The complete sample case when R = { 0 , 0 , ⋯ , 0 } and n s = m s .

1) The lifetimes T 1 i ≡ T , i = 1 , ⋯ , n 1 of items allocated to use condition, are independent and identically distributed random variables (i.i.d. r.v.’s) and follows a mixture of MTP distribution with PDF, given in (1).

2) The lifetimes T 2 i ≡ X , i = 1 , ⋯ , n 2 of items allocated to accelerated condition, are i.i.d r.v.’s.

3) The PDF, RF, CDF and HRF of an item tested at accelerated condition are given, respectively, by

f 2 Θ ( x ) = p 1 f 21 ( x ; θ 1 ) + p 2 f 22 ( x ; θ 2 ) , R 2 Θ ( x ) = p 1 R 21 ( x ; θ 1 ) + p 2 R 22 ( x ; θ 2 ) , F 2 Θ ( x ) = p 1 F 21 ( x ; θ 1 ) + p 2 F 22 ( x ; θ 2 ) , H 2 Θ ( x ) = f 2 Θ ( x ) R 2 Θ ( x ) , } , (7)

where for j = 1 , 2 , θ j = ( α j , β j , λ j ) , and

H 2 j ( x ; θ j ) = λ j H 1 j ( x , θ j ) = λ j α j ( β j + x ) − 1 , (8)

R 2 j ( x ; θ j ) = β j λ j α j ( β j + x ) − λ j α j , (9)

F 2 j ( x ; θ j ) = 1 − β j λ j α j ( β j + x ) − λ j α j , (10)

f 2 j ( x ; θ j ) = α j λ j β j λ j α j ( β j + x ) − ( λ j α j + 1 ) , (11)

where λ j is an accelerated factor satisfying λ j > 1 .

4) The i.i.d lifetimes T 1 i and T 2 i , i = 1 , 2 , ⋯ , n j are mutually statistically- independent.

Let, for s = 1 , 2 , T s 1 : m s : n s ( R s 1 , ⋯ , R s m s ) < T s 2 : m s : n s ( R s 1 , ⋯ , R s m s ) < ⋯ < T s m s : m s : n s ( R s 1 , … , R s m s ) denote two progressively type-II censored samples from two populations whose PDFs are as given by (1) and (2), respectively, with R s = ( R s 1 , ⋯ , R s m s ) being the two progressive censoring schemes. We denote also the observed values by, t s 1 : m s : n s < t s 2 : m s : n s < ⋯ < t s m s : m s : n s . The log-likelihood function l ( α , β , λ | x ¯ ) = log L ( α , β , λ | x ¯ ) without normalized constant is then given by

l ≡ ln L ( θ ; t ) = ∑ s = 1 2 ln A s + ∑ s = 1 2 ∑ i = 1 m s ln f s Θ ( t s i : m s : n s ) + ∑ s = 1 2 ∑ i = 1 m s R s i ln R s Θ ( t s i : m s : n s ) . (12)

Assuming that the parameters p , λ j and β j are unknown and α j , is known, the likelihood equations are given, for j = 1 , 2 , by

∂ l ∂ p j = ∑ s = 1 2 ∑ i = 1 m s ψ s ( t s i ) + ∑ s = 1 2 ∑ i = 1 m s R s i ψ s * ( t s i ) = 0 , ∂ l ∂ λ j = ∑ i = 1 m 2 p j ξ j ( t 2 i ) f 2 Θ ( t 2 i : m 2 : n 2 ) + ∑ i = 1 m 2 R 2 i p j ξ j * ( t 2 i ) R 2 Θ ( t 2 i : m 2 : n 2 ) = 0 , j = 1 , 2 ∂ l ∂ β j = ∑ s = 1 2 ∑ i = 1 m s p j ϑ s j ( t s i ) f s Θ ( t s i : m s : n s ) + ∑ s = 1 2 ∑ i = 1 m s R s i p j ϑ s j * ( t s i ) R s Θ ( t s i : m s : n s ) = 0 , j = 1 , 2 } (13)

where, for j = 1 , 2

ψ s ( t s i ) = f s 1 ( t s i ; θ 1 ) − f s 2 ( t s i ; θ 2 ) f s Θ ( t s i : m s : n s ) , ψ s * ( t s i ) = R s 1 ( t s i ; θ 1 ) − R s 2 ( t s i ; θ 2 ) R s Θ ( t s i : m s : n s ) , ξ j ( t 2 i ) = ∂ f 2 j ( t 2 i ; θ j ) ∂ λ j = α j 2 λ j β j λ j α j ( β j + t 2 i ) − ( λ j α j + 1 ) [ ln β j β j + t 2 i + 1 λ j α j ] ξ j * ( t 2 i ) = ∂ R 2 j ( t 2 i ; θ j ) ∂ λ j = α j β j λ j α j ( β j + t 2 i ) − λ j α j [ ln β j β j + t 2 i ] , ϑ 1 j ( t 1 i ) = ∂ f 1 j ( t 1 i ; θ j ) ∂ β j = α j β j α j − 1 ( β j + t 1 i ) − ( α j + 2 ) ( α j t 1 i − β j ) , ϑ 2 j ( t 2 i ) = ∂ f 2 j ( t 2 i ; θ j ) ∂ β j = α j λ j β j α j λ j − 1 ( β j + t 2 i ) − ( λ j α j + 2 ) ( α j λ j t 2 i − β j ) , ϑ 1 j * ( t 1 i ) = ∂ R 1 j ( t 1 i ; θ j ) ∂ β j = α j β j α j − 1 t 1 i ( β j + t 1 i ) − ( α j + 1 ) , ϑ 2 j * ( t 2 i ) = ∂ R 2 j ( t 2 i ; θ j ) ∂ β j = α j λ j β j α j λ j − 1 t 2 i ( β j + t 2 i ) − ( λ j α j + 1 ) , } . (14)

Equations (13) do not yield explicit solutions for p , λ j and β j , j = 1 , 2 , and have to be solved numerically to obtain the ML estimates of the five parameters, Newton-Raphson iteration is employed to solve (13).

For Bayesian approach, in order to select a single value as representing our “best” estimators of the unknown parameter, a loss function must be specified. A wide variety of loss functions have been developed in the literature to describe various types of loss structures. The balanced loss function which is introduced [

L Φ , Ω , δ o ( Ψ ( θ ) , δ ) = Ω ϒ ( θ ) Φ ( δ o , δ ) + ( 1 − Ω ) ϒ ( θ ) Φ ( Ψ ( θ ) , δ ) , (15)

where ϒ ( ⋅ ) is a suitable positive weight function and Φ ( Ψ ( θ ) , δ ) is an arbitrary loss function when estimating Ψ ( θ ) by δ . The parameter δ o is a chosen prior estimator of Ψ ( θ ) , obtained for instance from the criterion of ML, least squares or unbiasedness among others. They give a general Bayesian connection between the case of Ω > 0 and Ω = 0 where 0 ≤ Ω < 1 .

This section deals with studying the Bayes estimates of the parameters under consideration using the balanced square error loss (BSEL) function using the non-informative prior NIP distribution. It follows that a NIP for the acceleration factor λ j is given by

π 1 ( λ j ) ∝ 1 λ j , ( λ j > 1 ) . (16)

Also, the NIP’s for the scale parameter β j and the parameter p j are, respectively, as

π 2 ( β j ) ∝ 1 β j , ( β j > 0 ) , (17)

π 3 ( p j ) ∝ 1 p j , ( p j > 0 ) . (18)

Therefore, the joint NIP of the three parameters can be expressed by

π ( Θ ) = π 1 ( λ j ) π 2 ( β j ) π 3 ( p j ) ∝ 1 p j λ j β j , ( λ j > 1 , β j , p j > 0 ) , (19)

where Θ = ( p j , λ j , β j ) .

It is to be noted that our objective is to consider vague priors so that the priors do not have any significant roles in the analyses that follow. However, if one uses the prior beliefs different from (19) and resorts to sample based approaches for analyzing the posterior, one may use the concept of sampling-importance-re- sampling without working afresh with the new prior-likelihood setup (see, [

The symmetric square-error loss (SE) is one of the most popular loss functions. By choosing Φ ( Ψ ( θ ) , δ ) = ( δ − Ψ ( θ ) ) 2 and ϒ ( θ ) = 1 , in (15), the balanced loss function reduced to the BSEL function, used by [

L Ω , δ o ( Ψ ( θ ) , δ ) = Ω ( δ − δ o ) 2 + ( 1 − Ω ) ( δ − Ψ ( θ ) ) 2 , (20)

and the corresponding Bayes estimate of the function Ψ ( θ ) is given by

δ Ω , Ψ , δ o ( t ) = Ω δ o + ( 1 − Ω ) E ( Ψ ( θ ) | t ) . (21)

Under the BSEL function, the estimator of a parameter (or a given function of the parameters) is the posterior mean. Thus, Bayes estimators of the parameters are obtained by using the loss function (20). The Bayes estimators of a function u ≡ u ( p j , λ j , β j ) = p j , λ j or β j is given by

u ^ B S = Ω u ^ M L + ( 1 − Ω ) ∫ 0 ∞ u π * ( p j , λ j , β j | t ) d Θ , (22)

where, u ^ M L is the ML estimate of u . It is not possible to compute (22) analytically, therefore, we propose to approximate (22) by using MCMC technique to generate samples from the posterior distributions and then compute the Bayes estimators of the individual parameters.

The MCMC method is a useful technique for computing Bayes estimates of the function u ≡ u ( p j , λ j , β j ) . A wide variety of MCMC schemes are available, and it can be difficult to choose among them. An important sub-class of MCMC methods is Gibbs sampling and more general Metropolis within- Gibbs samplers. The advantage of using the MCMC method over the MLE method is that we can always obtain a reasonable interval estimate of the parameters by constructing the probability intervals based on the empirical posterior distribution. This is often unavailable in maximum likelihood estimation. Indeed, the MCMC samples may be used to completely summarize the posterior uncertainty about the parameters p j , λ j and β j , through a kernel estimate of the posterior distribution. This is also true of any function of the parameters. For more detailes about the MCMC methods see, for example, [

The Metropolis-Hasting algorithm generates sampling from an (essentially) arbitrary proposal distribution (i.e. a Markov transition kernel). From the product of Equations (19) and (6), the joint posterior density function of p j , λ j and β j given the data can be written as

π * ( p j , λ j , β j | t ) = B 1 ( p j λ j β j ) − 1 ∏ s = 1 2 { A s ∏ i = 1 m s f s Θ ( t s i : m s : n s ) [ R s Θ ( t s i : m s : n s ) ] R s i } , (23)

where

B 1 − 1 = ∫ Θ π ( Θ ) L ( p j , λ j , β j ) dΘ .

t = ( t i 1 , t i 2 , ⋯ , t i m i ) . The conditional posterior distribution of the parameters p j , λ j and β j can be computed and written, respectively, by

π * ( p j | λ j , β j , t ) ∝ p j − 1 ∏ s = 1 2 { A s ∏ i = 1 m s f s Θ ( t s i : m s : n s ) [ R s Θ ( t s i : m s : n s ) ] R s i } , (24)

π * ( λ j | p j , β j , t ) ∝ λ j − 1 ∏ s = 1 2 { A s ∏ i = 1 m s f s Θ ( t s i : m s : n s ) [ R s Θ ( t s i : m s : n s ) ] R s i } , (25)

π * ( β j | p j , λ j , t ) ∝ β j − 1 ∏ s = 1 2 { A s ∏ i = 1 m s f s Θ ( t s i : m s : n s ) [ R s Θ ( t s i : m s : n s ) ] R s i } . (26)

The posterior of p j , λ j and β j in (24), (25) and (26) is not known, but the plot of it shows that it is similar to normal distribution. Therefore to generate from this distribution, we use the Metropolis {Hastings method ( [

1) Start with initial guess of ( p j , λ j , β j ) say ( p j 0 , λ j 0 , β j 0 ) , respectively.

2) Set i = 1 .

3) Generate p i from (24) and λ i from (25).

4) Generate β i from (26).

5) Set i = i + 1.

6) Repeat steps 3 - 5 N times.

7) An approximate Bayes estimator of u under BSEL function is given by

E ( u | t ) = ( 1 / ( N − ν ) ) ∑ i = ν + 1 N u ( p i , λ i , β i ) , (27)

where ν is the burn-in period. So that, the Bayes estimators of u based on BSEL function is given by

u ^ B S = Ω u ^ M L + ( 1 − Ω ) E ( u | t ) . (28)

The two-sample prediction technique is considered to derive Bayesian prediction bounds for future order statistics based on progressively Type-II censored informative samples obtained from constant-PALT models. The coverage probabilities and the average interval lengths of the confidence intervals are computed via a Monte Carlo simulation to investigate the procedure of the prediction intervals. Suppose that, for S = 1 , 2 , the two sample scheme is used in which the informative sample ( T s 1 : m s : n s < T s 2 : m s : n s < ⋯ < T s m s : m s : n s ) re- presents an observed informative progressively type-II right censored sample of size m s obtained from a sample of size n s with progressive CS R s = ( R s 1 , ⋯ , R s m s ) drawn from a population whose PDFs are as given by (1) and (7). Suppose also that Y 1 : M : N , Y 2 : M : N , ⋯ , Y M : M : N represents a future (unobserved) independent progressively type-II right censored sample of size M obtained from a sample of size N with progressive CS R * = ( R 1 * , ⋯ , R M * ) , drawn from the population whose CDF is (9). We want to predict any future (unobserved) Y b , b = 1 , 2 , ⋯ , M , in the future sample of size M . The PDF of Y b , b = 1 , 2 , ⋯ , M , given the vector of parameters θ , is obtained as (see [

g * ( y b | θ ) = = C b − 1 f 2 Θ ( y b ) ∑ i = 1 b κ i [ 1 − F 2 Θ ( y b ) ] γ i − 1 , (29)

where

γ i = ∑ j = i M ( R j * + 1 ) = N − ∑ j = i i − 1 ( R j * + 1 ) , C b − 1 = ∏ i = 1 b γ i , κ i = ∏ j = 1 b 1 γ j − γ i , ∀ i ≠ j , b > 1 , and κ 1 = 1 for b = 1.

Substituting from (7) and (9) in (29), we have:

g * ( y b | θ ) = C b − 1 ( p 1 f 21 ( y ; θ 1 ) + p 2 f 22 ( y ; θ 2 ) ) ∑ i = 1 b κ i [ 1 − ( p 1 F 21 ( y ; θ 1 ) + p 2 F 22 ( y ; θ 2 ) ) ] γ i − 1 . (30)

Maximum likelihood prediction (MLP) can be obtained using (30) by replacing the parameters θ = ( p , β 1 , β 2 , λ 1 , λ 2 ) by θ ^ ( M L ) = ( p ^ ( M L ) , β 1 ^ ( M L ) , β 2 ^ ( M L ) , λ 1 ^ ( M L ) , λ 2 ^ ( M L ) ) .

1) Interval prediction:

The maximum likelihood prediction interval (MLPI) for any future observation y b , 1 ≤ b ≤ M can be obtained by

Pr [ y b ≥ υ | t ] = ∫ υ ∞ g * ( y b | θ ^ ( M L ) ) d y b . (31)

A ( 1 − τ ) × 100 % MLPI ( L , U ) of the future observation y b is given by solving the following two nonlinear equations

Pr [ y b ≥ L ( t ) | t ] = 1 − τ 2 , Pr [ y b ≥ U ( t ) t | ] = τ 2 . (32)

2) Point prediction:

The maximum likelihood prediction point (MLPP) for any future observation y b can be obtained by replacing the parameters θ = ( p , β 1 , β 2 , λ 1 , λ 2 ) by θ ^ ( M L ) = ( p ^ ( M L ) , β 1 ^ ( M L ) , β 2 ^ ( M L ) , λ 1 ^ ( M L ) , λ 2 ^ ( M L ) ) .

y ^ b ( M L ) = E [ y b | t ] = ∫ 0 ∞ y b g * ( y b | θ ^ ( M L ) ) d y b . (33)

The predictive density function of Y b , 1 ≤ b ≤ M is given by:

Ψ * ( y b | t ) = ∫ 0 ∞ g * ( y b | θ ) π * ( θ | t ) d θ , y b > 0 , (34)

1) Interval prediction:

Bayesian prediction interval (BPI), for the future observation Y b , 1 ≤ b ≤ M , can be computed using (34) which can be approximated using MCMC algorithm by the form

Ψ ⋆ ( y b | t ) = ∑ i = 1 μ g * ( y b | θ i ) ∑ i = 1 μ ∫ 0 ∞ g * ( y b | θ i ) d y b , (35)

where θ i , i − 1 , 2 , ⋯ , μ are generated from the posterior density function (23) using Gibbs sampler and Metropolis-Hastings techniques.

A ( 1 − τ ) × 100 % BPI ( L , U ) of the future observation y b is obtained by solving the following two nonlinear equations

∑ i = 1 μ ∫ L ∞ g * ( y b | θ i ) d y b ∑ i = 1 μ ∫ 0 ∞ g * ( y b | θ i ) d y b = 1 − τ 2 , (36)

∑ i = 1 μ ∫ U ∞ g * ( y b | θ i ) d y b ∑ i = 1 μ ∫ 0 ∞ g * ( y b | θ i ) d y b = τ 2 . (37)

Numerical methods such as Newton-Raphson are necessary to solve the above two nonlinear Equations (36) and (37), to obtain L and U for a given.

2) Point prediction:

a) Bayesian prediction point (BPP) for the future observation y b based on BSEL function can be obtained using

y ˜ b ( B S ) = Ω y ^ b ( M L ) + ( 1 − Ω ) E ( y b | t ) , (38)

where y ^ b ( M L ) is the ML prediction for the future observation y b which can be obtained using (36) and E ( y b | t ) can be obtained using

E ( y b | t ) = ∫ 0 ∞ y b Ψ * ( y b | t ) d y b . (39)

b) BPP for the future observation y b based on BLINX loss function can be obtained using

y ^ b ( B L ) = − 1 a ln [ Ω exp [ − a y ^ b ( M L ) ] + ( 1 − Ω ) E ( e − a y b | t ) ] , (40)

where y ^ b ( M L ) is the ML prediction for the future observation y b which can be obtained using (36) and E ( e − a y b | t ) can be obtained using

E ( e − a y b | t ) = ∫ 0 ∞ e − a y b Ψ * ( y b | t ) d y b . (41)

In this subsection, numerical examples are provided to demonstrate the theoretical results given in this paper. All computations were performed using (MA- THEMATICA ver. 8.0).

To generate progressively type-II censored Pareto samples, we used the algorithm proposed by [

1) For given values of the parameters, n s and m s ( 1 ≤ m s ≤ n s ) , s = 1 , 2 we generate type II progressively samples from the MTP distribution as follows:

a) For given values of m s , we generate two independent random samples of sizes m_{1} and m 2 from Uniform (0,1) distribution ( U s 1 , U s 2 , ⋯ , U s m s ) , s = 1 , 2.

b) For given values of the progressive censoring scheme

R s i , s = 1 , 2 , i = 1 , ⋯ , m s , we set E s i = 1 / ( i + ∑ κ = m s − i + 1 m s R s κ ) where

s = 1 , 2 , i = 1 , ⋯ , m s .

c) Set V s i = U s i E s i .

d) Set U s i * = 1 − ∏ κ = m s − i + 1 m s V s κ , s = 1 , 2 , i = 1 , ⋯ , m s .

e) For given values of p , α j , β j , λ j and n s , m s , set:

U s i * = p [ 1 − β 1 λ 1 ( s − 1 ) α 1 ( β 1 + t s i ) − λ 1 ( s − 1 ) α 1 ] + ( 1 − p ) [ 1 − β 2 λ 2 ( s − 1 ) α 2 ( β 2 + t s i ) − λ 2 ( s − 1 ) α 2 ] ,

which is the required progressive Type II censored samples of sizes m s from MTP distribution under constant PALT.

2) The MLEs of the parameters are obtained by solving the nonlinear equations (13) numerically.

3) Based on BSEL loss function the Bayes estimates of the parameters are computed, from (28) according to the above MCMC method.

Simulation studies have been performed using (Mathematica ver. 8.0) for illustrating the theoretical results of estimation problem. The performance of the resulting estimators of the acceleration, shape and scale parameters has been considered in terms of their average (AVG), relative absolute bias (RAB) and mean square error (MSE), where

Φ ^ ¯ k = ( 1 / M ) ∑ i = 1 M Φ ^ k ( i ) k = 1 , 2 , ⋯ , 5 , ( Φ 1 = p , Φ 2 = λ 1 , Φ 3 = λ 2 , Φ 4 = β 1 , Φ 5 = β 2 ) ,

R A B = | Φ ^ ¯ k − Φ k | Φ k ,

M S E = ( 1 / M ) ∑ i = 1 M ( Φ ^ k ( i ) − Φ k ) 2 .

In our study, we have used three different censoring schemes (C.S), namely:

Scheme I: R m = n s − m s , R i = 0 for i ≠ m s .

Scheme II: R 1 = n s − m s , R i = 0 for i ≠ 1 .

Scheme III: R ( ( m s + 1 ) / 2 ) = n s − m s , R i = 0 for i ≠ ( m s + 1 ) / 2 ; if m s odd, and R ( m s / 2 ) = n s − m s , R i = 0 for i ≠ ( m s / 2 ) ; if m s even.

In simulation studies, we consider two case separately:

a) The population parameter values ( α 1 = 1.1 , α 2 = 2.3 , β 1 = 0.3 , β 2 = 0.7 , λ 1 = 1.5 , λ 2 = 2 , p = 0.5 ) , the sample sizes ( n 1 = n 2 = n ) and observed failure times ( m 1 = m 2 = m ) the results shown in

b) The population parameter values

( α 1 = 1.1 , α 2 = 2.3 , β 1 = 0.3 , β 2 = 0.7 , λ 1 = 1.5 , λ 2 = 2 , p = 0.5 ) , the sample sizes

n | m | C.S | Parameters | ML method | Bayes method | ||||
---|---|---|---|---|---|---|---|---|---|

MLE | MSE | RAB | MLE | MSE | RAB | ||||

10 | 5 | I | 0.601454 | 0.0911552 | 0.202909 | 0.460629 | 0.0232686 | 0.078743 | |

1.9737 | 0.703832 | 0.315798 | 1.80706 | 0.214357 | 0.204707 | ||||

2.12271 | 0.516762 | 0.0613551 | 1.88423 | 0.139247 | 0.0578838 | ||||

0.369676 | 0.0239581 | 0.232252 | 0.477845 | 0.0365735 | 0.592818 | ||||

0.751637 | 0.0794354 | 0.0737667 | 0.72107 | 0.0199059 | 0.0300995 | ||||

II | 0.700854 | 0.0977859 | 0.401708 | 0.518353 | 0.014129 | 0.0367053 | |||

1.95765 | 0.708308 | 0.305101 | 1.82122 | 0.228966 | 0.214148 | ||||

2.21685 | 0.581832 | 0.108427 | 1.95556 | 0.148177 | 0.0222217 | ||||

0.378412 | 0.0360783 | 0.261372 | 0.471209 | 0.0371772 | 0.570698 | ||||

0.777305 | 0.109645 | 0.110436 | 0.727977 | 0.0273527 | 0.0399674 | ||||

III | 0.629658 | 0.09394 | 0.259317 | 0.472783 | 0.0192282 | 0.0544344 | |||

2.06651 | 0.824214 | 0.377674 | 1.86251 | 0.257217 | 0.241674 | ||||

2.26205 | 0.60211 | 0.131027 | 1.9515 | 0.135306 | 0.0242518 | ||||

0.377877 | 0.0427896 | 0.259589 | 0.481326 | 0.0418398 | 0.604419 | ||||

0.826555 | 0.121951 | 0.180793 | 0.757457 | 0.0300898 | 0.0820817 | ||||

10 | 7 | I | 0.660453 | 0.112751 | 0.320906 | 0.506635 | 0.0283939 | 0.0132703 | |

1.85369 | 0.648571 | 0.235796 | 1.74605 | 0.18166 | 0.164036 | ||||

2.16858 | 0.744627 | 0.0842877 | 1.90422 | 0.177836 | 0.0478888 | ||||

0.210994 | 0.0137477 | 0.296687 | 0.401985 | 0.0114582 | 0.339951 | ||||

0.699248 | 0.134125 | 0.00107442 | 0.688608 | 0.0320385 | 0.016274 | ||||

II | 0.537681 | 0.100897 | 0.0753617 | 0.442848 | 0.029858 | 0.114304 | |||

2.2181 | 0.924725 | 0.478734 | 1.92421 | 0.266521 | 0.282808 | ||||

2.03363 | 0.490193 | 0.0168168 | 1.85065 | 0.149335 | 0.074674 | ||||

0.459593 | 0.0958671 | 0.531976 | 0.520935 | 0.0664047 | 0.73645 | ||||

0.773434 | 0.170571 | 0.104906 | 0.732026 | 0.0413834 | 0.0457511 | ||||

III | 0.723168 | 0.151995 | 0.446336 | 0.519121 | 0.0271232 | 0.0382419 | |||

1.85078 | 0.432858 | 0.233853 | 1.746 | 0.141774 | 0.164001 | ||||

2.41489 | 0.78414 | 0.207443 | 2.02639 | 0.151202 | 0.0131956 | ||||

0.375755 | 0.0283916 | 0.252517 | 0.473603 | 0.0359016 | 0.578677 | ||||

0.870039 | 0.0594453 | 0.242912 | 0.780168 | 0.0127516 | 0.114525 | ||||

20 | 10 | I | 0.557594 | 0.0847459 | 0.115188 | 0.430601 | 0.0254593 | 0.138799 | |

1.90065 | 0.704428 | 0.267101 | 1.75898 | 0.208271 | 0.172653 | ||||

2.10902 | 0.498771 | 0.0545082 | 1.87512 | 0.142039 | 0.0624406 | ||||

0.362324 | 0.039027 | 0.207747 | 0.474889 | 0.0411224 | 0.582964 | ||||

0.742612 | 0.101949 | 0.0608743 | 0.717371 | 0.0261471 | 0.0248159 |

20 | 10 | II | 0.669976 | 0.1211 | 0.339952 | 0.490481 | 0.0242542 | 0.0190383 | |
---|---|---|---|---|---|---|---|---|---|

1.95717 | 0.68463 | 0.30478 | 1.8027 | 0.206388 | 0.201803 | ||||

2.04906 | 0.472107 | 0.0245278 | 1.85392 | 0.140938 | 0.0730379 | ||||

0.3308 | 0.0168379 | 0.102667 | 0.454641 | 0.0286291 | 0.515468 | ||||

0.787288 | 0.187934 | 0.124698 | 0.733932 | 0.045068 | 0.048474 | ||||

III | 0.632057 | 0.100993 | 0.264113 | 0.47326 | 0.0224658 | 0.0534806 | |||

1.84788 | 0.563256 | 0.231918 | 1.73577 | 0.171077 | 0.157178 | ||||

2.02863 | 0.545136 | 0.0143147 | 1.8346 | 0.167083 | 0.0826977 | ||||

0.320854 | 0.017344 | 0.0695136 | 0.453179 | 0.0281525 | 0.510598 | ||||

0.750958 | 0.15458 | 0.0727968 | 0.719845 | 0.0389693 | 0.0283506 | ||||

20 | 15 | I | 0.550374 | 0.118694 | 0.100747 | 0.497373 | 0.0349629 | 0.00525442 | |

1.56948 | 0.229736 | 0.046317 | 1.58626 | 0.0538567 | 0.0575068 | ||||

1.81417 | 0.514309 | 0.0929162 | 1.71431 | 0.198039 | 0.142846 | ||||

0.364483 | 0.0526832 | 0.214943 | 0.477257 | 0.045402 | 0.590855 | ||||

0.739965 | 0.0820597 | 0.0570934 | 0.717426 | 0.0186836 | 0.0248949 | ||||

II | 0.548206 | 0.0601328 | 0.0964128 | 0.42714 | 0.0182459 | 0.145721 | |||

1.84334 | 0.801833 | 0.228892 | 1.69273 | 0.186019 | 0.128484 | ||||

2.52571 | 0.595692 | 0.262855 | 2.08784 | 0.0885475 | 0.043922 | ||||

0.35615 | 0.0225767 | 0.187167 | 0.474979 | 0.0349709 | 0.583262 | ||||

0.813256 | 0.0874564 | 0.161794 | 0.756249 | 0.0249578 | 0.0803555 | ||||

III | 0.549382 | 0.0478625 | 0.0987647 | 0.441121 | 0.0164724 | 0.117758 | |||

1.75354 | 0.525229 | 0.169024 | 1.69137 | 0.168108 | 0.127583 | ||||

1.79942 | 0.472088 | 0.100288 | 1.70238 | 0.207905 | 0.14881 | ||||

0.354368 | 0.0425507 | 0.181226 | 0.46136 | 0.0369997 | 0.537867 | ||||

0.670072 | 0.0371572 | 0.0427541 | 0.675311 | 0.010769 | 0.0352695 | ||||

30 | 15 | I | 0.499598 | 0.104484 | 0.000803821 | 0.397832 | 0.0335915 | 0.204337 | |

2.15827 | 0.797189 | 0.438849 | 1.89995 | 0.234899 | 0.266632 | ||||

1.87373 | 0.398902 | 0.0631326 | 1.72948 | 0.18941 | 0.135258 | ||||

0.376413 | 0.0215011 | 0.254709 | 0.48996 | 0.041951 | 0.633201 | ||||

0.779513 | 0.0489548 | 0.11359 | 0.741129 | 0.0135147 | 0.0587554 | ||||

II | 0.76084 | 0.136086 | 0.521681 | 0.540043 | 0.0191063 | 0.0800852 | |||

1.78253 | 0.423387 | 0.188354 | 1.71719 | 0.130948 | 0.144796 | ||||

1.78884 | 0.641888 | 0.105579 | 1.74993 | 0.214919 | 0.125034 | ||||

0.31688 | 0.0129736 | 0.0562653 | 0.438228 | 0.0243289 | 0.460761 | ||||

0.583023 | 0.0623007 | 0.16711 | 0.644207 | 0.0155235 | 0.079704 | ||||

III | 0.57124 | 0.0459656 | 0.14248 | 0.447413 | 0.0156696 | 0.105174 | |||

1.88525 | 0.719008 | 0.256833 | 1.75951 | 0.2079 | 0.173008 | ||||

1.705 | 0.467048 | 0.147502 | 1.68106 | 0.222155 | 0.159471 | ||||

0.308311 | 0.014189 | 0.0277027 | 0.446638 | 0.0246022 | 0.488793 | ||||

0.632415 | 0.0380157 | 0.0965499 | 0.664642 | 0.0116061 | 0.0505112 |

30 | 20 | I | 0.393203 | 0.127093 | 0.213595 | 0.345447 | 0.0551168 | 0.309106 | |
---|---|---|---|---|---|---|---|---|---|

1.59049 | 0.36839 | 0.0603245 | 1.62542 | 0.106877 | 0.0836112 | ||||

2.26603 | 0.428951 | 0.133013 | 1.96647 | 0.0839317 | 0.0167674 | ||||

0.234789 | 0.0197048 | 0.217369 | 0.411934 | 0.0212264 | 0.373113 | ||||

0.822029 | 0.08424 | 0.174326 | 0.760107 | 0.02548 | 0.0858669 | ||||

II | 0.750849 | 0.0994436 | 0.501699 | 0.527088 | 0.0112714 | 0.0541766 | |||

1.78607 | 0.344742 | 0.190715 | 1.70742 | 0.108857 | 0.13828 | ||||

2.36571 | 0.719694 | 0.182853 | 1.96996 | 0.163298 | 0.0150208 | ||||

0.384138 | 0.0274193 | 0.28046 | 0.466012 | 0.0325419 | 0.553373 | ||||

0.715106 | 0.097689 | 0.0215797 | 0.700472 | 0.0258527 | 0.000674698 | ||||

III | 0.526054 | 0.113862 | 0.0521074 | 0.412145 | 0.0375391 | 0.17571 | |||

1.70714 | 0.334107 | 0.138096 | 1.65264 | 0.107972 | 0.10176 | ||||

1.85459 | 0.535046 | 0.0727042 | 1.72149 | 0.217767 | 0.139256 | ||||

0.272605 | 0.0108596 | 0.0913181 | 0.430056 | 0.0197156 | 0.433519 | ||||

0.698068 | 0.0602391 | 0.00276056 | 0.68964 | 0.0151716 | 0.0148004 |

n | m | C.S | ||
---|---|---|---|---|

I | II | III | ||

10 | 5 | |||

10 | 7 | |||

20 | 10 | |||

20 | 15 | |||

30 | 15 | |||

30 | 20 |

( n 1 ≠ n 2 ) and observed failure times ( m 1 ≠ m 2 ) the results shown in

The ML prediction (point and interval) and Bayesian prediction (point and interval) are computed according to the following steps:

Generate θ i = ( p i , β 1 i , β 2 i , λ 1 i , λ 2 i ) , from the posterior PDF using MCMC algorithm.

Solving Equation (32) we get the 95% MLPI for the b th order statistics in a future progressively Type-II censored sample also the MLPP for the future observation y b ,is computed using (33).

C.S | Parameters | ML method | Bayes method | ||||||
---|---|---|---|---|---|---|---|---|---|

MLE | MSE | RAB | MLE | MSE | RAB | ||||

15 20 | 9 12 | I | 0.582126 | 0.0667111 | 0.164253 | 0.43802 | 0.0196126 | 0.123961 | |

2.14351 | 1.15743 | 0.429006 | 1.83838 | 0.291516 | 0.225585 | ||||

1.92079 | 0.366342 | 0.039605 | 1.76003 | 0.142649 | 0.119987 | ||||

0.405359 | 0.0605374 | 0.351195 | 0.493668 | 0.0497705 | 0.645561 | ||||

0.64733 | 0.0564176 | 0.0752435 | 0.666565 | 0.0150668 | 0.0477636 | ||||

II | 0.690734 | 0.0901268 | 0.381469 | 0.508892 | 0.0109952 | 0.0177844 | |||

1.76961 | 0.509727 | 0.17974 | 1.70937 | 0.142868 | 0.139577 | ||||

1.873 | 0.64641 | 0.0635004 | 1.73129 | 0.257858 | 0.134355 | ||||

0.306905 | 0.0121153 | 0.0230154 | 0.448829 | 0.0259015 | 0.496097 | ||||

0.679614 | 0.137545 | 0.0291222 | 0.678015 | 0.0336289 | 0.0314067 | ||||

III | 0.701195 | 0.114975 | 0.40239 | 0.492905 | 0.0156311 | 0.0141904 | |||

1.87453 | 0.434023 | 0.249685 | 1.72842 | 0.129539 | 0.152283 | ||||

2.44222 | 0.834579 | 0.221111 | 2.05112 | 0.155312 | 0.0255614 | ||||

0.298962 | 0.0067289 | 0.00345946 | 0.430867 | 0.0192125 | 0.436223 | ||||

0.891142 | 0.276275 | 0.273059 | 0.784181 | 0.0686896 | 0.120258 | ||||

15 20 | 12 16 | I | 0.436934 | 0.147934 | 0.126133 | 0.424671 | 0.0443185 | 0.150657 | |

1.8192 | 0.729329 | 0.212799 | 1.72476 | 0.213135 | 0.149842 | ||||

1.99231 | 0.179585 | 0.0038437 | 1.80985 | 0.090785 | 0.0950757 | ||||

0.278366 | 0.0164655 | 0.0721123 | 0.442964 | 0.0253041 | 0.476548 | ||||

0.776163 | 0.0460672 | 0.108805 | 0.725809 | 0.0103525 | 0.0368705 | ||||

II | 0.654981 | 0.0720127 | 0.309961 | 0.485461 | 0.0107014 | 0.0290777 | |||

1.75885 | 0.424684 | 0.172564 | 1.6856 | 0.116873 | 0.123733 | ||||

1.76559 | 0.736282 | 0.117205 | 1.68275 | 0.280035 | 0.158626 | ||||

0.285547 | 0.0120893 | 0.0481779 | 0.441627 | 0.0235345 | 0.47209 | ||||

0.580928 | 0.0744739 | 0.170103 | 0.626942 | 0.0211994 | 0.104368 | ||||

III | 0.729426 | 0.170489 | 0.458852 | 0.518466 | 0.0316011 | 0.0369323 | |||

1.8595 | 0.41773 | 0.239663 | 1.70108 | 0.115046 | 0.134051 | ||||

2.10522 | 0.644808 | 0.0526092 | 1.89043 | 0.185788 | 0.0547844 | ||||

0.290329 | 0.00288853 | 0.032236 | 0.444839 | 0.0217482 | 0.482798 | ||||

0.620031 | 0.149604 | 0.114242 | 0.638284 | 0.0382819 | 0.0881652 | ||||

25 30 | 15 18 | I | 0.624484 | 0.0707967 | 0.248968 | 0.46255 | 0.0152373 | 0.0748991 | |

2.04136 | 0.641435 | 0.360909 | 1.84179 | 0.200086 | 0.227863 | ||||

2.6438 | 0.597395 | 0.321901 | 2.17796 | 0.0846467 | 0.0889785 | ||||

0.334961 | 0.0181699 | 0.116536 | 0.461212 | 0.0314131 | 0.537372 | ||||

0.853151 | 0.0425477 | 0.218787 | 0.77776 | 0.0131854 | 0.111086 |

25 30 | 15 18 | II | 0.579907 | 0.0865519 | 0.159814 | 0.431046 | 0.0231535 | 0.137909 | |
---|---|---|---|---|---|---|---|---|---|

1.71326 | 0.420802 | 0.142174 | 1.65845 | 0.123108 | 0.105636 | ||||

2.06298 | 0.534335 | 0.03149 | 1.85569 | 0.171584 | 0.0721568 | ||||

0.35053 | 0.0256934 | 0.168434 | 0.466404 | 0.0337755 | 0.554681 | ||||

0.692717 | 0.100299 | 0.0104043 | 0.697282 | 0.0213776 | 0.00388272 | ||||

III | 0.548462 | 0.117228 | 0.0969238 | 0.433495 | 0.0345365 | 0.133011 | |||

1.58744 | 0.458697 | 0.0582929 | 1.60793 | 0.1446 | 0.07195 | ||||

2.16086 | 0.290563 | 0.0804277 | 1.92313 | 0.0779529 | 0.0384327 | ||||

0.249868 | 0.0133291 | 0.167105 | 0.42914 | 0.0190754 | 0.430468 | ||||

0.793726 | 0.0741596 | 0.133894 | 0.742978 | 0.0211759 | 0.0613978 | ||||

25 30 | 20 24 | I | 0.566283 | 0.0899157 | 0.132566 | 0.466032 | 0.0438552 | 0.067936 | |

1.53468 | 0.39365 | 0.023117 | 1.27542 | 0.0491383 | 0.149719 | ||||

2.01437 | 0.740411 | 0.00718381 | 1.50366 | 0.185134 | 0.24817 | ||||

0.301303 | 0.011117 | 0.00434313 | 0.364586 | 0.0228807 | 0.215288 | ||||

0.683733 | 0.0780066 | 0.0232389 | 0.566201 | 0.0214978 | 0.191142 | ||||

II | 0.730676 | 0.151648 | 0.461353 | 0.540285 | 0.0312658 | 0.0805702 | |||

1.62499 | 0.192889 | 0.0833271 | 1.62759 | 0.0665582 | 0.0850597 | ||||

1.73909 | 0.418326 | 0.130453 | 1.67432 | 0.202752 | 0.162838 | ||||

0.285779 | 0.00779342 | 0.0474039 | 0.425446 | 0.0176157 | 0.418152 | ||||

0.599221 | 0.0828184 | 0.143969 | 0.642418 | 0.029898 | 0.0822602 | ||||

III | 0.74725 | 0.119373 | 0.4945 | 0.527233 | 0.0137002 | 0.054467 | |||

1.82079 | 0.28875 | 0.213863 | 1.69809 | 0.0967408 | 0.132062 | ||||

2.3759 | 0.669129 | 0.187952 | 1.97417 | 0.152521 | 0.0129148 | ||||

0.339932 | 0.0101032 | 0.133106 | 0.45045 | 0.0272884 | 0.5015 | ||||

0.956769 | 0.23684 | 0.366813 | 0.828082 | 0.05404 | 0.182975 |

n | m | C.S | ||
---|---|---|---|---|

I | II | III | ||

15 20 | 9 12 | |||

15 20 | 12 16 | |||

25 30 | 15 18 | |||

25 30 | 20 24 |

Point predictions | Interval predictions | |||
---|---|---|---|---|

ML | BSEL | ML | Bayes | |

(L, U) | (L, U) | |||

Length (CP) | Length (CP) | |||

(15, 9) (20, 12) | 0.0185867 | 0.0211427 | (0.000439728, 0.0725284) | (0.000472307, 0.101625) |

0.0720887 (95.66) | 0.101153 (96.70) | |||

(15, 12) (20, 16) | 0.0130117 | 0.0186568 | (0.000315783, 0.0497251) | (0.000460148, 0.103072) |

0.0494094 (92.49) | 0.102612 (96.94) | |||

(25, 15) (30, 18) | 0.0202449 | 0.0207763 | (0.000495333, 0.0768623) | (0.000378492, 0.0922851) |

0.076367 (95.93) | 0.0919067 (97.13) | |||

(25, 20) (30, 24) | 0.0192247 | 0.0238478 | (0.000453452, 0.0751992) | (0.000542946, 0.127486) |

0.0747457 (95.92) | 0.126943 (96.73) |

Point predictions | Interval predictions | |||
---|---|---|---|---|

ML | BSEL | ML | Bayes | |

(L, U) | (L, U) | |||

Length (CP) | Length (CP) | |||

(15, 9) (20, 12) | 0.0242643 | 0.023783 | (0.000582895, 0.0935202) | (0.000431341, 0.101425) |

0.0929373 (95.88) | 0.100993 (97.15) | |||

(15, 12) (20, 16) | 0.0223902 | 0.0210169 | (0.000551125, 0.0845838) | (0.000299128, 0.0841044) |

0.0840327 (96.00) | 0.0838053 (97.30) | |||

(25, 15) (30, 18) | 0.0162685 | 0.0190981 | (0.000386642, 0.0632501) | (0.000350237, 0.0919707) |

0.0628634 (94.87) | 0.0916205 (97.37) | |||

(25, 20) (30, 24) | 0.0138337 | 0.0181426 | (0.00033405, 0.0530935) | (0.000395425, 0.100244) |

0.0527595 (93.03) | 0.0998488 (97.37) |

The 95% BPI for the future observation y b are obtained by solving Equations (36) and (37).

Point predictions | Interval predictions | |||
---|---|---|---|---|

ML | BSEL | ML | Bayes | |

(L, U) | (L, U) | |||

Length (CP) | Length (CP) | |||

(15, 9) (20, 12) | 0.0165701 | 0.0192861 | (0.000407008, 0.0627065) | (0.000430123, 0.0940873) |

0.0622995 (95.04) | 0.0936572 (97.09) | |||

(15, 12) (20, 16) | 0.0149446 | 0.0184964 | (0.000365933, 0.0567025) | (0.000349813, 0.0985559) |

0.0563365 (94.29) | 0.0982061 (97.69) | |||

(25, 15) (30, 18) | 0.0217064 | 0.0220789 | (0.00051196, 0.0849101) | (0.000369217, 0.0986803) |

0.0843982 (96.12) | 0.0983111 (97.24) | |||

(25, 20) (30, 24) | 0.0152917 | 0.0188501 | (0.000380291, 0.057271) | (0.000342122, 0.101289) |

0.0568907 (94.04) | 0.100947 (97.77) |

The BPP for the future observation y b , is computed based on BSEL function using (38) and based on BLINX loss function using (40).

Generate 10 , 000 progressively Type-II censored samples each of size M from a population whose CDF is as (7) with R i * , i = 1 , 2 , ⋯ , M , then calculate the coverage percentage (CP) of Y b . For simplicity, we will consider R i * = 0 , i = 1 , 2 , ⋯ , M which represents the ordinary order statistics and M = N = 10.

The progressive Type-II censoring is of great importance in planning duration experiments in reliability studies. It has been shown by [

From the result, we observe the following:

It is noticed from the numerical calculations that the Bayes estimates under the BSEL function have the smallest MSEs as compared with their corresponding MLEs.

In general, for increasing the effective sample size m / n , the MSEs and ARBs of the considered parameters decrease.

For fixed values of the sample and failure time sizes, the Scheme II in which the censoring occurs after the first observed failure gives more accurate results through the MSEs and RABs than the other schemes and this coincides with Theorem [2.2] by [

The MLEs of β 1 are better than the BEs in general.

In most cases, we observed that when the sample size increased, the MSEs and RABs decreased for all censoring schemes.

The results in Tables 5-7 show that the lengths of the prediction intervals using the ML procedure are shorter than that of prediction intervals using the Bayes procedure.

The simulation results show that the proposed prediction levels are satisfactory compared with the actual prediction level 95%.

Abushal, T.A. and AL-Zaydi, A.M. (2017) Inference on Constant-Partially Accelerated Life Tests for Mixture of Pareto Distributions under Pro- gressive Type-II Censoring. Open Journal of Statistics, 7, 323-346. https://doi.org/10.4236/ojs.2017.72024