11
Author's personal copy A practical procedure for the selection of time-to-failure models based on the assessment of trends in maintenance data D.M. Louit a , R. Pascual b, , A.K.S. Jardine c a Komatsu Chile, Av. Americo Vespucio 0631, Quilicura, Santiago, Chile b Centro de Minerı ´a, Pontificia Universidad Cato ´lica de Chile, Av. Vicun ˜a Mackenna 4860, Santiago, Chile c Department of Mechanical and Industrial Engineering, University of Toronto, 5 King’s College Road, Toronto, Ont., Canada M5S 3G8 article info Article history: Received 24 April 2008 Received in revised form 6 April 2009 Accepted 10 April 2009 Available online 18 April 2009 Keywords: Trend testing Time to failure Model selection Repairable systems NHPP abstract Many times, reliability studies rely on false premises such as independent and identically distributed time between failures assumption (renewal process). This can lead to erroneous model selection for the time to failure of a particular component or system, which can in turn lead to wrong conclusions and decisions. A strong statistical focus, a lack of a systematic approach and sometimes inadequate theoretical background seem to have made it difficult for maintenance analysts to adopt the necessary stage of data testing before the selection of a suitable model. In this paper, a framework for model selection to represent the failure process for a component or system is presented, based on a review of available trend tests. The paper focuses only on single-time-variable models and is primarily directed to analysts responsible for reliability analyses in an industrial maintenance environment. The model selection framework is directed towards the discrimination between the use of statistical distributions to represent the time to failure (‘‘renewal approach’’); and the use of stochastic point processes (‘‘repairable systems approach’’), when there may be the presence of system ageing or reliability growth. An illustrative example based on failure data from a fleet of backhoes is included. & 2009 Elsevier Ltd. All rights reserved. 1. Introduction As described by Dekker and Scarf [1] maintenance optimiza- tion consists of mathematical models aimed at finding balances between costs and benefits of maintenance, or the most appro- priate moment to execute maintenance. Many times, these models are fairly complex and maintenance analysts have been slow to apply them, since often data are scarce or, due to lack of statistical theoretical knowledge, models are very difficult to implement correctly in an industrial setting. Other, more qualitative techniques such as reliability centered maintenance (RCM) or total productive maintenance (TPM) have then played an important role in maintenance optimization. Nevertheless, data analysis and statistical modeling are definitely very valuable tools engineers can employ to optimize the maintenance of assets under their supervision. Acknowledging that many reliability studies or maintenance optimization programs do not require sophisticated statistical inputs, Ansell and Phillips [2] reinforce that even at a basic level, we should always be critical of the analysis and ask whether a technique is appropriate. The gap between researchers and practitioners of maintenance has resulted in the fact that although many models rely on very specific assumptions for their proper application, these are not normally discriminated by the practitioner according to the real operating conditions of their plants or fleets, i.e. real-world data [3,22,43]. O’Connor (cited in [2]) points out that much reliability analysis is done under false premises such as independence of components, constant failure rates, identically distributed vari- ables, etc. As critical constituents of any reliability analysis, time- to-failure models are not excluded of this situation; thus many times the use of conventional time-to-failure analysis techniques is adopted when they are, in fact, not appropriate. The aim of this paper is to provide practitioners with a review of techniques useful for the selection of a suitable time-to-failure model, specifically looking at the case when the standard use of statistical distributions is useless, given the presence of long-term trends in the maintenance failure data. The paper focuses on the selection of single time variable models, since they are the most commonly applied in practice, rather than in more complex multivariate models such as the proportional hazards model, which have also shown great value in their application to maintenance and reliability [3]. The above does not imply that we propose that time-to-failure models should be the center of attention in a reliability improvement study; on the contrary, they should only act as a ARTICLE IN PRESS Contents lists available at ScienceDirect journal homepage: www.elsevier.com/locate/ress Reliability Engineering and System Safety 0951-8320/$ - see front matter & 2009 Elsevier Ltd. All rights reserved. doi:10.1016/j.ress.2009.04.001 Corresponding author. E-mail address: [email protected] (D.M. Louit). Reliability Engineering and System Safety 94 (2009) 1618–1628

Data Para El Manteimiento

Embed Size (px)

DESCRIPTION

Base de datos para el desarrollo del programa de mantenimiento basado en la criticidad

Citation preview

  • Author's personal copy

    A practical procedure for the selection of time-to-failure models based on theassessment of trends in maintenance data

    D.M. Louit a, R. Pascual b,, A.K.S. Jardine c

    a Komatsu Chile, Av. Americo Vespucio 0631, Quilicura, Santiago, Chileb Centro de Minera, Ponticia Universidad Catolica de Chile, Av. Vicuna Mackenna 4860, Santiago, Chilec Department of Mechanical and Industrial Engineering, University of Toronto, 5 Kings College Road, Toronto, Ont., Canada M5S 3G8

    a r t i c l e i n f o

    Article history:

    Received 24 April 2008

    Received in revised form

    6 April 2009

    Accepted 10 April 2009Available online 18 April 2009

    Keywords:

    Trend testing

    Time to failure

    Model selection

    Repairable systems

    NHPP

    a b s t r a c t

    Many times, reliability studies rely on false premises such as independent and identically distributed

    time between failures assumption (renewal process). This can lead to erroneous model selection for the

    time to failure of a particular component or system, which can in turn lead to wrong conclusions and

    decisions. A strong statistical focus, a lack of a systematic approach and sometimes inadequate

    theoretical background seem to have made it difcult for maintenance analysts to adopt the necessary

    stage of data testing before the selection of a suitable model. In this paper, a framework for model

    selection to represent the failure process for a component or system is presented, based on a review of

    available trend tests. The paper focuses only on single-time-variable models and is primarily directed to

    analysts responsible for reliability analyses in an industrial maintenance environment. The model

    selection framework is directed towards the discrimination between the use of statistical distributions

    to represent the time to failure (renewal approach); and the use of stochastic point processes

    (repairable systems approach), when there may be the presence of system ageing or reliability

    growth. An illustrative example based on failure data from a eet of backhoes is included.

    & 2009 Elsevier Ltd. All rights reserved.

    1. Introduction

    As described by Dekker and Scarf [1] maintenance optimiza-tion consists of mathematical models aimed at nding balancesbetween costs and benets of maintenance, or the most appro-priate moment to execute maintenance. Many times, thesemodels are fairly complex and maintenance analysts have beenslow to apply them, since often data are scarce or, due to lack ofstatistical theoretical knowledge, models are very difcult toimplement correctly in an industrial setting. Other, morequalitative techniques such as reliability centered maintenance(RCM) or total productive maintenance (TPM) have then played animportant role in maintenance optimization. Nevertheless, dataanalysis and statistical modeling are denitely very valuable toolsengineers can employ to optimize the maintenance of assetsunder their supervision.

    Acknowledging that many reliability studies or maintenanceoptimization programs do not require sophisticated statisticalinputs, Ansell and Phillips [2] reinforce that even at a basic level,we should always be critical of the analysis and ask whether atechnique is appropriate.

    The gap between researchers and practitioners of maintenancehas resulted in the fact that although many models rely on veryspecic assumptions for their proper application, these are notnormally discriminated by the practitioner according to the realoperating conditions of their plants or eets, i.e. real-world data[3,22,43]. OConnor (cited in [2]) points out that much reliabilityanalysis is done under false premises such as independence ofcomponents, constant failure rates, identically distributed vari-ables, etc. As critical constituents of any reliability analysis, time-to-failure models are not excluded of this situation; thus manytimes the use of conventional time-to-failure analysis techniquesis adopted when they are, in fact, not appropriate.

    The aim of this paper is to provide practitioners with a reviewof techniques useful for the selection of a suitable time-to-failuremodel, specically looking at the case when the standard use ofstatistical distributions is useless, given the presence of long-termtrends in the maintenance failure data. The paper focuses on theselection of single time variable models, since they are the mostcommonly applied in practice, rather than in more complexmultivariate models such as the proportional hazards model,which have also shown great value in their application tomaintenance and reliability [3].

    The above does not imply that we propose that time-to-failuremodels should be the center of attention in a reliabilityimprovement study; on the contrary, they should only act as a

    ARTICLE IN PRESS

    Contents lists available at ScienceDirect

    journal homepage: www.elsevier.com/locate/ress

    Reliability Engineering and System Safety

    0951-8320/$ - see front matter & 2009 Elsevier Ltd. All rights reserved.

    doi:10.1016/j.ress.2009.04.001

    Corresponding author.E-mail address: [email protected] (D.M. Louit).

    Reliability Engineering and System Safety 94 (2009) 16181628

  • Author's personal copy

    tool for the objective the engineers assign to it. Actually, thelogical priority is that of objective, data and, nally, modelselection (as shown in Fig. 1). In other words, as suggested byAnsell and Phillips [4] an analysis should be problem led ratherthan technique or model centered. Nevertheless, correctassessment of the failure process and of time to failure isusually of critical importance to the (posterior) economicanalysis required to nd an optimal solution to the problem thatoriginated the analysis. Discussion of approaches to maintenanceand reliability optimization and models mixing reliability andeconomics can be found in several references, for example [57].

    When dealing with reliability eld data, frequently somepractical problems such as the unavailability of large sets of dataoccur. This paper will briey touch on this and other problems, asthey are relevant to the discussion of model selection techniques.

    This document is structured as follows. Section 2 refers tocommon practical problems found in the analysis of reliabilitydata. Section 3 describes the concept of repairable systems andidenties some of the models available for their representation.Section 4 presents a series of graphical and analytical tests used todetermine the existence of trends in the data. Section 5 proposes aprocedure based on these tests to correctly select a time-to-failuremodel, discriminating between a renewal approach and the use ofan alternative, non-stationary model, such as the non-homo-geneous Poisson process (NHPP). Section 6 presents numericalexamples using data coming from a eet of backhoes. Finally,Section 7 contains a summary of the paper.

    2. Some practical problems in reliability data

    2.1. Scarce data

    One major problem associated with reliability data is,ironically, the lack of sufcient data to properly run statisticalanalyses, as many authors mentioned repeatedly. In fact, Bendell[8] points out that all statistical methodologies are limited whendone based on small data sets, since the amount of informationcontained by such sets is by nature small. Furthermore, asmentioned in [9], empirical evidences indicate that sets of failuretimes typically contain ten or fewer observations, which empha-size the need to develop methods to deal adequately with smalldata sets (naturally the larger the data set, the more precise thestatistical analysis). Also, many data sets are collected formaintenance management rather than reliability. Hence theinformation content is often very poor and can be misleadingwithout careful scrutiny of the material and cleaning if necessary.

    Thomas, a discussant of Ansell and Phillips [10], states that theinsufcient data quantity problemwould never disappear, since asthe aim of maintenance is to make failures rare events, one couldexpect that as maintenance improves, fewer failures should occur.Thus the solution, he says, is based on better a priori modeling.Bayesian techniques, directed to incorporate into the models allthe prior information available, are useful in this case. Barlow andProschan [11], Lindley and Singpurwalla [12], Singpurwalla [13],Walls and Quigley [44] and Guikema and Pate-Cornell [45] arerelevant sources for Bayesian methods in reliability.

    Although it escapes the reach of this paper, it is valuable tomention Bayesian analysis, because it provides a means to reachoptimal decisions, using standard models as the ones discussedlater on in this paper, when lack of data is a problem. In simplewords, and as described in [14], Bayesian estimation methodsincorporate information beyond that contained in the data,such as:

    previous systems estimates; generic information coming from actual data from similar

    systems; generic information from reliability sources and expert judgment and belief.

    This information is converted into a prior distribution, which isthen updated using new data gathered during the operation intoposterior distributions representing the failure process of thecomponent or system of interest. Scarf [3] warns about the factthat many times the expert judgment comes from the samepeople responsible for the maintenance actions; thus it is possiblethat prior distributions may reect the current practice ratherthan the real underlying failure process; so special care should betaken when attempting to use a Bayesian approach.

    2.2. Data censoring

    Another common practical problem in a reliability study is thepresence of censored data. A censored observation corresponds toa non complete time to failure or to a non-failure event, but thisdoes not mean it does not contain relevant information for thereliability modeler. Censoring can usually be classied in right,left or interval censoring. Truncation may also be a practicalproblem in some data sets, commonly confused with censoring.An example of the latter is the case when the time to failure canonly be registered if it lies within a certain observation window(failures that occurred outside this interval are not observed, thusno information is available to the modeler).

    Another situation that may arise is when data collection beginsin a specied moment of time and the operating time of the itemsunder analysis is unknown before the start of the monitoringperiod. The monitored life to failure of a component underobservation can then be called residual life.

    If time-to-failure data are found to be subject of representationby a renewal process (RP), a statistical distribution can be tted totimes between failures (see Section 3). There are well-knowntechniques to determine parameters for many distributions in thepresence of these censoring types. Detailed descriptions of suchtechniques can be found in [1518], among others. Tsang andJardine [19] propose a methodology for the estimation of theparameters of a Weibull distribution using residual-life data.

    2.3. Combining data

    A valid alternative when data are scarce is the combination orpooling of data from similar pieces of equipment. This is a

    ARTICLE IN PRESS

    Objectives

    Maintenance Data

    Model Selection Failure Process

    Optimization Model

    Solution

    Fig. 1. General framework in a reliability improvement study (modied from [2]).

    D.M. Louit et al. / Reliability Engineering and System Safety 94 (2009) 16181628 1619

  • Author's personal copy

    normally used procedure in reliability analysis, especially inoperations where a large number of identical systems are utilized,such as eets of mobile equipment or parallel production lines.Stamatelatos et al. [14] provide a check list of coupling factors forcommon cause failure events (one triggering cause originatesseveral failures), found to be helpful in the denition of conditionsneeded for proper data pooling, as usually these same conditionsare met by equipment subject to data pooling:

    same design; same hardware; same function; same installation, maintenance or operations people

    (and conditions); same procedures; same systemcomponent interface; same location and same environment.

    The word same in the list above could be replaced bysimilar in many cases, as engineers should use their judgmentand experience in assessing the similarities between two or moreitems before they are combined for analysis. If a more rigorousanalysis is needed, when in the presence of two or more samplesof data from possibly different populations, various statisticalmethods can be used to determine if there are signicantdifferences between two populations (or two-sample problem),even in heavily censored data sets [20,21].

    2.4. Effect of repair actions

    Usually in practice, components can be repaired or adjusted,rather than replaced, whenever a breakdown occurs. Theseinterventions (here referred to as repair actions) are likely tomodify the hazard rate of the component; so it can be argued thatthe expected time to failure after an intervention takes place isdifferent from the expected time to the rst failure of a newcomponent. But the most common approach to reliabilityassessment does not take this into account, as time to failure ismodeled using statistical distributions assumed to be valid forevery failure of the component or system (rst, second, third, etc.).This is called the as good as new or renewal assumption. In fact,in the reliability literature, as Ascher and Feingold [22] notice,there is a xation in mortality, or rather, its equivalent in reliabilityterms, time to failure in a non-repairable item or time to rst failure

    of a repairable system.In order to take repair actions into account (when they

    effectively affect the behavior of the component or system understudy), a so called repairable systems approach has to beadopted.

    3. Repairable systems

    A non-repairable system is one which, when it fails, isdiscarded (as repair is physically not feasible or non-economical).The reliability gure of interest is, then, the survival probability.The times between failures of a non-repairable system areindependent and identically distributed, iid [23]. This is the mostcommon assumption made when analyzing time-to-failure data,but as many authors mention, it might be unrealistic in somesituations. Many examples have been given of systems that ratherthan being discarded (and replaced) on failure, are repaired. Inthis case, the usual non-repairable methodologies (statistical

    distribution tting such as Weibull analysis, for instance) simplycannot be appropriate [24].

    Repairable systems, on the other hand, are those that can berestored to their fully operational capabilities by any method,other than the replacement of the entire system [22]. In this sense,reliability is interpreted as the probability of not failing for aparticular period. This analysis does not assume that timesbetween failures are independent or identically distributed. Whendealing with repairable systems, reliability is not modeled interms of statistical distributions, but using stochastic pointprocesses.

    The number of failures in an interval of time can berepresented through a stochastic point process. Furthermore, inthis case the point process can be interpreted as a countingprocess, and what it counts is the number of events (failures) in acertain time interval. In reliability analysis, such a process is saidto be time truncated when it stops counting at a particular instant.It is called failure truncated when it stops counting when a certainnumber of failures is reached.

    The ve main stochastic process models applied to modeling ofrepairable systems are [22]

    The renewal process (RP). The homogeneous Poisson process (HPP). The branching Poisson process (BPP). The superposed renewal process (SRP) and The non-homogeneous Poisson process (NHPP).

    The RP assumes that the system is returned to an as newcondition every time it is repaired, so that it actually converts timebetween failures in time to rst failure of a new system or, inother words, leads to a non-repairable system approach, inwhich time to failure can be modeled by a statistical distributionand the iid assumption is valid. The HPP is a special case of the RP,that assumes that times between failures are independent andidentically exponentially distributed, so the iid assumption is alsovalid and the time to failure is described by an exponentialdistribution (constant hazard rate).

    The BPP is used to represent time-to-failure data that can beassumed to be identically distributed, but not independent. AsAscher and Feingold [22] mention, this process is applicable whena primary failure (or a sequence of primary failures having iidtimes to failure) can trigger one or more subsidiary failures; thusthere is dependence between the subsidiary failures and theoccurrence of the primary, triggering failure. Very few practicalapplications of this model are found in the literature. A thoroughdescription of the BPP and its application to the study ofrepairable systems can be found in [25].

    The SRP is a process derived from the combination of variousindependent RPs, and in general it is not an RP. For example, thinkof a set of parts within a system that are discarded and replacedevery time they fail, independently. Each part can be modeled asan RP, and then the systemwould be modeled using an SRP. But asa possibility exists to investigate the times between failures forthe system as a whole, then the question whether this approach isjustied or not arises [4]. In addition, the superposition ofindependent RPs converges to a Poisson process (possibly non-homogeneous), when the number of superimposed processesgrows (by the theorem of Grigelionis, see [26]).

    Since the RP and the HPP are equivalent to the regular, non-repairable items methods, and the BPP and SRP have either notbeen largely applied or, in the case of the latter, can beapproximated by an NHPP (with a relatively large number ofconstituent processes), they will not be described in greaterdetail here. These models are covered in detail by Ascher andFeingold [22].

    ARTICLE IN PRESS

    D.M. Louit et al. / Reliability Engineering and System Safety 94 (2009) 161816281620

  • Author's personal copy

    When the repair or substitution of the failed part in a complexsystem does not involve a signicant modication of thereliability of the equipment as a result of the repair action, theNHPP is able to correctly describe the failure-repair process. Then,the NHPP can be interpreted as a minimal repair model [27]. Notethat for (i) hazardous maintenance, i.e. when condition of theequipment is worse after repair than it was before failure or for (ii)imperfect repair, i.e. when reliability after repair is better than justbefore failure, but not as good as new, other models have beenproposed. These models are even more exible than the NHPP, asthey allow for better representation of imperfect repair scenarios(see, e.g. [28,29], among others). Nevertheless, we concentrate onthe NHPP given its simplicity, along with the following reasons(as listed by Coetzee [30]):

    i. it is generally suitable for the purpose of modeling data with atrend, due to the fact that the accepted formats of the NHPPare monotonously increasing/decreasing functions;

    ii. NHPP models are mathematically straightforward and theirtheoretical base is well developed;

    iii. models have been tested fairly well, and several examples areavailable in the literature for their application.

    Under the NHPP, times between successive failures are neitherindependent nor identically distributed, which makes this modelthe most important and widely used in the modeling of repairablesystems data [31,32]. Actually, whenever a trend is found to bepresent in time between failures data, a non-stationary modelsuch as the NHPP is mandatory, and the regular distribution ttingmethods are not valid.

    The next section of the paper reviews a series of trend-testingtechniques found very helpful for model selection purposes,focusing on the discrimination between a renewal approach andthe need for an alternative model, such as the NHPP. Should thereader decide to pursue the modeling of times to failure usingother non-stationary models (i.e. imperfect repair models), thetechniques presented in this paper are equally valuable toestablish the existence of trends in the data, which justify thedecision of not using the standard distribution tting methodssuch as Weibull analysis.

    Fig. 2 shows three theoretical situations that may occur inpractice, in relation to time to failure of a particular system. Fromthe gure, it can be noticed that the three data sets generatedfrom systems AC are very different. For A, no clear trend can be

    observed from the failure data, thus an iid assumption may bemade, since time between failures is apparently independent fromthe age of the equipment. For B and C, however, a trend is clearlypresent; for B a decrease in reliability is evident, while for Creliability growth is taking place. Whenever these latter situationsoccur, and there is signicant evidence for recognition of anageing process taking place, the usual RP approach has to bedisregarded, and an alternative non-stationary approach wouldusually be used to model time between failures for the system.Note that equipment age refers to the age of the system underanalysis, measured from the moment it was rst put intooperation, as apposed to the time elapsed since the last repair(which is signicant for a RP).

    3.1. The non-homogeneous Poisson process

    In practical terms, the NHPP permits the modeling of trend inthe number of failures to be found in an interval in relation to totalage of the system, through the intensity function. Two popularparameterizations for the intensity function of an NHPP are thepower law intensity (also called Weibull intensity) and the log-linear intensity.

    The power law intensity gives its name to the power lawprocess (NHPP with Weibull intensity) and is given by

    lt Zbtb1, (1)with Z, b40 and tX0.

    The log-linear intensity function has the form

    lt eabt , (2)with Noa,boN and tX0.

    Several practical examples reviewed show that the power-lawprocess is preferred, because of its similarities with the Weibulldistribution tting method used regularly for non-repairable data.Actually, Z is the scale parameter and b the shape parameter, andthe intensity function is of the same form of the failure rate of aWeibull distribution.

    Details concerning the tting of a power-law and log-linearintensity functions to data from water pumps in a nuclear plantare discussed in [23]. A practical example using jet engine datacombined from several pieces of equipment is found in [24].Ascher and Kobbacy [33], and Baker [34] also provide applicationsof both log-linear and power-law processes.

    When using an NHPP, engineers could imagine it as a two-stage problem: the rst (say, inner stage) relates to the tting ofan intensity function to data and the second (outer stage) usesthe cumulative intensity function to estimate reliability(or probability of failure) for the system. When trends are notpresent and data can be assumed to be iid, only one stage isneeded, where directly a time-to-failure distribution is tted todata and reliability (or probability of failure) estimates areobtained from it.

    4. Trend testing techniques

    Given that a clear need for trend testing in time betweenfailures data is identied, a rst step in model selection should bethat of assessing the existence of trend or time dependency in thedata. Several techniques accomplish this task, and a selection ofthem is described here.

    Before presenting the testing techniques, it is of greatimportance to identify the possible trends one may encounterwhen analyzing reliability data. A trend in the pattern of failurescan be monotonic or non-monotonic. In the case of a monotonictrend (such as the ones shown in Fig. 2), the system is said to be

    ARTICLE IN PRESS

    Age

    Age

    Age

    For each of the following diagrams, represents the occurrence of a

    failure:

    Fig. 2. Possible trends in time between failures.

    D.M. Louit et al. / Reliability Engineering and System Safety 94 (2009) 16181628 1621

  • Author's personal copy

    improving (or happy system) if the time between arrivals tendsto get longer (decreasing trend in number of failures); and it issaid to be deteriorating (or sad system) if the times tend to getshorter (increasing trend in number of failures). Non-monotonictrends are said to occur when trends change with time or theyrepeat themselves in cycles. One common form of non-monotonictrend is the bath-tub shape trend, in which time between failuresincreases in the beginning of the equipment life, then tends to bestable for a period and decreases at the end.

    It should be remembered, when testing for trend, that thechoice of the time scale could have an impact on the pattern offailures; so special attention has to be given to the selection of thetime unit (calendar hours, operating hours, production throughput, etc. [31]).

    4.1. Graphical methods

    4.1.1. Cumulative failures vs. time plot

    The simplest method for trend testing is a plot of thecumulative failures against time for the system observed (New-ton, in Ansell and Phillips [4]). When a linear plot results, data canbe assumed to have no trend and the same distribution for timebetween failures is accepted. Fig. 3 shows generic plots expected.Plot A clearly shows the existence of a trend in the data, while plotB shows no evidence of trend. Sometimes, a curve like plot Coccurs, where instead of a smooth trend, two or more straightlines can be plotted. This may be the consequence of changes inthe maintenance policy or changes in the operational conditionsof the equipment; for example, dividing the failure behavior of thesystem into two or more clearly different periods. When thissituation arises, one alternative is to discard data notrepresentative of the current situation; thus a no-trend plotwould result, for the most recent data set, and a renewalassumption could then be made. When a plot like D occurs, anon-monotonic trend may be present in the data set.

    This kind of test is very simple to perform, does not require anycalculations and is very powerful when there are strong trends inthe data. When slight trends are present, this solution may not beenough and an analytical test should be performed. A weaknessfor this test is that assessment of trend is based on interpretation(as in all graphical procedures). Ascher and Feingold [22] providean example of the use of this graphical test for diesel propulsionengines in a US Navy ship.

    Also in [22], it is noted that this test may result in maskinglocal variations when very large samples are available. Analternative procedure is to divide the total observation intervalinto several equally sized intervals, calculating (and then plotting,if necessary) the average rate of occurrence of failures for each ofthem, using

    lit Nit Ni1t

    Dtwith i 1DtptpiDt; (3)

    where Ni(t) is the total number of failures observed from time zeroto the ith interval and Dt the length of each interval. If there is atrend in the data, then it will be reected in the average rate ofoccurrences calculated. Then, if the system is improving, thesuccessive values of li(t) calculated will decrease and vice versa.

    4.1.2. Scatter plot of successive service lives

    A complementary test to the cumulative failures against timeplot is one consisting in plotting the service life of the ith failure,against that of the (i1)th failure If no trend is observable, onlyone cluster of points should be obtained. Two or more clusters, orlinear plots, indicate trend.

    This test is also very helpful in checking for unusual values forthe failure times in a set of data, which may be related to poordata collection, accidents or other situations not representative ofthe failure process, and thus providing a means for identicationof candidates for data ltering. Knights and Segovia [35], forexample, applied this test to data coming from mining shovelcables (see Fig. 4 for an example of this type of plot, points out ofthe cluster suggest a revision of some failure times).

    The tests described up to this point are for a single system only.When in presence of multiple systems, two alternatives areavailable for combination of the systems for assessment of trend.The rst is based on the assumption that all systems follow thesame failure process (independently), and leads to the use of thetotal time on test (TTT) transform of failure times (see denitionin Section 4.1.4). This approach results in one single process withtimes to failure given by the TTT transformed values; thus single-system tests can be applied to the transformed data set. Thesecond alternative assumes that all systems follow possiblydifferent failure processes, and leads to the combination of the

    ARTICLE IN PRESS

    Time Time Time Time

    Cum

    ulat

    ive

    Failu

    res

    Cum

    ulat

    ive

    Failu

    res

    Cum

    ulat

    ive

    Failu

    res

    Cum

    ulat

    ive

    Failu

    res

    Fig. 3. Cumulative failures vs. time plotsexamples (A: Increasing trend, B: no trend, C: two clearly different periods, D: non-monotonic trend).

    Serv

    ice

    life

    (ith f

    ailure

    )

    Service life ((i-1)th failure) Fig. 4. Example of a successive service life plot (highlighted points indicateanomalies).

    D.M. Louit et al. / Reliability Engineering and System Safety 94 (2009) 161816281622

  • Author's personal copy

    results of various single-system tests into what is called acombined test. This is usually performed by combining teststatistics from single systems (for examples see Sections 4.2.2and 4.2.4).

    4.1.3. NelsonAalen plot

    Another useful graphical test is the NelsonAalen plot. This testuses a non-parametric estimate of the cumulative intensityfunction of an NHPP, L(t), and plots it against time [31]. Theestimate is given by

    L^t

    XTijpt

    1

    YTij, (4)

    where Tij is the time to the ith failure of the jth process underobservation, Y(Tij) the number of systems operating immediatelybefore time Tij and L(t) 0 for tomin{Tij}. The formula in Eq. (4)is valid for multiple systems under observation (multipleprocesses, j 1,2,y,m).

    If there is no trend, then the plot would tend to be linear, andany deviation from a straight line indicates some kind of trend. Itshould be noted that when only one system is observed, then theNelsonAllen plot is equivalent to the cumulative failures vs. timeplot. It is also interesting to notice that the NelsonAalen plotcounts the number of systems operating before a certain time;thus it may include suspensions to assess trend.

    4.1.4. Total time on test (TTT) plot

    As mentioned above, sometimes we are in the presence ofseveral pieces of equipment. Now, the combined failure processfor the entire group of components observed may or may notpresent a trend. This test is directed to the identication of trendfor the combined behavior. So, if there are m independentprocesses with the same intensity function (i.e. several identicalsystems under observation) and the observation intervals for eachone are all contained in the interval [0,S], then the total number offailures will be N Pmi1ni, where ni is the number of failuresobserved for each process in its particular observation interval.

    For the superposed process (combination of the m individualprocesses), let Sk denote the time to the kth failure time. And letp(u) denote the number of processes under observation at time u.If all processes are observed from time 0 to time S, then p(u) isequal to m. Then, Tt R t0 pudu is the total time on test fromtime 0 to time t (this is known as the total time on test, or TTT,transformsee [36]).

    The TTT plot test for NHPPs is given by a plot of the total timeon test statistic, calculated as

    TSkTS

    R Sk0 puduR S0 pudu

    , (5)

    against the scaled failure number k/N, with k 1,2,y,N. Whenp(u) m, that is, when all processes are observed during thecomplete interval, the TTT plot is also called a scaled NelsonAllenplot with axes interchanged [31]. Fig. 5 shows different formspossible to obtain when constructing a TTT plot. As in othergraphical techniques, a linear plot is representative of a no-trendsituation (thus validating a renewal assumption for the entiregroup of observed items). The TTT plot is especially useful toidentify non-monotonic trends in time-to-failure data, such as thebath-tub failure intensity (see Fig. 5C).

    It is important to mention that spacing between points will notbe constant in a TTT Plot. Rather, in an increasing trend (Fig. 5A),larger spacement between points will occur at the lower leftsection of the plot. In the presence of a decreasing trend curve(Fig. 5B) points should tend to be further from each other at the

    upper right section, whereas for a bath-tub shape (Fig. 5C), furtherspacing will occur in the middle section of the curve.

    Some other graphical tools, such as control charts for reliabilitymonitoring (described by Xie et al. [37]), can also constitute auseful method to identify if improvement or deterioration hasoccurred in a particular parameter of interest, such as the rate ofoccurrences of failures (ROCOF) or failure intensity. Nevertheless,they rely on an RP assumption and are not directed to test fortrend when evaluating the use of a repairable systems approach.

    4.2. Analytical methods

    If preferred over the graphical approach, analytical testingmethods are available to test data for trends. Additionally, the nulland alternative hypotheses of these tests are of great help in thedetermination of the most suitable model for the data.

    Ascher and Feingold [22] provide a very complete survey ofanalytical trend tests, and present them organized according totheir null hypothesis (i.e. RP, HPP, NHPP, monotonic trend, non-monotonic trend, etc.). Hereby, only the most popular tests will bedescribed, according primarily to Elvebakk [38]. Other methodsare described and referenced in [46].

    4.2.1. The Mann test

    The null hypothesis for this non-parametric test is an RP. Then,if this hypothesis is accepted, we can continue the reliabilityanalysis, tting a distribution to time-to-failure data. Thealternative hypothesis is a monotonic trend.

    The test statistic is calculated counting the number of reversearrangements,M, among the times between failures. Let T1,T2,y,Tnbe the interarrival times of n failures. Then a reverse arrangementoccurs whenever TioTj for ioj. For example, if the following timesto failure were observed for a system:

    21;17;48;37;64;13;

    then, for the rst failure time, 3 reverse arrangements occur, as21o48, 21o37 and 21o64. We have that for the sample:

    M 3 3 1 1 0 8.In general

    M Xn1i1

    Xnji1

    ITioTj (6)

    I( ) is an indicator variable used for counting the reversearrangements present in the data set. It takes the value of 1whenever the condition is met, in this case, when (TioTj). Mann[39] who originally developed the test, showed that M isapproximately normally distributed for nX10 and tabulatedprobabilities for smaller samples.

    If the hypothesis of an RP is correct, then the expected numberof reverse arrangements is equal to n(n1)/4, so large deviations

    ARTICLE IN PRESS

    1 1 1

    0 1 0 1 0 1

    Fig. 5. Typical shape of TTT plots. A: increasing, B: decreasing, C: bath-tub shapeintensities. Modied from [31]).

    D.M. Louit et al. / Reliability Engineering and System Safety 94 (2009) 16181628 1623

  • Author's personal copy

    from this number would indicate presence of a trend. This testconsiders a single system under observation.

    4.2.2. The Laplace test

    This well-known test has a null hypothesis of HPP vs. analternative hypothesis of NHPP with monotonic intensity. In otherwords, if the null hypothesis is not rejected, then we can assumethat times between failures are iid exponentially distributed. If not,then a NHPP should be used. The test is optimal for NHPP withlog-linear intensity function.

    The general idea behind the test is to compare the mean valueof the failure times in an interval with the midpoint of theinterval. If the mean of the failure times tends to deviate from themidpoint, then a trend is present and data cannot be assumed tobe independent and identically distributed.

    The test statistic, L, approximately follows a standard normaldistribution under the null hypothesis, and is calculated as

    L Pn^

    j1Tj 12n^b a

    112n

    ^b a2q , (7)

    where Tj is the age at failure for the jth failure, [a,b] is the intervalof observation and n

    ^is given by:

    n^

    nobserved number of failures if the process is time truncatedn 1 if the process is failure truncated:

    (

    Note that Eq. (7) can be simplied when the starting point ofobservation is time t 0, since (b+a) and (ba) both equal the endpoint of the observation interval. The statistic above is applicablefor the case when only one process is being observed. General-ization of the laplace Test to more than one process is fairlysimple, and for m processes, the statistic is given by the followingexpression (combined Laplace test statistic):

    L Smi1S

    ni^

    j1Tij Smi112ni^ bi ai

    112S

    mi1n

    ^ibi ai2

    q (8)L is expected to be equal to zero (or very close to zero) when notrend is present in the data. Then, the null hypothesis is rejectedwith small or large values of L, and the sign is an indicationof the type of trend. If L40, then an increasing trend(deterioration) is detected. Analogously, if Lo0, then a decreasingtrend (improvement) is detected.

    4.2.3. The LewisRobinson test

    This test, used for testing of the RP assumption, was derived byLewis and Robinson [40]. The test statistic LR is obtained bydividing the Laplace test statistic L by the estimated coefcient ofvariation of the times between failures, ccv, which is calculated asccv sX

    X, (9)

    where X is a random variable representing the times betweenfailures of the system. Then, the LR statistic is given by

    LR Lccv , (10)with L given by Eq. (7). If the failure times follow a HPP, then LR isasymptotically equivalent to L, as ccv is equal to 1 when the timesbetween failures are exponentially distributed. That is, LR isasymptotically standard normally distributed. As in the Laplacetest, the expected value of the statistic is zero when no trend ispresent; thus deviations from this value indicate trend. The sign isan indication of the type of trend.

    4.2.4. The military handbook test

    As in the Laplace test, the null hypothesis for this one is a HPP,and the alternative a NHPP with monotonic intensity. This test isoptimal for NHPP with increasing power-law intensity (reliabilitydeterioration with Weibull intensity function).

    The test statistic for a single system (process) is w2 distributedwith 2n

    ^degrees of freedom under the null hypothesis, and is

    dened as

    MH 2Xn^j1

    lnb aTj a

    , (11)

    where a, b, Tj and n^have the same meaning as in the Laplace test.

    As before, the generalization to m processes is given by(combined MH test statistic):

    MH 2Xmi1

    Xni^j1

    lnbi aiTij ai

    . (12)

    In this case, the MH statistic is w2 distributed with 2p degrees offreedom under the null hypothesis of HPPs, where p Pmi1n^i.

    TTT-based statistics for both the Laplace and the MilitaryHandbook test are also available for the pooling of data fromseveral systems (see [31]).

    Another test, known as the AndersonDarling Test (derived byAnderson and Darling [41]), has been found to be very powerfulagainst non-monotonic trends, but normally simpler graphicaltests are able to detect this situation. For this reason, it will not bedescribed here.

    5. Model selection procedure

    Vaurio [42] and Ascher and Feingold [22] proposed proceduresbased on various trend tests, directed to the proper selection ofmodels for time-to-failure data. Both methodologies are robustand incorporate a set of tests leading to the selection of a model,but are subject to simplication in order to achieve a larger use ofthe testing techniques by maintenance analysts. Based on this, anew diagram consisting of several steps to model selection isproposed. This procedure only considers explicitly two models withpractical applicationthe RP and the NHPP (though it leaves theoption open to the user to select other non-stationary models).The procedure also reduces the number of tests considered inorder to concentrate the users efforts on the techniques that seemto be subject to easier practical implementation. The diagrampresented below is similar to that of Vaurio [42], which thoughvery complete in its procedure for model selection, appears to betoo complex for regular industrial application. The Ascher andFeingold [22] ow diagram (AF ow diagram) is simpler, but asthey consider a broader review of tests and do not include themexplicitly in the graphical representation of the procedure, it canpossibly result in misguiding the practitioner.

    Fig. 6 presents the suggested guideline for model selection,applying the testing techniques reviewed here. As mentioned, thisprocedure is believed to be a simple way for maintenance analyststo correctly assess the failure processes in their operations and todiscriminate whether a standard renewal approach or arepairable systems approach should be used to represent them.

    Although the use of an NHPP is suggested in this paper as it iscapable of representing data with a trend, the reader should notethat the NHPP is best interpreted as a minimal repair (or as badas old) model, thus it will not be necessarily be the mostappropriate model for imperfect repair situations (neither as goodas new nor as bald as old system after repair). Minimal repair isdened here as the situation when the components reliability

    ARTICLE IN PRESS

    D.M. Louit et al. / Reliability Engineering and System Safety 94 (2009) 161816281624

  • Author's personal copy

    characteristics are not noticeable changed by the repair action.Nevertheless, the procedure remains valid for the testing of therenewal assumption, even in the scenario of imperfect repair.When imperfect repair is observed, more exible models (e.g.models I and II by Kijima [29]) could provide a better t to thefailure data.

    The box for testing for renewal assumption and the box fortesting against NHPP have been grouped in Fig. 6, under TEST.Intuitively, all tests considered (graphical tests, Mann, Laplace,LewisRobinson and Military Handbook) have the sameobjective of testing for trends in the data, but only the Laplaceand Military Handbook tests explicitly have a HPP as nullhypothesis (that is, if no trend is found when applying thesetests, then a HPP can be used, and an exponential distributionshould be tted to failure times). It is important to noticethat suspended data is not considered in trend testing, as notechnique is currently available to assess the existence of trendsincorporating the effect of censored observations, to my knowl-edge, with the exception of the NelsonAalen plot. This plot,as it counts the number of systems in operation before eachfailure, may indirectly use suspensions in the trend assessmentcalculations. This is not necessarily a shortcoming of the frame-work proposed, since the tests reviewed usually need few failuretimes to be validated (using only failures and ignoring censoredobservations).

    Two-sample techniques allow for evaluation of the poolingof censored data sets, so that intensity functions or distributionsmay be tted using more information. In a repairable systemsapproach, the simplest extension of the treatment of a singlesystem to multiple systems is when they are all observed from thesame time, as in this combination the group may be thought of asa single system. In this case, the rate of occurrences of failures forthe combined set must be divided by the number of systems todetermine the ROCOF for a single unit.

    6. Case study

    Failure data coming from a eet of backhoes, collectedbetween 1998 and 2003, are used in the following numericalexample to illustrate the use of the trend tests and selectionprocedure described in the paper. These equipments are operatedby a construction rm in the United States. The data consist of theage at failure for each of 11 pieces of equipment, with a total of 43failures. Table A1 in the Appendix presents the complete data set.The following example will consider two cases: (i) single-systemanalysis, for which all calculations are based on backhoe #7 (with7 failures during the observation period) and (ii) multiple-systemsanalysis, using the pooled data for all 11 backhoes. Time to failureis expressed in operating hours.

    ARTICLE IN PRESS

    Collect operating time for each failure registered

    Order them chronologically (failures only)

    Test for Renewal Assumption

    Evaluate Bayesian

    techniques

    CMMS Databases

    1. Define object of study

    2. Identify similar systems

    Nodata?

    Valid to combine?

    Fitdistributions

    to data

    Notrend RPValid?

    Evaluate goodness of

    fit

    Graphical tests (any)

    Mann Test

    Test against NHPP Laplace Test LR test Military HB Test

    HPPValid?

    Weibull Exponential

    other

    HPPRejected?

    Determine intensityfunction

    Weibull Log-linear

    Evaluate goodness of

    fit

    RP

    NHPP(or other non-stationary model)

    TIME TO FAILUREMODEL

    TEST

    Fig. 6. Framework for time-to-failure model selectiona practitioners approach.

    D.M. Louit et al. / Reliability Engineering and System Safety 94 (2009) 16181628 1625

  • Author's personal copy

    6.1. Single-system example

    Backhoe number 7 had seven failures during the years19982003. The times between failures are: 3090, 850, 904, 166,395, 242 and 496h. By looking at these numbers, deteriorationappears to be present in the data, that is, times between failuresseem to be shorter as the equipment ages. Fig. 7 (cumulativefailures vs. time plot) conrms this belief. Similar plots areobtained for 7 out of the 11 backhoes analyzed. When plotting thesuccessive service lives, only two points lay outside the maincluster (implying that one failure time might be an anomaly, seeFig. 8). This is the rst service life of 3090hours, which is muchlarger than the rest. This failure time (as well as all other failuretimes in the data set) was validated by the eet operator;thus noneed for further revision or elimination of data is identied.According to the procedure depicted in Fig. 6, the Mann test couldbe used to conrm the ndings obtained by the cumulativefailures vs. time plot, that a trend is present in the data. In thisexample, M 0+1+0+3+1+1 6. This number differs signicantlyfrom the expected value of 10.5. Actually, it is signicantly low, asPMoM=H1 0:881, implying that the data show a degradationtrend (H1 is the alternative hypothesis, of degradation trend).

    As a trend was identied, a renewal approach is not valid formodeling time to failure, and a repairable systems approach isrequired. Alternatively, testing through parametric tests such asthe Laplace, LewisRobinson or Military Handbook tests wasperformed, with similar results (NHPP with monotonic intensity isthe model selected). The Laplace statistic L equals 2.189 forbackhoe #7. This is given that the process in this case is failuretruncated (e.g. nal point of the observation interval given by thefailure time of the last recorded failure event). This result is

    statistically signicant (at 0.05 signicance level), indicating anincreasing trend in the intensity of the failure process (i.e.degradation). The LR statistic equals 10.187 and the MH statisticequals 3.5698, leading us to the same conclusion (the latter at 0.01signicance level). Fitting of an intensity function to the data isthe next step considered in the procedure presented in Fig. 6, butwill not be included here for brevity. It is important to mentionthat not all tests need be performed in this case. On the contrary,the idea is to choose a test that accommodates the user and, onlyin the case that not enough evidence is available, then validate itsresults using a second testing technique.

    6.2. Multiple-system example

    If we were interested in modeling the time to failure for thepooled group of backhoes, existence of trends in the pooledbehavior should be assessed. A common mistake is to assume thatevery piece of equipment operated for the same number of hoursover the entire interval, which in this case is not true. As thesebackhoes are operated by a construction company in differentprojects, during the same observation period of 19982003, someof them operated more than 6000h, whereas others barelyreached 2500h of operation. Then, for the performance ofgraphical tests such as the TTT or NelsonAalen plots, the modelershould have special care in identifying the number of units inoperation for different ages of the eet. Note that time isexpressed in operating hours, so in the superposed failure process,11 backhoes can be considered to be in operation just for theinterval between t 0 and t 2028 operating hours. For moreadvanced ages of the eet, the number of backhoes in operationdecreases. Fig. 9 shows a TTT plot constructed for this example.From the form of the curve, a clear indication of an increasingtrend in the intensity of failures is observed (i.e. degradation).Fig. 10 presents a Nelson-Aalen plot for the same data, againsuggesting the same conclusion.

    Results for the combined Laplace and Military Handbook testsare the following: L 3.096 (signicant indicator of degradation,at 0.01 signicance level) and MH 31.859 (signicant indicatorof degradation, at 0.01 signicance level). Then, the pooled failurebehavior for the eet effectively presents a trend, thus a RP cannot

    ARTICLE IN PRESS

    012345678

    0Age (hours)

    Cum

    ulat

    ive

    failu

    res

    1000 2000 3000 4000 5000 6000 7000

    Fig. 7. Cumulative failures vs. time plotbackhoe #7.

    0

    500

    1000

    1500

    2000

    2500

    3000

    3500

    0

    Serv

    ice

    life

    (ith fa

    ilure

    ) (ho

    urs)

    500 1000 1500 2000 2500 3000 3500Service life ((i-1)th failure) (hours)

    Fig. 8. Successive service life plotbackhoe #7.

    0

    1

    10Scaled failure number

    TTT

    stat

    istic

    Fig. 9. TTT plotall backhoes combined. Increasing intensity is suggested.

    D.M. Louit et al. / Reliability Engineering and System Safety 94 (2009) 161816281626

  • Author's personal copy

    be assumed (the use of an NHPP is suggested). This was expectednot only because the graphical tests indicated a similar trend, butalso because the results obtained for single systems showed thatmany of the backhoes presented individually a deteriorationtrend. The next step in this example would be, then, to estimateparameters for an NHPP representing the failure process of thebackhoes. Coetzee [30] contains expressions for parameterestimation and goodness of t procedures for an NHPP.

    7. Conclusions

    This paper reviews several tests available to assess theexistence of trends, and proposes a practical procedure todiscriminate between (i) the common renewal approach to modeltime to failure and (ii) the use of a non-stationary model such asthe NHPP, which is a model believed to be subject to an easypractical implementation, within the alternatives availablefor a repairable systems approach. The procedure suggestedis simple, yet it is believed that it will lead to better represen-tation of the failure processes commonly found in industrialoperations. Through numerical examples, the use of the severaltests reviewed is illustrated. Some practical problems thatone may encounter when analyzing reliability data are alsobriey discussed and references are given in each case for furtherreview.

    Acknowledgements

    We would like to thank Dr. Dragan Banjevic, of the Center forMaintenance Optimization and Reliability Engineering at theUniversity of Toronto, for his valuable comments on an earlierversion of this paper.

    Appendix. Data used in the numerical example

    See Table A1.

    References

    [1] Dekker R, Scarf PA. On the impact of optimisation models in maintenancedecision making: the state of the art. Reliability Engineering and SystemSafety 1998;60:1119.

    [2] Ansell JI, Phillips MJ. Strategies for reliability data analysis. In: Comer P, editor.Proceedings of the 11th advances in reliability technology symposium.London: Elsevier; 1990.

    [3] Scarf PA. On the application of mathematical models in maintenance.European Journal of Operational Research 1997;99:493506.

    [4] Ansell JI, Phillips MJ. Practical problems in the statistical analysis of reliabilitydata (with discussion). Applied Statistics 1989;38:20531.

    [5] Jardine AKS, Tsang AHC. Maintenance, replacement and reliability: theory andapplications. Boca Raton: CRC Press; 2006.

    [6] Campbell JD. Uptime: strategies for excellence in maintenance management.Portland: Productivity Press; 1995.

    [7] Campell JD, Jardine AKS, editors. Maintenance excellence: optimizingequipment life-cycle decisions. New York: Marcel Dekker; 2001.

    [8] Bendell T. An overview of collection, analysis, and application of reliabilitydata in the process industries. IEEE Transactions on Reliability 1998;37:1327.

    ARTICLE IN PRESS

    0

    5

    10

    15

    20

    25

    0

    Cum

    ulat

    ive

    inte

    nsity

    2000 4000 6000Tij

    Fig. 10. NelsonAalen plotall backhoes combined. Increasing intensity issuggested.

    Table A1Failure times for a eet of 11 backhoes.

    Fleet of backhoesfailure data

    Equipment # Failure # Age at failure (h) TBF (h)

    1 1 346 346

    2 1925 1579

    3 3108 1183

    4 3610 502

    5 3892 282

    2 1 875 875

    2 2162 1287

    3 3248 1086

    4 4422 1174

    3 1 2920 2920

    2 4413 1493

    3 4691 278

    4 4801 110

    4 1 1234 1234

    2 1911 677

    3 2352 441

    4 3063 711

    5 1 896 896

    2 1885 989

    3 2028 143

    6 1 1480 1480

    2 3648 2168

    3 5859 2211

    7 1 3090 3090

    2 3940 850

    3 4844 904

    4 5010 166

    5 5405 395

    6 5647 242

    7 6143 496

    8 1 1710 1710

    2 1787 77

    3 2297 510

    4 2915 618

    9 1 1885 1885

    2 2500 615

    3 2815 315

    10 1 1691 1691

    2 2230 539

    3 2500 270

    11 1 1210 1210

    2 2549 1339

    3 2621 72

    D.M. Louit et al. / Reliability Engineering and System Safety 94 (2009) 16181628 1627

  • Author's personal copy

    [9] Percy DF, Kobbacy KAH, Fawzi BB. Setting preventive maintenance scheduleswhen data are sparse. International Journal of Production Economics1997;51:22334.

    [10] Ansell JI, Phillips MJ. Discussion of practical problems in the statisticalanalysis of reliability data (with discussion). Applied Statistics 1989;38:23147.

    [11] Barlow RE, Proschan F. Inference for the exponential life distribution. In:Serra A, Barlow RE, editors. Theory of reliability, Proceedings of theInternational School of Physics Enrico Fermi. Amsterdam: North-Holland;1986. p. 14364.

    [12] Lindley DV, Singpurwalla ND. Reliability and fault tree analysis using expertopinions. Journal of the American Statistical Association 1986;81:8790.

    [13] Singpurwalla ND. Foundational issues in reliability and risk analysis. SIAMReview 1988;30:26481.

    [14] Stamatelatos M, et al. Probabilistic risk assessment procedures guide forNASA managers and practitioners. Washington, DC: Ofce of Safety andMission Assurance NASA Headquarters; 2002.

    [15] Meeker WQ, Escobar LA. Statistical methods for reliability data. New York:Wiley; 1998.

    [16] OConnor PDT. Practical reliability engineering. 3rd ed. New York: Wiley; 1991.[17] Mann NR, Shafer RE, Singpurwalla ND. Methods for statistical analysis of

    reliability and life data. New York: Wiley; 1974.[18] Barlow RE, Proschan F. Mathematical theory of reliability. New York: Wiley;

    1965.[19] Tsang AH, Jardine AKS. Estimators of 2-parameter Weibull distributions from

    incomplete data with residual lifetimes. IEEE Transactions on Reliability1993;42:2918.

    [20] Bohoris GA. Parametric statistical techniques for the comparative analysis ofcensored reliability data: a review. Reliability Engineering and System Safety1995;48:14955.

    [21] Bohoris GA, Walley DM. Comparative statistical techniques in maintenancemanagement. IMA Journal of Mathematics Applied in Business and Industry1992;3:2418.

    [22] Ascher HE, Feingold H. Repairable systems reliability. Modeling, inference,misconceptions and their causes. New York: Marcel Dekker; 1984.

    [23] Saldanha PLC, de Simone EA, Frutoso e Melo PF. An application of non-homogeneus Poisson point processes to the reliability analysis of servicewater pumps. Nuclear Engineering and Design 2001;210:12533.

    [24] Weckman GR, Shell RL, Marvel JH. Modeling the reliability of repairablesystems in the aviation industry. Computers and Industrial Engineering2001;40:5163.

    [25] Rigdon SE, Basu AP. Statistical methods for the reliability of repairablesystems. New York: Wiley; 2000.

    [26] Thompson WA. On the foundations of reliability. Technometrics 1981;23:113.

    [27] Calabria R, Pulcini G. Inference and test in modeling the failure/repair processof repairable mechanical equipments. Reliability Engineering and SystemSafety 2000;67:4153.

    [28] Brown M, Proschan F. Imperfect repair. Journal of Applied Probability1983;20:8519.

    [29] Kijima M. Some results for repairable systems with general repair. Journal ofApplied Probability 1989;26:89102.

    [30] Coetzee J. The role of NHPP models in the practical analysis of maintenancefailure data. Reliability Engineering and System Safety 1997;56:1618.

    [31] Kvaloy JT, Lindqvist BH. TTT-based tests for trend in repairable systems data.Reliability Engineering and System Safety 1998;60:1328.

    [32] Miller AG, Kaufer B, Carlsson L. Activities on component reliability under theOECD Nuclear Energy Agency. Nuclear Engineering and Design 2000;198:32534.

    [33] Ascher HE, Kobbacy KAH. Modelling preventive maintenance for repairablesystems. IMA Journal of Mathematics Applied in Business and Industry1995;6:85100.

    [34] Baker RD. Some new tests of the power law process. Technometrics 1996;38:25665.

    [35] Knights PF, Segovia R. Reliability model for the optimal replacement of shovelcables. Transactions of the Institution of Mining and Metallurgy, Section A:Mining Industry 1999;108:A8A16.

    [36] Bergman B. On age replacement and the total time on test concept.Scandinavian Journal of Statistics 1979;6:1618.

    [37] Xie M, Goh TN, Ranjan P. Some effective control chart procedures forreliability monitoring. Reliability Engineering and System Safety 2002;77:14350.

    [38] Elvebakk G. Extending the use of some traditional trend tests for repairablesystems data by resampling techniques, 1999. /http://www.math.ntnu.no/preprint/statistics/1999/S19-1999.psS.

    [39] Mann HB. Nonparametric tests against trend. Econometrica 1945;13:24559.[40] Lewis PA, Robinson DW. Testing for monotone trend in a modulated renewal

    process. In: Proschan F, Sering RJ, editors. Reliability and biometry.Philadelphia: SIAM; 1974. p. 16382.

    [41] Anderson TW, Darling DA. Asymptotic theory of certain goodness of tcriteria based on stochastic processes. Annals of Mathematical Statistics1952;23:193212.

    [42] Vaurio JK. Identication of process and distribution characteristics by testingmonotonic and non-monotonic trends in failure intensities and hazard rates.Reliability Engineering and System Safety 1999;64:34557.

    [43] Tukey JW. The future of data analysis. Annals of Mathematical Statistics1962;33:167.

    [44] Walls L, Quigley J. Building prior distributions to support Bayesian reliabilitygrowth modelling using expert judgement. Reliability Engineering andSystem Safety 2001;74(2):11728.

    [45] Guikema SD, Pate-Cornell ME. Probability of infancy problems for spacelaunch vehicles. Reliability Engineering and System Safety 2005;87:30314.

    [46] Viertava J, Vaurio JK. Testing statistical signicance of trends in learning,ageing and safety indicators. Reliability Engineering and System Safety2009;94:112832.

    ARTICLE IN PRESS

    D.M. Louit et al. / Reliability Engineering and System Safety 94 (2009) 161816281628