Human Error & Risk Factor Affecting Reliability & Safety

Preview:

Citation preview

SEMINAR

ON

“HUMAN ELEMENT FACTOR AFFECTING RELIABLITY & SAFETY-Treating Human errors

and risk factors in probabilistic analysis”

BY

DUSHYANT KALCHURI

M.TECH(PRODUCTION ENGG.)

DEPARTMENT OF

MECHANICAL ENGINEERING

CONTENTS :• Introduction• Faulty measures undertaken during the handling of equipment

And Unsafe Acts• Case Study On:

Computer System Interruptions By Human Errors

Utility Distributions Interruptions By Human Errors• Classification of Human Error according to system orientation• Human Reliability analysis• Technique for Human Error Rate Prediction• Error Prevention / Remediation• Accident Injury Sequence Model• Safety analysis• References

INTRODUCTION:• Many system reliability predictive methods are based

solely on equipment failures, neglecting the human component of man–machine systems (MMS).

• The reliability and safety of industrial and commercial power systems and processes (i.e., MMS) are dependent upon human characteristics and many dependent and dynamic interactive factors .

• The consequences of human errors are very diverse and can range from damage to equipment and property, injury to personnel or fatalities, to disruption of scheduled system operation, all of which represent a significant cost to society.

HUMAN ERROR:

• A failure on the part of the human to perform a prescribed act or task within specified limits of accuracy, sequence, or time, which could result in damage to equipment and property and disruption of scheduled operations or have no consequences at all.

• “ most of the human errors occur because humans are capable of doing so many different things in many diverse ways.”

• Generally 20%–50% of all equipment failures are due to human errors.

WHY A HUMAN PERFORMANCE IMPROVEMENT APPROACH??

80% Human Error 30% Individual

20% Equipment Failures

Human Error

Unwanted Outcomes70% Latent Organization Weaknesses

INDUSTRY EVENT CAUSESDUE TO HUMAN PERFORMANCE/ERROR

Source: INPO, Event Database, March 2000. For all events during 1998 and 1999.

215

26 3988

192

654

9 20

160

82

806

73118

0

100

200

300

400

500

600

700

800

900

Num

be

r o

f C

ause

s

1,676 = Org behavior (68%)

806 = Individual behavior (32%)

TAXONOMY OF HUMAN ERROR:

Interpretation

Situation Assessment

Plan

Intention ofAction

Action Execution

StimulusEvidence

Memory

MISTAKES SLIPS

LAPSES &MODE ERRORS

Knowledge Rule

TAXONOMY OF HUMAN ERROR MISAKES:

• Mistakes – failure to come up with appropriate solution• Takes place at level of perception, memory, or

cognition• Knowledge-based Mistakes – wrong solution because

individual did not accurately assess the situation.• Caused by poor heuristics/biases, insufficient info,

info overload• Rule-based Mistakes – invoking wrong rule for given

situation• Often made with confidence

TAXONOMY OF HUMAN ERRORSLIPS:

• Slips – Right intention incorrectly executed (oops!)• Capture errors – similar situation elicits action, which

may be wrong in “this” situation. Likely to result when:• Intended action is similar to routine behavior

• Hitting enter key when software asks, “sure you want to exit without saving?”

• Either stimulus or response is related to incorrect response• Hit “3” instead of “#” on phone to hear next

message, because “3” is what I hit to hear the first message

TAXONOMY OF HUMAN ERRORLAPSES & MODE ERRORS:

• Lapses – failure to carry out an action• Error of Omission (working memory)– Examples: Forgetting to close gas cap, failure to

put safety on before cleaning gun, failure to remove objects from surgical patient

• Mode Errors – Making the right response, but while in the wrong mode of operation• Examples: leave keyboard in shift mode while

trying to type a numeral, driving in wrong gear, going wrong direction because display was north-up when thought it was nose-up

FAULTY MEASURES UNDERTAKEN DURING THE HANDLING OF

EQUIPMENTS:1. Loose connections;

2. Faulty installation;

3. Improper grounding;

4. Defective parts;

5. Ground faults in equipment;

6. Unguarded live parts.

These Conditions lead to plant interruptions and disruption of processes and degrades Reliability.

• According to many safety and health laws, employers must provide a workplace where workers will not be exposed to hazards, where practicable.

• Workers must receive training, instruction, supervision, and information so they are not exposed to hazards.

Examples of FAULTY/ UNSAFE acts:

1. Failure to de-energize, Lockout and Tag-Out hazards during maintenance, repair, or inspections;

2. Use of defective and unsafe tools;

3. Use of tools or equipment too close to energized parts;

4. Etc..

CASE STUDIES ON THE FREQUENCY OF HUMAN ERRORS:

• Computer System Interruptions Caused By Human Errors

• Utility Distributions Interruptions By Human Errors

COMPUTER SYSTEM INTERRUPTION CAUSED BY HUMAN ERROR:

• A ten-year study at the University of Alberta’s central computer system was conducted analyzing the frequency of computer system interruptions caused by operator errors.

• A human or computer operator error is defined as an act or set of acts which results in a computer system interruption, and the system is restored to an operational state either by initial program loading or restarting.

• The computer system runs continuously 24 hours a day, except for maintenance periods early in the mornings on the weekends to minimize the impact of the scheduled interruptions on the users.

• The annual number of computer system interruptions caused by operator errors is shown in Fig (Next Slide).

YEAR OF STUDY:

• The total number of computer system interruptions caused by operator errors per year averaged approximately 25.

• Table 1 reveals the various percentages of interruptions attributed to the primary causes of computer system interruptions in which operators errors accounted 7.4% of computer system interruptions.

1. DAY OF THE WEEK OF COMPUTER SYSTEM INTERRUPTION:

• A ten year study of the average frequency of computer system interruptions per given day of the week confirmed belief of operators, as is shown in Fig. 2.

• The average frequency of computer interruptions was higher during the “weekdays” (i.e., Monday through Friday) than on the weekends, when the system loading was reduced.

• This supported the operators belief that “weekdays” were more prone to computer system interruptions than Saturday and Sunday.

DAY OF WEEK

2. TIME OF THE DAYOF COMPUTER SYSTEM INTERRUPTION:

• Many users of the computer system claimed that there appeared to be more interruptions in the morning than during the remainder of the day.

• The loading on the system peaked between 8–9 a.m., and the load dropped off between 4–5 p.m. and remained fairly steady for the remaining time periods.

• It is clear that the sudden increase in computer system loading and operator stress between 8–9 a.m. was directly correlated with a significant increase in the frequency of operator errors resulting in computer system interruptions.

TIME OF DAY

UTILITY DISTRIBUTION INTERRUPTION CAUSED BY HUMAN ERROR:

• The electric utility distribution system customer interruptions were recorded for the past 30 years by the Canadian Electricity Association (CEA) in Canada.

• It can be seen that the human element accounts for approximately 1.7% of the total number of distribution system interruptions.

• Other factors, such as scheduled outages, lightning, and defective equipment were the dominant causes of distribution system interruptions.

ELECTRIC UTILITY LOST TIME DUE TO INJURY ACCIDENTS:

• To measure the impact of injury accidents on productivity in terms of hours in the workplace, the CEA uses an index called the severity rate.

• The severity rate equals the number of calendar days lost due to injury accidents per millions of hours worked. Typical rates are shown in Fig. 5

• The severity rate remains fairly constant for several years, averaging about 500 days lost per million hours worked.

SEVERITY RATE

CLASSIFICATION OF HUMAN ERRORS ACCORDING TO SYSTEM

ORIENTATION:

• Human errors can occur at any stage in the life of a system.

• It occurs from the original design inadequacies, to installation deficiencies and operating and maintenance human anomalies.

Continued…

Design

Installation

Assembly

Inspection

Operating

Maintenance

Classification Of Human

Error

DESIGN ERROR:

• It can be attributed to the physical structure of a system with basically the following three types of inadequacies:

1. failure to implement human needs in the design.

2. assigning inappropriate functions to persons, e.g., lack of definition of primary work tasks;

3. failure to ensure the effectiveness of the man and machine component interactions.

• INSTALLATION ERROR:• This are primarily due to the failure to install

equipment by humans according to instructions or blueprints, assuming these drawings are correct, and poor workmanship when operating under severe time constraints.

• The inspection criteria of evaluation is dependent

upon the inspector’s knowledge of the system and the relation between its interacting parts.

• According to study an average inspection effectiveness is close to 85%.

• INSPECTION ERRORS:

ASSEMBLY ERROR:

• This errors are the result of poor workmanship. These errors are often discovered after the installation process when they disrupt scheduled system operations.

• Examples are:

1) use of incorrect component;

2) use of incorrect tools;

3) omitting a component;

4) improper connections;

5) improper handling of equipment.

OPERATION ERRORS:• This error is subject to human operating errors.

Situations that lead to these errors are as follows:

1) lack of proper procedures;

2) task complexity and overload conditions;

3) poor personnel selection and training;

4) operator carelessness and lack of interest;

5) poor environmental conditions.

MAINTENANCE ERROR:

• This errors are primarily due to the incorrect repair/replacement/service activities of equipment.

• Examples of maintenance errors are the following:

1) incorrect calibration of instruments, e.g., relays, computer controls, and sensors;

2) failure to follow maintenance schedules and procedures;

3) incorrect equipment cleaning procedures.

HUMAN RELIABILITY ANALYSIS

• Human Reliability Analysis – predict reliability of system in terms of probability of failure or mean time between failures (MTBF) when system is designed to work in parallel or series

.9 .9

.9

.9

Series

Parallel

Reliability = .9 x .9 = .81P(failure) = 1 - .81 = .19

Reliability = 1 – [(1 - .9) (1 - .9)] = 1 - .01 = .99 P(failure) = 1 - .99 = .01

TECHNIQUE FOR HUMAN ERROR RATE PREDICTION (THERP)

THERP components

1. Human Error Probability• Ratio of errors made to possible errors

2. Event Tree• Diagram showing sequence of events

• Probability of success or failure for each component

3. Other Moderating Factors• May add in multiplier to account for variables such

as experience level, time, stress, etc.

THERP EVENT TREE

a A

b\a B\a

SS

b\A B\A

FS

FS

FF

SeriesParallel

Series:P[S] = a(b\a)P[F] = 1 – a(b\a) = a(B\a) + A(b\A) + A(B\A)Parallel:P[S] = 1 – A(B\A) = a(b\a) + a(B\a) + A(b\A)P[F] = A(B\A)

P(successful task B given A)

P(unsuccessful task B given A)

P(success of task B given a)

P(Unsuccessful task B given a)

P(successful task A) P(unsuccessful task A)

Task A = first task

Task B = second task

ERROR PREVENTION / REMEDIATION

1. Task Design – design tasks with working memory capacity in mind

2. Equipment Design a) Minimize perceptual confusions – ease of

discrimination• Ex: airplane controls that feel like what they do

(flaps, wheels)b) Make consequences of action visible – immediate

feedback• Ex: preview window in some software programs

c) Lockouts – design to prevent wrong actions• Ex: car that will not let you lock door from outside

without keyd) Reminders – compensate for memory failures

• Ex: ATM reminds you to take your card

ERROR PREVENTION / REMEDIATION(Cont.….)

3. Training – provide opportunity for mistakes in training, so can learn from them

• Ex: Simulation

4. Assists and Rules – checklists to follow• Ex: Pilot pre-flight checklist

5. Error-tolerant systems – system allows for error correction or takes over when operator makes serious error

• Ex: Undo button

ACCIDENT-INJURY SEQUENCE MODEL :

• This Model provide a framework for identifying the possible root cause of electrical accidents.

• This Model provide a basis for developing accidents prevention and injury control strategies to minimize

1. impact of disruption to system operation.

2. occurrences of injuries.

SAFETY ANALYSISSequence for identifying potential hazards and recommendations for hazard

reduction: (Weinstein et al. 1978)

1. Task Analysis – How will product be used?2. Environment Analysis – Where will product be used?3. User Analysis – Who will use product?4. Hazard Identification – What is likelihood of hazard

with product?5. Generate Methods for Hazard Control – What might

eliminate hazards?6. Evaluate Alternatives – How will alternative designs

affect product performance?7. Select Hazard Control – Given alternatives, what is best

design to minimize hazards?

ACCIDENT INVESTIGATION LEVELS OF CAUSES

Management Safety Policy & DecisionsPersonal Factors

Environmental factors

Unsafe Act

Unsafe Condition

Unplanned Release of EnergyAnd/or

Hazardous Material

ACCIDENTPersonal Injury

Property Damage

BASICCAUSES

INDIRECTCAUSES(SYMPTOMS)

DIRECTCAUSES

SAFETY PROGRAMS

1. Identify risks to the company identify hazards, hazard controls, accident

frequency, & company losses due to accidents/incident claims

2. Implement safety programs, includes: management involvement, accident

investigation, recommendations for equipment, safety rules, personal protective equipment, employee training, safety promotion

3. Measuring program effectiveness evaluated by assessing changes in safety

behaviors, accident/incident rates, number of injuries or death, and number of days off due to injury

CONCLUSION / SUGGESTION:• Risk-Taking as a Decision Process

• People must know a hazard exists, know what actions are available, & know the consequences of the safe behavior vs. alternative behaviors

• Written Warnings and Warning Labels• Accurately convey the hazards of a product • Should include a signal word, info pertaining to the

hazard, consequences, & necessary behavior• Danger: Immediate hazard likely results in severe

injury• Warning: Hazard could result in injury• Caution: Hazard or unsafe use my result in minor

injury

REFERENCES:

• Human Element Factors Affecting Reliability and Safety, Don O. Koval and H. Landis Floyd, IEEE Transactions on Industry Applications, Vol. 34.

• IRACST- International Journal of Research in Management & Technology (IJRMT), ISSN: 2249-9563 ,Vol. 2, No. 1, 2012,Human Reliability Analysis: A review of the state of the art

• 4th European-American Workshop on Reliability of NDE - Th.4.A.1, Integrating Human Factors in Safety and Reliability Approaches by Babette FAHLBRUCH, TÜV NORD SysTec, Berlin, Germany http://www.ndt.net/index.php?id=8338

Thank You!!!

Recommended