Upload
joel-miles
View
224
Download
5
Tags:
Embed Size (px)
Citation preview
Theoretical Basis—Systems Theory• Systems Theorist—Burns and Stalker
(1972), Azumi and Hage (1972), Lincoln (1985) , Gharajedaghi (1985), Morgan (1986).
• Systems theorist in education—(Henry Bernard, Horace Mann, William Harris, Carleton Washburne).
• Mechanical/linear constructions of the world versus organic/systems constructions.
• Closed versus Open systems. The role of the environment.
References: Patton, 2002: Fitzpatrick, Sanders, & Worthen, 2004: Scott, 2003.
Defining and conceptualizing a system
• A system is a whole that is both greater than and different from its parts
• The effective management of a system requires managing the interactions of its parts, not the action of its parts taken separately (Gharajedaghi and Ackoff, 1985).
Describe that system—volunteers!
Reference: Patton, 2002.
Management Oriented Evaluation
• The primary focus of management oriented evaluation is to serve the decision-maker (s).
• The needs of the decision-makers guide the direction of evaluation.
Stufflebeam’s CIPP• CIPP serves decision-makers facing
four types of decisions:
Context Evaluation—planning decisions
Input Evaluation—structuring decisions
Process Evaluation—implementation decisions
Product Evaluation—recycling decisions
Context EvaluationObjective To define context, identify
target population, assess needs, diagnose problems
Questions to ask
What are the needs to be addressed?Who is the target population? Any existing problems?
Method System analysis, survey, document review, interviews, hearings, tests, Delphi technique
Relation to decision-making
Decide on setting, goals, planning, a basis for judging outcomes
Input EvaluationObjective To assess system capability,
alternative strategies, procedural designs, budgets, and schedules
Questions to ask
What resources are available?What plan has the best potential? What alternatives should be considered?
Method Inventory, literature review, visits to other programs, advocate teams, and pilot trials
Relation to decision-making
Select sources of support, solutions, and procedural designs. A basis for judging implementation
Process Evaluation
Objective To identify/predict defects in design or implementation, record/judge procedural events/activities.
Questions to ask
How well is the plan being implemented? What barriers threaten its success? What revisions are needed?
Method Monitoring, describing process, interacting, observing
Relation to decision-making
Effecting process control, save information for future use in interpreting outcomes
Product EvaluationObjective To judge outcomes, relate
to objectives, context, inputs and process, interpret worth
Questions to ask
What results are obtained? Were the goals met? What should be done with the program after it has run its course?
Method Defining criteria, stakeholders’ judgments, qualitative and quantitative analyses.
Relation to decision-making
Deciding to continue, terminate, modify, or refocus. Present record
Stufflebeam’s Evaluation Steps
• Focusing the evaluation• Collection of information• Organization of information• Analysis of information• Reporting of information• Administration of the evaluation
Alkin’s UCLA Model
• Types of evaluation– Systems assessment (context)– Program planning (input)– Program implementation– Program improvement (process)– Program certification (product)
Strengths• Focuses on informational needs and
pending decisions of decision-makers• Systematic and comprehensive• Provides a wide variety of information• Stresses importance of utility of
information• Evaluation happens throughout the
program’s life• Provides timely feedback and
improvement• CIPP—heuristic tool that helps
generate important questions to be answered. Easy to explain.
Weaknesses• Narrow focus
– Inability to respond to issues that clash with concerns of decision makers
– Indecisive leaders unlikely to benefit• Possibly unfair or undemocratic
evaluation• May be expensive and complex• Unwarranted assumptions
– Important decisions may be correctly identified up front
– Orderliness and predictability of decision-making process
Application of CIPP• Evaluation framework for nursing
education programs: application of the CIPP model
• Critical success factors– Create an evaluation matrix– Form a program evaluation
committee including representatives from all partners
– Determine who will conduct the evaluation: internal or external
– Ensure the evaluators understand and adhere to the program evaluation standards
It’s Time for…..
• Management Oriented Evaluation Trivia.– Split up into two groups.– Each group selects a team
captain.– The group with the most money
wins.
References• Fitzpatrick, J.L., Sanders, J.R., & Worthen, B.R. (2004).
Program evaluation: alternative approaches and practical guidelines. White Plains, NY: Longman.
• Patton, M. Q. (2002). Qualitative evaluation and research methods (3rd ed.). Newbury Park, California: Sage.
• Scott, W. R. (2003). Organizations: Rational, natural, and open systems (5th ed.).Upper Saddle River, NJ: Prentice Hall.
• Singh, M.D. (2004). Evaluation framework for nursing education programs: Application of the CIPP model. International Journal of Nursing Education Scholarship, 1, Issue 1.