Upload
jenny-katherine-l-henson
View
212
Download
0
Embed Size (px)
Citation preview
7/31/2019 Program Evaluation and Management
1/28
PROGRAM EVALUATION ANDMANAGEMENTJoining Theory and Practice
By: Jenny Katherine L. Henson, R.N.
7/31/2019 Program Evaluation and Management
2/28
Program Evaluation
Is intended to be a flexible andsituation specific means ofanswering questions, testinghypotheses, or describing programprocesses.
Can be formative or summative.
Its purposes clearly affect
relationships between evaluators andmanagers.
7/31/2019 Program Evaluation and Management
3/28
Program Evaluation
Generally, managers are more likelyto view formative evaluations as
friendly evaluations, and are morelikely to be willing to cooperate withthe evaluators.
7/31/2019 Program Evaluation and Management
4/28
Program Evaluation
From an evaluators standpoint, the
experience of conducting a formative
evaluation can be quite different fromconducting a summative evaluation.
7/31/2019 Program Evaluation and Management
5/28
Prospects for Building Culturesthat Support Evaluation
Instead of seeing evaluation as anactivity that challenges management,
they are encouraged to believe thatevaluators can work with managers todefine and execute evaluations thatcombine the best of what both parties
bring to that relationship.
7/31/2019 Program Evaluation and Management
6/28
Prospects for Building Culturesthat Support Evaluation
Utilization focused evaluation, forexample is premised on producing
evaluations that managers and otherstakeholders will use and ensuringthat means developing a workingrelationship between evaluators and
managers.
7/31/2019 Program Evaluation and Management
7/28
Prospects for Building Culturesthat Support Evaluation
Patton (1997) characterizes the role ofevaluator as the one who facilitates judgmentand decision making by intended users
rather than acting as a distant, independentjudge. Since no evaluation can be valuefree, utilization focused evaluation answers
the question of whose values will frame theevaluation by working with clearly identified,primary intended users who haveresponsibility to apply evaluation findings and
implement recommendations.
7/31/2019 Program Evaluation and Management
8/28
Love (1993) outlined six stages in thedevelopment of internal evaluation
capacity Ad hoc evaluations focused on single programs
Process focused regular evaluations
Goal setting, measurement of program outcomemonitoring, adjustment
Evaluations of program effectiveness, improving
organizational performance Evaluations of technical efficiency and cost
effectiveness
Cost benefit analyses
7/31/2019 Program Evaluation and Management
9/28
Learning Organizations asSelf Evaluating Organizations
Morgan (1997), in Images of Organizations,elaborates a organizational metaphor that
suggests that the organization can be seenas a brain. Within that broad metaphor, heelaborates a metaphor for learning
organizations.
7/31/2019 Program Evaluation and Management
10/28
Learning Organizations asSelf Evaluating Organizations
Using the work of Senge (1990), Morgan suggests
that learning organizations must developcapacities to:
Scan and anticipate in the wider environment todetect significant variations
Develop an ability to question, challenge, andchange operating norms and assumptions
Allow an appropriate strategic direction andpattern of organization to emerge
7/31/2019 Program Evaluation and Management
11/28
Learning Organizations asSelf Evaluating Organizations
Key to estabishing a learning organizationis what Morgan (1997) calls double - loop
learning; that is learning how to learn. Garvin (1993) has suggested five building
blocks for creating learning organizations.
By reviewing these steps, one can see akey role for evaluations.
7/31/2019 Program Evaluation and Management
12/28
Building Blocks for Creating Learning Organizations
1. Systematic problem solving
Tackling problems using a sequence of hypothesis generating hypothesis testingactions
Insisting on data rather than assumptions
Attention to details
2. Experimentation
Small, controlled modifications and tests of existing programs
Searching for and testing new knowledge
Managers must have both the incentives and skills to experiment
3. Learning from past experience
Systematically recording, displaying, and reviewing the evidence from pastperformance
Both this information and the skills to use and interpret it need to be widely distributedin the organization
4. Learning from others
Steal ideas shamelessly Find out who is the best learn why there are, and adapt their practices to your
organization
5. Transferring knowledge
Knowledge must be spread quickly and efficiently throughout the organization
Knowledge is treated as a resource
7/31/2019 Program Evaluation and Management
13/28
Learning Organizations asSelf Evaluating Organizations
Fetterman (Fetterman, 2001; Fetterman,Kaftarian, & Wndersman, 1996) argues thatone way to contribute to the development of
a learning organization is through theprocess of empowerment evaluation.
Empowerment evaluation is defined as theuse of evaluation concepts, techniques, and
findings to help program managers and staffevaluate their own programs and thusimprove practice and foster selfdetermination in organizations.
7/31/2019 Program Evaluation and Management
14/28
Learning Organizations asSelf Evaluating Organizations
Empowerment evaluation can only besuccessful in the right kind of organizational
environment, one which is guided by acommitment to truth and honesty.
7/31/2019 Program Evaluation and Management
15/28
Can Program Managers EvaluateTheir Own Programs?
7/31/2019 Program Evaluation and Management
16/28
Can Program Managers EvaluateTheir Own Programs?
Clearly, expecting managers to evaluatetheir own programs, given the incentivesalluded to above, can result in biased
program evaluations. Love (1993) envisions evaluators working
closely with program managers to produce
evaluations on issues that are of directrelevance to managers.
7/31/2019 Program Evaluation and Management
17/28
Can Program Managers EvaluateTheir Own Programs?
Patton (1997) stresses that among thefundamental premises of utilization focusedevaluation, the first is the commitment to
working with the intended users to ensure thatthe evaluation actually gets used.
Fetterman (2001) argues that the best data aresecured through close interaction andobservation with program managers and staff,because they are typically the mostknowledgeable about their program and its
strengths and weaknesses.
7/31/2019 Program Evaluation and Management
18/28
Can Program Managers EvaluateTheir Own Programs?
An important dissenting voice in the chorus thatadvocates evaluator manager contact and evencollaboration is Scrivens (1997) view that program
evaluators should keep distance from theorganizations and people with whom they work.
Getting too close to program managers amountsto compromising the objectivity of the evaluation
process, and undermines the key contribution thatevaluators can make: speaking the truth andoffering an unbiased view of a program.
7/31/2019 Program Evaluation and Management
19/28
Can Program Managers EvaluateTheir Own Programs?
Objectivity has been a criterion for highquality evaluations historically (Office of the
Comptroller General of Canada, 1981), andcontinues to have a scientific appeal to
practitioners and clients.
7/31/2019 Program Evaluation and Management
20/28
Striving for Objectivity in ProgramEvaluations
For Scriven (1997), objectivity is defined as withbasis and without bias and an important part of
being able to claim that an evaluation is objective isto maintain distance between the evaluator andwhat is being evaluated.
Objectivity has a certain cachet, and as apractitioner, it would be appealing to be able to
assert prospective clients that ones work will beobjective.
7/31/2019 Program Evaluation and Management
21/28
Criteria for Best Practices in ProgramEvaluation: Assuring Stakeholders that
Evaluations are High Quality
A review of several of these guideline documentsindicates that there are no specific mentions ofobjectivity among the criteria suggested for good
evaluations (AERA, 2000; American EvaluationAssociation, 1995; Australasian EvaluationSociety, 2002;Organization for Economic
Cooperation and Development, 1998).
7/31/2019 Program Evaluation and Management
22/28
It would seem that although some programevaluators and perhaps clients/stakeholders are
prepared to make objectivity a criterion for soundpractice, the evaluation profession as a whole isnot.
Professional evaluation organizations tend tomention the accuracy and credibility of evaluationinformation, the honesty and integrity of evaluators
and the evaluation process, the completeness andfairness of evaluation assessments and the validityand reliability of evaluation information.
7/31/2019 Program Evaluation and Management
23/28
In addition, professional guidelines emphasize theimportance of declaring and avoiding conflicts ofinterest and the importance of impartiality inreporting findings and conclusions. As well asguidelines tend to emphasize the importance ofcompetence in conducting evaluations and the
importance of upgrading evaluation skills.
Collectively, these guidelines cover many of thecharacteristics of evaluators and evaluations that
we might associate with objectivity it is a processthat involves corroboration of ones findings by
ones peers.
7/31/2019 Program Evaluation and Management
24/28
Ethics and Evaluation Practice
The evaluation guidelines, standards, andprinciples that have been developed for theevaluation profession all speak, in different ways,to ethical practice. Although evaluation practice isnot guided by a set of professional norms that areenforceable, ethical guidelines are an importantreference point for evaluators.
7/31/2019 Program Evaluation and Management
25/28
Ethics and Evaluation Practice
Newman and Brown (1996) have undertaken anextensive study of evaluation practice to establishethical principles that are important for evaluatorsin the roles they play. Underlying their work areethical principles, which they trace to Kitcherners
(1984, 1985) discussions of ethical norms.
Relationships Bet een the Ameri an E al ation Asso iation Prin iples and Ethi al Prin iples for E al ation
7/31/2019 Program Evaluation and Management
26/28
American Evaluation Association Guiding Principles
Systematic inquiry
Evaluators conduct systematic, data
basedinquiries about the subject of evaluation
Competence
Evaluators provide competent performance to
stakeholders
Integrity/honesty
Evaluators ensure the honesty and integrity of the
entire evaluation process
Respect for people
Evaluators respect the security, dignity, and self
worth of the respondents, program participants,clients, and other stakeholders with whom they
interact
Responsibilities for the general and public welfare
Evaluators articulate and take into account the
diversity of interests and values that may be related
to the general and public welfare
Ethical Principles for Evaluators
Maximizing Benefits
Minimizing harmsBalancing harms and benefits
Minimizing harms
Being honest
Keeping promises
No conflict of interest
Free and informal consent
Privacy and confidentiality
Respect for vulnerable persons
Procedural justice ethical reviews of projects
are fair, independent and transparent
Distributive justice persons are not
discriminated against, and there is respect for
vulnerable persons
Relationships Between the American Evaluation Association Principles and Ethical Principles for Evaluation
7/31/2019 Program Evaluation and Management
27/28
Reference:
Mc. David, Hawthorn Program Evaluation& Performance Measurement: An
Introduction to Practice, 2006, ThousandOaks, California
7/31/2019 Program Evaluation and Management
28/28