Upload
ashley-briggs
View
28
Download
0
Embed Size (px)
Citation preview
icfi.com |
Evaluation 2014: Visionary Evaluation for a Sustainable, Equitable Future
28th Annual Conference of the American Evaluation Association
Denver, CO, USA
Approaches to Mixed Methods in Education, Community, and Social Programs
Kristen Peterson, Chair
October 16, 2014 | 3:00 - 4:30 PM
As you arrive, please discuss with your neighbor what you have done or want to do with mixed methods in your own work.
icfi.com |
Session Goals
• Highlight a few program evaluation research projects at ICF
• Share processes for designing and conducting mixed methods research
• Offer examples of how mixed methods have provided added breadth and depth of understanding
• Identify next steps to improve the ways we use mixed methods to paint a holistic picture of programs
icfi.com |
Session Agenda
1. Weaving Together Outcome Data and Site Visit Findings for a National Evaluation of Drug-Free Communities, Kazuaki Uekawa
2. Evaluation of a State GEAR UP Program: Understanding Implementation through the Integration of Annual Performance Report, site visit and survey data, Ashley Briggs
3. The Diplomas Now Validation Study: Integrating case study, fidelity, and outcome data to assess program implementation, Kathleen Wang
4. Discussion/Q&A
icfi.com |
Weaving Together Outcome Data and Site Visit Findings for a National Evaluation of Drug-Free Communities Kazuaki Uekawa
October 16, 2014
Presented to:
American Evaluation Association 2014 Conference
icfi.com | 5
Drug-Free CommunitiesOVERVIEW
DFC— Drug-Free Communities
Funding and Support—Office of National Drug Control Policy (ONDCP), with support from the Substance Abuse and Mental Health Services Administration (SAMHSA)
Grantees ––Nearly 2,000 DFC grants across the nation in various locales (660 grantees in FY 14)
Target Area—Combined population of 114.3 million, or 37.0% of the U.S. population
Map downloaded from:http://www.whitehouse.gov/ondcp/drug-free-communities-support-program
icfi.com | 6
DFC Grantees
Award Amount—Grantees receive awards of up to $125,000 per year for up to five years per award (maximum of 10 award years)
Goal— (1) reduce substance use among youth and (2) increase collaboration in the community to address substance use and associated problems
Approach—Prevention efforts through environmental approaches
Community Coalition Sectors—Youth, parents, schools, law enforcement, business, media, government (local, state, and tribal), youth-serving organizations, law enforcement agencies, religious or fraternal organizations, civic and volunteer groups, healthcare professionals
OVERVIEW
icfi.com | 7
DFC INFOGRAPHIC
Infographic downloaded from:http://www.whitehouse.gov/ondcp/drug-free-communities-support-program
icfi.com | 8
DFC National Evaluation: ChallengesEVALUATION DESIGN AND METHODOLOGY
Challenge: Not an Experimental Design– Use a natural variation design to understand how and in what situations DFC
works through mixed-methods
Challenge: Stakeholders from different sectors interested in understanding their role in improving outcomes– Conduct 36 site visits (including interviews with members from all 12
sectors)
Challenge: Annual reports are heavily focused on outputs but need to integrate qualitative information– Solution: Draw out lessons learned from site visits in issue briefs (e.g.,
parent engagement, sustainability) that draw from multiple data sources. Newsletters are also under consideration.
icfi.com | 9
Data Sources, QuantitativeEVALUATION DESIGN AND METHODOLOGY
– Progress report data on the use of Tobacco, Alcohol, Marijuana, and prescription drugs: core measures of 30-day use, perception of risk, perception of parental disapproval, perception of peer disapproval (reported every other year via. COMET system)
– Progress report on target population demographics, membership capacity, coalition processes, planning, and implementation (reported twice a year via. COMET system). Open-ended questions included.
– Coalition Classification Tool (CCT): latent constructs such as leadership, collective efficacy, sustainability, etc.
Example of 30-day use survey
Example of strategy activity details collected as part of implementation data. Interim report 2012. (http://www.whitehouse.gov/sites/default/files/dfc_2012_interim_report_annual_report_-_final.pdf)
icfi.com | 10
Data Sources, QualitativeEVALUATION DESIGN AND METHODOLOGY
– Site visits (36): Initial set of 12 high-performing coalitions, and then sequenced by special issue (e.g., inner cities, border towns).• Period of data collection: 2011-2014• Number of Sites: 36• Participants: DFC staff, coalition members, youth, parents, and community stakeholders• Role-specific, Semi-structured Protocol Topics: Community Context; Coalition Capacity;
Strategies and Activities; Lessons and Recommendations; Challenges/Successes
– Open-ended responses from semi-annual progress reports (via. COMET system)• Examples
– Grantee Coalition Information Section: Please provide a brief summary of your coalition. This is your 'Elevator Speech.' (Coalitions describe target population, goals, activities, accomplishments, challenges, etc.)
– Challenges and Protective Assets Section: Please report any additional details about your challenges and protective assets that were not captured above.
– Assessment Activities Section: Please report any notable accomplishments related to assessment activities achieved during this reporting period.
icfi.com | 11
Example 1. Application of Mixed Methods (in measurement)EVALUATION DESIGN AND METHODOLOGY
The goal: Refine the service categories of strategy activities – informed by both qualitative and quantitative analysis
Result: Reduced 42 service categories/7 strategies to 14 to 3 orientations. This was a shift from a typology to a more clear delineation of service strategies that relate to outcomes.
Method– Recoded service categories based on qualitative understanding of site visit
data and legacy data– Descriptive analysis to identify the size of variance, Factor analysis to
understand the subgroups of items, Correlations analysis to understand cross-use of legacy service strategies by the new categories
icfi.com | 12
Service Strategies Refined
Notes: DFC Coalitions Strategies: A Strategy Orientation Approach to Understanding How DFC Grantees Utilize Strategies
Typology of 7 environmental strategies w/ 42 service categories
PI: Providing informationES: Enhancing skillsPS: Providing supportLP: Modifying/changing policiesC: Changing ConsequencesAB: Enhancing access/Reducing barriersPD: Physical Design
3 new orientations
Building Individual Capacity
Building Protective Capacity
Building Supportive Capacity
Examples:• Physical Design : Clean up &
beautification • Providing Support: You/family
community involvement
Two of the most frequently associated activities across coalitions are promoting youth / family community involving (a skills enhancement activity) and cleanup and beautification (a physical design activity). The orientation toward building individual capacity links the skills enhancement and the opportunity to apply those skills.
Examples: Providing support• Alternative drug-free social activity
events• Youth organizations/ drop-in centers• Organized youth recreation programs• Youth/family community involvement• Youth/family support groups
icfi.com | 13
Example 2: Application of Mixed Methods (in analysis)ANALYSIS
Issue briefs for DFC stakeholders that blend site visit data with COMET data– Summary of trends across all coalitions who submitted progress report data between
2011 and 2013, Rich descriptions from site visits of strategies used, factors for success, and continued challenges, Draft briefs focus on sustainability and parent involvement.
– Example: Parent Engagement Brief• Basic quantitative information (e.g., average number of parent members)• Parents are key but engagement was often reported as a challenge, especially
engaging parents beyond those who also represent another sector• Age of youth may impact engagement• Cultural competence• Summary of common challenges in engaging parents, a) Parents may link certain
life events with permissive attitudes (e.g., prom, graduation, quinceañera, branding), b) Helping can be perceived as attacking
• Examples of parent involvement practices including parent-led workshops and small group discussions
• Recommendations from the field to engage parents strong school partnerships, PTAs, etc.
icfi.com | 14
Example 3: Application of Mixed Methods (in analysis)ANALYSIS
Ad-hoc requests from ONDCP regarding emergent topics– In response to inquiries such as “What do we know about the rise in marijuana use?,”
the evaluation team conducted a literature review on marijuana use, scanned open-ended responses on strategies used to combat marijuana, and analyzed quantitative data on the number of coalitions that target marijuana use.
– Marijuana Overview Example• Marijuana prevalence of use and trends• COMET data on marijuana as a target substance• Emerging issues (e.g., legalization and descriminalization)• Common strategy to target marijuana (e.g., social media campaigns, education, targeting
parents’ awareness)
icfi.com | 15
Lessons Learned about Using Mixed MethodsCONCLUSION
Given the complexity of this evaluation, mixed-methods research offers an opportunity to gain a more holistic understanding of coalitions.
Quantitative data is used in our primary “impact” analyses, and qualitative data expands the breadth of that data for a deeper understanding of complex processes.
Open-ended text fields offered all grantees a place for them to tell their story without necessitating site visits to every coalition.
icfi.com | 16
Next Steps in the EvaluationCONCLUSION
Synthesize results from all site visits to offer a deep understanding about challenges, successes, and context influences (e.g., inner city, border town)
Conduct quantitative analyses that help contextualize the qualitative data, supported by qualitative write-ups that help contextualize the quantitative data
Continue producing issue briefs and reports that present various types of data cohesively
icfi.com |
Evaluation of a State GEAR UP Program: Understanding Implementation through the Integration of Annual Performance Report, site visit and survey data
Ashley Briggs
October 16, 2014
Presented to:
American Evaluation Association 2014 Conference
icfi.com | 18
GEAR UP Nationally
GEAR UP— Gaining Early Awareness and Readiness for Undergraduate Programs
U.S. Department of Education (USED) discretionary grant— State and local partnership grants
Program goals— Raise postsecondary participation, strengthen academic preparation, and increase postsecondary awareness
Cohort approach— Students from 7th grade through to postsecondary education
FY12 – 132 federal grants serving 647,772 students
OVERVIEW
icfi.com | 19
About the Texas GEAR UP State Grant (SG)OVERVIEW
Texas Education Agency FY12 state grant– Approx. $5M per year (2012-19)
Single cohort of students starting in Grade 7— students are in Grade 9 in 2014–15 (4 districts and 7 high schools)
District services– Increase students’ college readiness through support services (e.g., college
visits, summer learning, and tutoring)– Provide teacher professional development to support course rigor and
postsecondary planning (e.g., project-based learning with college visits)– Promote vertical alignment of core subject teachers across the grades
Statewide Services– Postsecondary information dissemination to students and families (state
GEAR UP website)– Online communication and teaching platform – Statewide coalition of state and local GEAR UP grantees
?
icfi.com | 20
About the Texas GEAR UP SG EvaluationEVALUATION DESIGN AND METHODOLOGY
The external evaluation is a longitudinal mixed-method study using a quasi-experimental design to:
Provide a formative evaluation of implementation (facilitators/barriers, promising practices, and recommendations)
Determine impact including short, intermediate, and long-term student outcomes
Identify impact on parent perceptions
Examine access to and use of statewide services
Understand cost, spending, and sustainability
icfi.com | 21
Data SourcesEVALUATION DESIGN AND METHODOLOGY
Extant Data– Documents: Texas GEAR UP SG Grant Application, Notice and Grant
Awards (NOGAs), and implementation plans– Student-level data: Demographics, attendance, high school course
completion, and graduation– School-level data: Profile information about campus-level
performance, staff, finances, and programs
Student Tracking System (Annual Performance Report – APR)– Format: Submission by 4 subgrantee districts using a prepopulated
spreadsheet – Topics: Advanced course-taking; Academic services; Student
services; Student events and attendance; Parent events and attendance; Teacher professional development and enrollment; Community partners
icfi.com | 22
Annual Performance Report (APR) DataEVALUATION DESIGN AND METHODOLOGY
Period of data collection: March 2014
Format: Submission by 4 subgrantee districts
Topics– Advanced course taking (baseline)– Student support services (tutoring, mentoring, counseling/advising)– Student events– Parent events– Summer programs– Teacher professional development– Vertical teaming– Community partners– Statewide Information Services
icfi.com | 23
Data Sources (cont.)EVALUATION DESIGN AND METHODOLOGY
Surveys with Parents and Students– Format: Online and paper-based versions in English and Spanish– Topics: Aspirations and expectations; Knowledge of financial aspects;
Knowledge of college requirements; Perceptions of Texas GEAR UP SG
Site Visits to Texas GEAR UP SG Schools– Format: 1-1.5-day visits including interviews and focus groups with school
staff, teachers, students, parents, and community partners– Topics: GEAR UP activities and events (school and statewide); Knowledge of
college requirements and financial aspects; Perceptions of Texas GEAR UP SG; Readiness for success in college
Interviews with Key Leaders from TEA and Partner Organizations– Format: Telephone interviews– Topics: Level of partner involvement; Perceptions of program; Progress on
statewide implementation
icfi.com | 24
Application of Mixed MethodsEVALUATION DESIGN AND METHODOLOGY
Comprehensive logic model depicts how program and school staff work together to promote college readiness.
Detailed crosswalk tracks the following:–All 65 evaluation questions–The data source(s), as specific items, to inform each question–Suggestions for the type of analysis
Using ATLAS.ti made data from these sources easily accessible to supplement quantitative findings: –Site visit data–Interview data–Documents
icfi.com | 25
Logical Model (Condensed)EVALUATION DESIGN AND METHODOLOGY
E.G.: student characteristics, school graduation rate, & parent education level
E.G.: Advances course taking, student services, professional development, & community partnership
E.G.: Participation levels, # of PD sessions, & # statewide materials
E.G.: Course completion, promotion rates, changes in vertical team meetings, & increased parent expectations
E.G.: proficiency on state tests, SAT scores, college credits, & perceptions of workshops
E.G.: average # of college applications, % enrolled in college, % in remedial courses
icfi.com | 28
Selected FindingsFINDINGS
APR data–Descriptive data: Mix of services across schools as well as levels of
student participation–Chi-square analyses: number of subject areas in which students were
being tutored (1 subject- 78%; 3 subjects- 5%)–ANOVA: average number of hours in tutoring (32 hours per student in
math)
?
icfi.com | 29
Selected Findings (cont.)FINDINGS
Site visit data–Details about how schools decided which students received tutoring
(diagnostic data or teacher input)–Clarity on who provided the services (local partners)–Understanding of implementation barriers/facilitators (training of tutors)–Clarification of APR data (virtual tutoring)
Survey data–Perceptions about tutoring:
• Students reported, on average that tutoring in any subject was slightly effective.
• Parents reported, on average that tutoring in any subject was mostly effective.
icfi.com | 30
Next Steps in the EvaluationCONCLUSION
Forthcoming reports will include analysis of – Service factors: Provision type, frequency of delivery, mix of
services, and quality of implemented activities– Qualitative themes: District and school level trends to understand
details about the implementation of services
Types of analysis may include – Hierarchical linear modeling (student, school, and district levels)– Cluster analysis– Comparisons using propensity score matching– Impact analysis and linkages to implementation
icfi.com | 31
Lessons Learned about Using Mixed MethodsCONCLUSION
Tools such as logic models, crosswalks, and ATLAS.ti are helpful ways to organize large amounts of data and streamline their integration
Confirmed importance of how multiple data sources can illuminate contradictions and offer different layers of detail
Appreciation for the labor-intensive nature of examining many dimensions of a program through multiple lenses
Acknowledgement of tradeoffs in delegating the work
?
32
http://www.tea.state.tx.us/WorkArea/linkit.aspx?LinkIdentifier=id&ItemID=25769807659&libID=25769807662#25769807659
The second annual implementation and comprehensive reports are forthcoming.
For more information, first annual implementation report is available at:
icfi.com | 33
The Diplomas Now Validation Study: Integrating case study, fidelity, and outcome data
October 18, 2014
Presented to:
American Evaluation Association 2014 Conference
Kathleen Wang
icfi.com | 34
500,000 kids who drop out of school every year come from just 12 percent of the nation’s high schools or 1,700 “dropout factories.”
icfi.com | 35
JHU found that students who are most at risk of dropping out can be identified through key indicators also known as the ABCs:
Poor AttendanceUnsatisfactory Behavior
Course failure in math and EnglishWhen any one of these off-track indicators is seen as early as 6th
grade, that child has only a 10-20% likelihood of graduating from high school.
Diplomas NowOVERVIEW
icfi.com | 36
Diplomas Now
Program Goal: Reduce the number of students who drop out and provide students with a college- and career-preparatory education.
OVERVIEW
icfi.com | 38
Diplomas Now Study
U.S. Department of Education awarded Johns Hopkins University an Investing in Innovation (i3) validation grant in 2010 for the Diplomas Now model.
The grant includes funds for a mixed method study which will be conducted through the end of 2015.
OVERVIEW
icfi.com | 39
Evaluation Design
Impact Study:• Experimental evaluation (or Randomized Control
Trial-RCT) measuring the impact on secondary school students' outcomes.
Implementation study: • Fidelity assessment at 32 schools examining fidelity of
implementation.• Case studies at seven schools in year 1 and nine
schools in year 2.
icfi.com | 40
Available: http://www.mdrc.org/sites/default/files/Diplomas_Now_First_Year_FR_0.pdf
Year 1 Evaluation Report
icfi.com | 42
Rationale for Mixed Methods
To corroborate and confirm findings that examine the same research questions (simultaneous triangulation).
Expand and elaborate using different methods to answer different research questions.
Creswell, JW, & Plano Clark, VL .(2011). Designing and conducting mixed methodsresearch. (2nd ed.). Thousand Oaks, CA: Sage.
icfi.com | 43
Implementation Study Research Question Matrix
RESEARCH QUESTION
Fidelity Assessment
Case Studies
How did the intervention as implemented compare to the intervention as planned?
X X
What were the factors that facilitated or challenged implementation of the intervention?
X X
How did treatment schools compare to control schools in their structures, programs, and practices?
X
What are the perceived benefits of the intervention?
X
What lessons can be learned from the implementation of the intervention?
X
Expand
Confirm
icfi.com | 47
“Mixing” Fidelity and Case Study Data
EXPAND AND ELABORATE
“According to FIDELITY DATA, Diplomas Now schools were less successful in adapting new curricula and implementing peer coaching models…Teachers from CASE STUDY schools expressed mixed feelings about the usefulness of the peer coaching. In addition, focus group and interview participants noted difficulty implementing the Diplomas Now curriculum.”
icfi.com | 48
“Mixing” Fidelity and Case Study Data
CORROBORATE and CONFIRM
“According to FIDELITY DATA, Diplomas Now schools were the most successful in implementing a tiered intervention model and adding student supports into the existing school structure…Similarly, CASE STUDY respondents most frequently discussed Tiered Intervention Model and Student Supports in the context of effective rather than challenging implementation.”
icfi.com | 50
Next Steps… Year 2 Evaluation Report
Focused on early impacts and updating implementation study data.
Documenting CONTEXTUAL FACTORS (using case study data) that had a strong influence on program implementation and understanding why we are seeing little change in some areas due to:
- School closures- School/program staff turnover- Budget cuts- Teacher layoffs
icfi.com | 51
Key Takeaways
MIX your findings don’t just layer them–Don’t make your reader have to do all the work!–Good mixed methods will interweave the findings throughout
the evaluation report to help draw those connections for the reader.
Logic model is the backbone of good mixed methods–It is critical when evaluating a complex model like DN to provide
a common thread for different types of data sources and analyses.
icfi.com | 52
Key Takeaways
The “gold standard”– RCTs are widely seen as the “gold standard” of research since they answer
the question, “does this program work?”– This doesn’t negate the importance of other methods, but further
necessitates their use.
Good evaluation is both an art and a science.– Our design and methods are the science, but the “mixing” of multiple
methods is the “art.”– Mixed method evaluations typically have multiple analysts that divide up
quantitative and qualitative analyses. Very important that analysts/authors meet regularly to discuss confirming and conflicting evidence and determine best outline for reporting.
icfi.com | 53
Year 1 Evaluation Reporthttp://www.mdrc.org/sites/default/files/
Diplomas_Now_First_Year_FR_0.pdf
Year 2 Evaluation Report (in draft)
Additional information on Diplomas Now Study
icfi.com | 54
Discussion Benefits of mixed methods include:
1) complementarity that clarifies results, 2) initiation that illuminates contradiction, and 3) development to inform other methods (Onwuegbuzie & Teddlie, 2003).
Strong methodologists who appreciate mixed methods design allow for stronger inferences (Tashakkori & Teddlie, 2003)
Examining the interconnectedness of multiple data sources offers a better understanding of the phenomena than either one alone (Creswell & Plano Clark, 2007).
Understanding a program, beyond outcomes, is essential in sustaining the interventions put in place in schools and communities.
CALL TO ACTIONReports that thoughtfully integrate data, make it more useful and accessible to
readers. Evaluators need to know the ART and SCIENCE of mixed methods.
icfi.com | 55
Continue the conversation with your comments and questions at:
http://padlet.com/abriggs5/zpbqovr8o8uo
AUDIENCE DISCUSSION
THANK YOU
icfi.com | 56
Contact Information
Program Evaluation
Kristen Peterson, Senior Associate, [email protected]
Drug-Free Communities
Kazuaki Uekawa, Technical Specialist, [email protected]
Texas GEAR UP State GrantAshley Briggs, Technical Specialist, [email protected]
Diplomas NowKathleen Wang, Expert Consultant, [email protected]