This book addresses the challenges of conducting program evaluations in real-world contexts where evaluators and the agencies face budget and time constraints and where critical data is missing. The book is organized around a seven-step model developed by the authors, which has been tested and refined in workshops. Vignettes and case studiesrepresenting evaluations from a variety of geographic regions and sectorsdemonstrate adaptive possibilities for small projects with budgets of a few thousand dollars to large-scale, long-term evaluations. The text incorporates quantitative, qualitative, and mixed-method designs and this Second Edition reflects important developments in the field over the last five years.
|Edition description:||Second Edition|
|Product dimensions:||6.90(w) x 9.90(h) x 1.50(d)|
About the Author
Michael Bamberger has been involved in development evaluation for fifty years. Beginning in Latin America where he worked in urban community development and evaluation for over a decade, he became interested in the coping strategies of low-income communities, how they were affected by and how they influenced development efforts. Most evaluation research fails to capture these survival strategies, frequently underestimating the resilience of these communities – particularly women and female-headed households. During 20 years with the World Bank he worked as monitoring and evaluation advisor for the Urban Development Department, evaluation training coordinator with the Economic Development Department and Senior Sociologist in the Gender and Development Department. After retiring from the Bank in 2001 he has worked as a development evaluation consultant with more than 10 UN agencies as well as development banks, bilateral development agencies, NGOs and foundations. Since 2001 he has been on the faculty of the International Program for Development Evaluation Training (IPDET). Recent publications include: (with Jim Rugh and Linda Mabry) Real World Evaluation: Working under budget, time, data and political constraints (2012 second edition); (with Marco Segone) How to design and manage equity focused evaluations (2011); Engendering Monitoring and Evaluation ( 2013 ); (with Linda Raftree) Emerging opportunities: Monitoring and evaluation in a tech-enabled world (2014); (with Marco Segone and Shravanti Reddy) How to integrate gender equality and social equity in national evaluation policies and systems (2014).
Jim Rugh has had 41 years of professional involvement in rural community development in Africa, Asia, and Appalachia. He has specialized in evaluation for 25 yearsthe past 10 years as head of Design, Monitoring and Evaluation for CARE International, a large nongovernmental organization (NGO). His particular skills include promoting strategies for enhanced capacity for evaluation throughout this worldwide organization. He is a recognized leader in evaluation among colleagues in the international NGO community, including Inter Action. He has been an active member of the American Evaluation Association since 1986, currently serving on the Nominations and Election Committee. He was a founding member of the Atlanta-area Evaluation Association. He has experience in promoting community development and evaluating and facilitating self-evaluation by participants in such programs. He has provided training for and/or evaluated many different international NGOs. He brings a perspective of the “big picture,” including familiarity with a wide variety of community groups and assistance agencies in many countries, plus an eye to detail and a respect for inclusiveness and the participatory process.
Linda Mabry is a faculty member at Washington State University specializing in program evaluation, student assessment, and research and evaluation methodology. She currently serves as president of the Oregon Program Evaluation Network and on the editorial board for Studies in Educational Evaluation. She has served in a variety of leadership positions for the American Evaluation Association, including the Board of Directors, chair of the Task Force on Educational Accountability, and chair of the Theories of Evaluation topical interest group. She has also served n the Board of Trustees for the National Center for the Improvement of Educational Assessments and on the Performance Assessment Review Board of New York. She has conducted evaluations for the U.S. Department of Education, National Science Foundation, National Endowment for the Arts, the Jacob Javits Foundation, Hewlett-Packard Corporation, Ameritech Corporation, ATT-Comcast Corporation, the New York City Fund for Public Education, the Chicago Arts Partnerships in Education, the Chicago Teachers Academy of Mathematics and Science, and a variety of university, state, and school agencies. She has published in a number of scholarly journals and written several books, including Evaluation and the Postmodern Dilemma (1997) and Portfolios Plus: A Critical Guide to Performance Assessment (1999).
Table of Contents
Part 1. The Real World Evaluation ApproachChapter 1. Overview: Real World Evaluation and the Contexts in Which It Is UsedChapter 2. [Step 1] Scoping the Evaluation: First clarify the purposeChapter 3. [Step 2] Budget Constraints: Not enough moneyChapter 4. [Step 3] Time Constraints: Addressing scheduling and other time constraintsChapter 5. [Step 4] Data Constraints: Critical information is missing or difficult to collectChapter 6. [Step 5] Addressing Political Influences: Reconciling different priorities and perspectivesChapter 7. [Step 6] Strengthening the Evaluation Design and the Validity of ConclusionsChapter 8. [Step 7] Making it Useful: Helping clients and other stakeholders use the evaluation findings and recommendationsPart 2. A Review of Evaluation Methods and Approaches and their Application in Real World Evaluation: For those who would like to dig deeper on particular evaluation topicsChapter 9. Ensuring competent and ethical practice in the conduct of the evaluationChapter 10. Theory-based evaluationChapter 11. Evaluation designsChapter 12. Quantitative evaluation approachesChapter 13. Qualitative evaluation approachesChapter 14. Mixed method evaluationsChapter 15. SamplingChapter 16. Evaluating complex, multi-component development interventionsPart 3. Organizing and Managing Evaluations and Strengthening Evaluation Capacity: For readers involved with the funding and management of evaluationsChapter 17. Organizing and managing evaluationsChapter 18. Strengthening evaluation capacityChapter 19. Conclusions and challenges and the road ahead