Entry
Reader's guide
Entries A-Z
Subject index
Utilization of Evaluation
Evaluation utilization (or evaluation use) refers to the way that evaluations and information from evaluations affect the programs on which evaluations are conducted. Were progressive changes made in a program as a consequence of an evaluation? Did an evaluation generate a new understanding of or a modification of beliefs about the program or aspects of the program?
The concern for evaluation utilization initially developed in the United States during the 1960s. There were many reasons for this, including the enormous growth in demand in these years for evaluations to be conducted. Major social programs were initiated that often mandated end-of-year evaluation reports. The subsequent rush to evaluate attracted many researchers who lacked awareness of the distinctions between research and evaluation and who failed to understand properly the contexts in which evaluations were to take place. The result was a multitude of evaluation reports that lacked relevance to program personnel. This was particularly true for evaluations conducted at the local program level, such as those in schools, school districts, mental health clinics, social welfare agencies, and the like. Thus, many evaluation reports sat on the shelf, fulfilling the requirement of evaluation but generally disregarded.
Although prior to the 1960s some conceptualization had been done on aspects of knowledge and research use, there had not been much interest in extending this to a consideration of evaluation use. The distinction between research and evaluation is an important one: Research refers to generalizable knowledge and thus potentially may be used very broadly and in a variety of contexts. Research may also have influence many years after it is completed. Evaluation refers to the activity of examining the process and the impact of a specific program at a specific time. Thus, evaluation use is restricted to the program under examination and, roughly, the time frame during which the evaluation is implemented. With an evaluation, program personnel are known, the specific context is understood, and the evaluator has a role in helping to define the particular areas to be evaluated. Further, evaluators can, through their actions, seek to define an evaluative role, relationships, and procedures that may enhance utilization. Consequently, use is potentially more achievable for evaluation than for research.
Research on Utilization
The concern expressed by Carol Weiss in 1972 for greater understanding of evaluation utilization ushered in a period in which great attention was paid to conducting the necessary research. In addition to Weiss, important studies were conducted by Michael Patton, Marvin Alkin and colleagues, Jean King and colleagues, Robert Braskamp, Robert Brown, and Diane Newman and others. This research identified factors associated with the likelihood of greater evaluation utilization. These factors can be categorized into three groupings: characteristics of users, context, and the evaluation itself.
In considering the first of these, researchers found a number of characteristics and attitudes of potential users of which evaluators should be aware. These included users' expectations for the evaluation, their prior experience and current disposition toward evaluation, and their perception of the risks they faced in having the evaluation performed.
...
- Concepts, Evaluation
- Personnel Evaluation
- Advocacy in Evaluation
- Evaluand
- Evaluation
- Evaluator
- Evaluator Roles
- External Evaluation
- Formative Evaluation
- Goal
- Grading
- Independence
- Internal Evaluation
- Judgment
- Logic of Evaluation
- Merit
- Metaevaluation
- Objectives
- Personnel Evaluation
- Process Evaluation
- Product Evaluation
- Program Evaluation
- Quality
- Ranking
- Standard Setting
- Standards
- Summative Evaluation
- Synthesis
- Value Judgment
- Values
- Worth
- Concepts, Methodological
- 360-Degree Evaluation
- Accountability
- Achievement
- Affect
- Analysis
- Applied Research
- Appraisal
- Appropriateness
- Assessment
- Audience
- Best Practices
- Black Box
- Capacity Building
- Client
- Client Satisfaction
- Consumer
- Consumer Satisfaction
- Control Conditions
- Cost
- Cost Effectiveness
- Criterion-Referenced Test
- Critique
- Cut Score
- Description
- Design Effects
- Dissemination
- Effectiveness
- Efficiency
- Feasibility
- Hypothesis
- Impact Assessment
- Implementation
- Improvement
- Indicators
- Inputs
- Inspection
- Interpretation
- Intervention
- Interviewing
- Literature Review
- Longitudinal Studies
- Measurement
- Modus Operandi
- Most Significant Change Technique
- Norm-Referenced Tests
- Opportunity Costs
- Outcomes
- Outputs
- Peer Review
- Performance Indicator
- Performance Program
- Personalizing Evaluation
- Rapport
- Reactivity
- Reliability
- Sampling
- Score Card
- Secondary Analysis
- Services
- Setting
- Significance
- Situational Responsiveness
- Social Indicators
- Sponsor
- Stakeholder Involvement
- Treatments
- Triangulation
- Concepts, Philosophical
- Verstehen
- Aesthetics
- Ambiguity
- Amelioration
- Argument
- Authenticity
- Authority of Evaluation
- Bias
- Conclusions, Evaluative
- Consequential Validity
- Construct Validity
- Context
- Credibility
- Criteria
- Difference Principle
- Empiricism
- Epistemology
- Equity
- External Validity
- Falsifiability
- Generalization
- Hermeneutics
- Inference
- Internal Validity
- Interpretation
- Interpretivism
- Logical Positivism
- Meaning
- Means-End Relations
- Moral Discourse
- Objectivity
- Ontology
- Paradigm
- Pareto Optimal
- Pareto Principle
- Phenomenology
- Point of View
- Positivism
- Postmodernism
- Postpositivism
- Praxis
- Probative Logic
- Proxy Measure
- Rationality
- Relativism
- Subjectivity
- Tacit Knowledge
- Trustworthiness
- Understanding
- Validity
- Value-Free Inquiry
- Values
- Veracity
- Concepts, Social Science
- Capitalism
- Chaos Theory
- Constructivism
- Critical Incidents
- Deconstruction
- Dialogue
- Disenfranchised
- Experimenting Society
- Feminism
- Great Society Programs
- Ideal Type
- Inclusion
- Lesbian, Gay, Bisexual, and Transgender Issues in Evaluation
- Minority Issues in Evaluation
- Persuasion
- Policy Studies
- Politics of Evaluation
- Qualitative-Quantitative Debate in Evaluation
- Social Class
- Social Context
- Social Justice
- Ethics and Standards
- The Program Evaluation Standards
- Certification
- Communities of Practice (CoPs)
- Confidentiality
- Conflict of Interest
- Ethical Agreements
- Ethics
- Guiding Principles for Evaluators
- Honesty
- Human Subjects Protection
- Impartiality
- Informed Consent
- Licensure
- Profession of Evaluation
- Propriety
- Public Welfare
- Reciprocity
- Social Justice
- Teaching Evaluation
- Evaluation and Approaches
- Accreditation
- Action Research
- Appreciative Inquiry
- Artistic Evaluation
- Auditing
- CIPP Model (Concept, Input, Process, Product)
- Cluster Evaluation
- Community-Based Evaluation
- Connoisseurship
- Cost-Benefit Analysis
- Countenance Model of Evaluation
- Critical Theory Evaluation
- Culturally Responsive Evaluation
- Deliberative Democratic Evaluation
- Democratic Evaluation
- Developmental Evaluation
- Empowerment Evaluation
- Evaluative Inquiry
- Experimental Design
- Feminist Evaluation
- Fourth-Generation Evaluation
- Goal-Free Evaluation
- Illuminative Evaluation
- Inclusive Evaluation
- Institutional Self-Evaluation
- Judicial Model of Evaluation
- Kirkpatrick Four-Level Evaluation Model
- Logic Model
- Models of Evaluation
- Multicultural Evaluation
- Naturalistic Evaluation
- Objectives-Based Evaluation
- Participatory Action Research (PAR)
- Participatory Evaluation
- Participatory Monitoring and Evaluation
- Quasiexperimental Design
- Realist Evaluation
- Realistic Evaluation
- Responsive Evaluation
- Success Case Method
- Transformative Paradigm
- Utilization-Focused Evaluation
- Evaluation Practice around the World, Stories
- Evaluation Planning
- Evaluation Theory
- Laws and Legislation
- Organizations
- Abt Associates
- Active Learning Network for Accountability and Performance in Humanitarian Action (ALNAP)
- American Evaluation Association (AEA)
- American Institutes for Research (AIR)
- Buros Institute
- Center for Instructional Research and Curriculum Evaluation (CIRCE)
- Center for Research on Evaluation, Standards, and Student Testing (CRESST)
- Center for the Study of Evaluation (CSE)
- Centers for Disease Control and Prevention (CDC)
- Centre for Applied Research in Education (CARE)
- ERIC Clearinghouse on Assessment and Evaluation
- Evaluation Center, The
- Evaluation Research Society (ERS)
- Evaluators' Institute™, The
- General Accounting Office (GAO)
- International Development Evaluation Association (IDEAS)
- International Development Research Center (IDRC)
- International Organization for Cooperation in Evaluation (IOCE)
- International Program in Development Evaluation Training (IPDET)
- Joint Committee on Standards for Educational Evaluation
- Mathematica Policy Research
- MDRC
- National Assessment of Educational Progress (NAEP)
- National Institutes of Health (NIH)
- National Science Foundation (NSF)
- Organisation for Economic Co-operation and Development (OECD)
- Performance Assessment Resource Centre (PARC)
- Philanthropic Evaluation
- RAND Corporation
- Research Triangle Institute (RTI)
- United States Agency of International Development (USAID)
- Urban Institute
- Westat
- WestEd
- World Bank
- World Conservation Union (IUCN)
- People
- Abma, Tineke A.
- Adelman, Clem
- Albæk, Erik
- Alkin, Marvin C.
- Altschuld, James W.
- Bamberger, Michael J.
- Barrington, Gail V.
- Bhola, H. S.
- Bickel, William E.
- Bickman, Leonard
- Bonnet, Deborah G.
- Boruch, Robert
- Brisolara, Sharon
- Campbell, Donald T.
- Campos, Jennie
- Chalmers, Thomas
- Chelimsky, Eleanor
- Chen, Huey-Tsyh
- Conner, Ross
- Cook, Thomas D.
- Cooksy, Leslie
- Cordray, David
- Cousins, J. Bradley
- Cronbach, Lee J.
- Dahler-Larsen, Peter
- Datta, Lois-ellin
- Denny, Terry
- Eisner, Elliot
- Engle, Molly
- Farrington, David
- Fetterman, David M.
- Fitzpatrick, Jody L.
- Forss, Kim
- Fournier, Deborah M.
- Freeman, Howard E.
- Frierson, Henry T.
- Funnell, Sue
- Georghiou, Luke
- Glass, Gene V
- Grasso, Patrick G.
- Greene, Jennifer C.
- Guba, Egon G.
- Hall, Budd L.
- Hastings, J. Thomas
- Haug, Peder
- Henry, Gary T.
- Hood, Stafford L.
- Hopson, Rodney
- House, Ernest R.
- Hughes, Gerunda B.
- Ingle, Robert
- Jackson, Edward T.
- Julnes, George
- King, Jean A.
- Kirkhart, Karen
- Konrad, Ellen L.
- Kushner, Saville
- Leeuw, Frans L.
- Levin, Henry M.
- Leviton, Laura
- Light, Richard J.
- Lincoln, Yvonna S.
- Lipsey, Mark W.
- Lundgren, Ulf P.
- Mabry, Linda
- MacDonald, Barry
- Madison, Anna Marie
- Mark, Melvin M.
- Mathison, Sandra
- Mertens, Donna M.
- Millet, Ricardo A.
- Moos, Rudolf H.
- Morell, Jonathan A.
- Morris, Michael
- Mosteller, Frederick
- Narayan, Deepa
- Nathan, Richard
- Nevo, David
- Newcomer, Kathryn
- Newman, Dianna L.
- O'Sullivan, Rita
- Owen, John M.
- Patel, Mahesh
- Patton, Michael Quinn
- Pawson, Ray
- Pollitt, Christopher
- Porteous, Nancy L.
- Posavac, Emil J.
- Preskill, Hallie
- Reichardt, Charles S. (Chip)
- Rist, Ray C.
- Rog, Debra J.
- Rogers, Patricia J.
- Rossi, Peter H.
- Rugh, Jim
- Russon, Craig W.
- Ryan, Katherine E.
- Sanders, James R.
- Scheirer, Mary Ann
- Schwandt, Thomas A.
- Scriven, Michael
- Shadish, William R.
- Shulha, Lyn M.
- Simons, Helen
- Smith, M. F.
- Smith, Nick L.
- Stake, Robert E.
- Stanfield, John II
- Stanley, Julian C.
- Stufflebeam, Daniel L.
- Tilley, Nick
- Torres, Rosalie T.
- Toulemonde, Jacques
- Trochim, William
- Tyler, Ralph W.
- VanderPlaat, Madine
- Wadsworth, Yoland
- Walberg, Herbert J.
- Walker, Rob
- Weiss, Carol Hirschon
- Whitmore, Elizabeth
- Wholey, Joseph S.
- Wildavsky, Aaron B.
- Worthen, Blaine R.
- Wye, Christopher G.
- Publications
- American Journal of Evaluation
- Evaluation & the Health Professions
- Evaluation and Program Planning
- Evaluation Review: A Journal of Applied Social Research
- Evaluation: The International Journal of Theory, Research and Practice
- New Directions for Evaluation (NDE)
- Practical Assessment, Research on Evaluation (PARE)
- The Personnel Evaluation Standards
- The Program Evaluation Standards
- EvalTalk
- Guiding Principles for Evaluators
- Qualitative Methods
- Archives
- Checklists
- Comparative Analysis
- Constant Comparative Method
- Content Analysis
- Cross-Case Analysis
- Deliberative Forums
- Delphi Technique
- Document Analysis
- Emergent Design
- Emic Perspective
- Ethnography
- Etic Perspective
- Fieldwork
- Focus Group
- Gendered Evaluation
- Grounded Theory
- Group Interview
- Key Informants
- Mixed Methods
- Narrative Analysis
- Natural Experiments
- Negative Cases
- Observation
- Participant Observation
- Phenomenography
- Portfolio
- Portrayal
- Qualitative Data
- Rapid Rural Appraisal
- Reflexivity
- Rival Interpretations
- Thick Description
- Think-Aloud Protocol
- Unique-Case Analysis
- Unobtrusive Measures
- Quantitative Methods
- Aggregate Matching
- Backward Mapping
- Benchmarking
- Concept Mapping
- Correlation
- Cross-Sectional Design
- Errors of Measurement
- Fault Tree Analysis
- Field Experiment
- Matrix Sampling
- Meta-analysis
- Multitrait-Multimethod Analysis
- Panel Studies
- Pre-Post Design
- Quantitative Data
- Quantitative Weight and Sum
- Regression Analysis
- Standardized Test
- Statistics
- Surveys
- Time Series Analysis
- Representation, Reporting, Communicating
- Systems
- Technology
- Utilization
- Loading...
Get a 30 day FREE TRIAL
-
Watch videos from a variety of sources bringing classroom topics to life
-
Read modern, diverse business cases
-
Explore hundreds of books and reference titles
Sage Recommends
We found other relevant content for you on other Sage platforms.
Have you created a personal profile? Login or create a profile so that you can save clips, playlists and searches