Beruflich Dokumente
Kultur Dokumente
WHAT IS EVALUATION ?
The systematic assessment of the worth or
merit of some objects (Trochim, 2000)
Evaluation research is a robust arena of
activity directed at collecting,
analyzing, and interpreting
information on the need for,
implementation of, and effectiveness and
efficiency of intervention efforts to better
the lot of humankind (Rossi and Freeman,
1989).
MONITORING VS EVALUATION
Establishing factual premises about public
policies answering what happened,
how, and why
MONITORING
Concerned with establishing the value
premises necessary to produce
information about the performance of
policies answering what differences
EVALUATION
does it make
Dunn, 1994
EVALUATING PLAN:
PLANNING TO PERFORM
Evaluation Models for City Planning
(Leora Susan Waldner, 2004)
Waldner, 2004
Goal attainment,
outcome,
conformance, or
compliance
Impact evaluation,
effectiveness
attempts to
evaluate the
effectiveness of the
plan along some
dimension
Waldner, 2004
Strategically evaluate
the effect of the plan
on the community
E.g. What kind of
community resulted
from the zoning
patterns we set forth?
Waldner, 2004
Goal attainment?
Effects of the plan?
Its use in future decision-making?
A successful process?
Some other criterion?
Or a combination thereof?
Waldner, 2004
Plan assessment
Plan testing and evaluation
Plan critique
Comparative research and professional
evaluations
Post hoc evaluation of plan outcomes
Baer, 1997
Baer, 1997
PLAN CRITIQUE
WHO
WHEN
WHAT
Plan as a package
METHOD
Baer, 1997
WHEN
WHAT
METHOD
Baer, 1997
WHEN
WHAT
METHOD
Baer, 1997
WHEN
WHAT
METHOD
Baer, 1997
Baer, 1997
Baer, 1997
EVALUATING POLICY
PERFORMANCE
Dunn, 1994: Chapter 9
WHAT IS POLICY?
Value focus
Fact-value interdependence
Present and past orientation
Value duality
FUNCTIONS OF EVALUATION
Evaluation provides reliable and valid
information about policy performance
Evaluation contributes to the
clarification and critique of values
that underlie the selection of goals and
objectives
Evaluation may contribute to the
application of other policy-analytic
methods, including problem structuring
and recommendation
Type of
Criterion
Question
Illustrative Criteria
Effectiveness
Units of service
Efficiency
Unit cost
Net benefits
Cost-benefit ratio
Adequacy
Equity
Pareto criterion
Kaldor-Hicks criterion
Raws criterion
Responsivene
ss
Appropriatene
ss
Approaches to Evaluation
Approach
Aims
Assumptions
Major Forms
Pseudoevaluation
Use descriptive
methods to produce
reliable and valid
information about
policy outcomes
Measures of worth
or value are selfevident or
uncontroversial
Social
experimentation
Social systems
accounting
Social auditing
Research and
practice synthesis
Formal
evaluation
Use descriptive
methods to produce
reliable and valid
information about
policy outcomes that
have been formally
announced as policyprogram objectives
Formally announced
goals and objectives
of policy makers and
administrators are
appropriate
measures of worth
or value
Developmental
evaluation
Experimental
evaluation
Retrospective process
evaluation
Retrospective
outcome evaluation
Decisiontheoretic
evaluation
Use descriptive
methods to produce
reliable and valid
information about
policy outcomes that
are explicitly valued
by multiple
Formally announced
as well as latent
goals and objectives
of stakeholders are
appropriate
measures of worth
or value
Evaluability
assessment
Multiattribute utility
analysis
PSEUDO-EVALUATION
An approach that uses descriptive methods to
produce reliable and valid information about
policy outcomes, without attempting to
question the worth or value of these outcomes
to persons, groups, or society as a whole
The major assumption is that measures of
worth or value are self-evident or
uncontroversial
Variety of methods: quasi-experimental
design, questionnaires, random sampling,
statistical techniques
FORMAL EVALUATION
An approach that uses descriptive methods to produce
reliable and valid information about policy outcomes
but evaluates such outcomes on the basis of policyprogram objectives that have been formally announced
by policy makers and program administrators.
The major assumption is that formally announced goals
and objectives are appropriate measures of the worth
or value of policies and programs
The difference with pseudo-evaluation is that formal
evaluations use legislation, program documents, and
interviews with policy makers and administrators to
identify, define, and specify formal goals and
objectives
Control
Over Policy
Actions
Summative
Direct
Developmental
evaluation
Experimental evaluation
Indirect
Retrospective
process evaluation
Retrospective outcome
evaluation
1. DEVELOPMENTAL EVALUATION
refers to evaluation activities that are
explicitly designed to serve the day-today needs of program staf.
involves some measure of direct control
over policy actions
can be used to adapt immediately to new
experience acquired through systematic
manipulations of input and process variable
2. RETROSPECTIVE PROCESS
EVALUATION
Involves the monitoring and evaluation of programs
after they have been in place for some time
Often focuses on problems and bottlenecks
encountered in the implementation of policies and
programs
Does not permit the direct manipulation of inputs
and processes, rather it relies on ex post facto
(retrospective) descriptions of ongoing program
activities, which are subsequently related to outputs
and impacts
Requires a well-established internal reporting system
that permits the continuous generation of programrelated information
3. EXPERIMENTAL EVALUATION
Involves the monitoring and evaluation of outcomes under conditions
of direct controls over policy inputs and processes
All factors that might influence policy outcomes except one
a particular input or process variable are controlled, held
constant, or treated as plausible rival hypotheses
Must meet rather severe requirements before they can be carried out:
A clearly defined and directly manipulable set of treatment variables
that are specified in operational terms
An evaluation strategy that permits maximum generalizability of
conclusions about performance to many similar target group or settings
(external validity)
An evaluation strategy that permits minimum error in interpreting policy
performance as the actual result of manipulated policy inputs and
processes (internal validity)
A monitoring system that produces reliable data on complex
interrelationships among preconditions, unforeseen events, inputs,
processes, outputs, impacts, and side effects and spillovers.
DECISION
THEORETIC
EVALUATION
Decision-theoretic Evaluation
One of the main purposes of this evaluation is to
link information about policy outcomes with the
values of multiple stakeholders.
The assumption is that formally announced as well
as latent goals and objectives of stakeholders are
appropriate measures of the worth or value of
policies and programs
The two major forms of decision theoretic
evaluation are evaluability assessment and
multiattribute utility analysis, both of which
attempt to link information about policy outcomes
with the values of multiple stakeholders
Stakeholder identification
Specification of relevant decision issues
Specification of policy outcomes
Identification of attributes of outcomes
Attribute ranking
Attribute scaling
Scale standardization
Outcome measurement
Utility calculation
Evaluation and presentation
Approach
Pseudo
Evaluation
Techniques
- Graphic displays
- Interrupted time-series
analysis
- Tabular displays
- Index numbers
analysis
- Control-series analysis
- Regression-discontinuity
Formal
Evaluation
- Objectives mapping
Decision
Theoretic
Evaluation
- Brainstorming
- Value clarification
- Value critique
- Constrains mapping
- Cross impact analysis
- Discounting
- Argumentation analysis
- Policy Delphi
- User-survey analysis
References
1. Baer, W.C. (1997) General Plan Evaluation Criteria: An Approach to
Making Better Plans, Journal of the American Planning Association
63(3), American Planning Association, Chicago, IL.
2. Waldner, L.S. (2004) Planning to Perform: Evaluation Models for City
Planners, Berkeley Planning Journal 17(1), eScholarship, University
of California.
3. Dunn, W. (1994) Public Policy Analysis: 2nd Edition, Chapter 9:
Conclusion: Evaluating Policy Performances, Prentice-Hall, Inc., New
Jersey.