California State University Long Beach
Graduate Center for Public Policy and Administration
Summer 2002, Third Session
PPA 590 WOMEN & PUBLIC POLICY
EVALUATION RESEARCH
Stages in Evaluation Research
Evaluation research is research which is characterized by 1) a focus on a plan or program of deliberate intervention, not just a natural or accidental event; 2) an objective or goal which is considered desirable or which has some positive value; 3) a method for determining the degree to which the planned program achieves the desired objectives.
Stages in Evaluation Research
STAGE |
DONE BY |
FOCUS |
METHODS |
I. Efficiency & the New Deal: 1910-1945 |
Industrial Research and New Deal Programs |
Efficiency in Productivity; Evaluating Social Change |
Early application of experimental & applied social research |
II. Field Research: 1945-1963 |
Sociology, Psychology, Armed Forces |
Behavioral and attitudinal change in social programs |
Field Research |
III. Social Experimentation: 1963-1974 |
Great Society Programs |
Turbulent setting of social programs; Utilization problems |
Social Experimentation |
IV. Public Policy-making: 1974-Present |
OMB, GAO, Federal agencies, State & Local government |
Congressional oversight; Rational policymaking; Implementation studies |
Experimentation, Quasi-experiments, variety of techniques & methods |
Stage I. From Efficiency to the New Deal.
As early as the late 19th century, researchers were doing evaluations: Frederick Taylor evaluated the effectiveness of particular approaches to work and management; Elton Mayo evaluated the effects of production policies at the Hawthorne plant on the informal organization of small work groups. However not until the 20th century was it an accepted research technique.
During the New Deal, government agencies and programs were created which provided social scientists excellent opportunities to use experimental evaluation research methods. Studies were made of the effects of work relief compared to direct relief, public housing projects, and the treatment of juvenile delinquents.
The application of experimental and quasi-experimental methods yielded important methodological discoveries (e.g., the Hawthorne effect).
Stage II. Field Research
The experience during the previous paradigm was mostly in controlled environments. During the second paradigm, field research was born as a way of designing, planning and managing investigations in real-life settings, such as schools, work camps, street corners, homes and the battle-field. The researchers were primarily sociologists and psychologists, who evaluated programs which were meant to affect attitudes, such as improvement of morale. Field research combined both qualitative and quantitative aspects of research, and acknowledged the role of the social scientist as a participant in the social world.
Stage III. Social Experimentation.
During the Johnson administration, initiation of Great Society programs provided unique opportunities for the application of research methods to government programs. These programs marked the first incursion of the Federal Government into the arena of social welfare, which was formerly the province of local and State authorities, if not private agencies. Almost immediately concerns arose over the integrity of these programs, their administration and effect. At first the programs were evaluated by the agencies themselves, but by 1970 Congress mandated or authorized evaluations of a wide variety of programs and gave the GAO increased responsibility for conducting program impact studies.
For the first time evaluators realized the difficulties of conducting studies in a "turbulent environment", marked by 1) the tendency of a program to shift as it was being evaluated; 2) the fact that the program was embedded in the organizational structure and that the structure itself has an effect on program outcomes; and 3) conflicting role expectations between program administrators and program evaluators, including high turnover among program evaluators, use of part-time evaluators, the use of poorly qualified evaluators, and a general carelessness in the conduct of evaluations.
Major social experiments were conducted, including negative income taxes, housing vouchers, and compensatory educational programs, which were seen as allowing for better planning and greater efficiency.
Stage IV. Public Policymaking.
By the mid 1970s, evaluation research had emerged as a central component in the intergovernmental power structure and in the policymaking process in general. Congress mandated evaluation of state and local programs which received federal funds, thereby augmenting its power through constitutionally mandated powers of oversight. Federal funding for evaluation grew from nearly nothing in 1960 to $63.6 million in 1977. Additionally, this greatly increased the prevalence of evaluation research at the State and local level. Reagan's New Federalism lead to an even greater reliance on evaluation research in the policy decision process at this level, in deciding how to distribute limited funds among competing programs.
While policymaking is still a political process, evaluation research was seen as infusing rationality into public policymaking, and as meeting the demands of decision makers at all levels for control of social programs. As part of their need for control, decisionmakers turned to social scientists for help in understanding the causes and solutions to social problems. The growth of the field of evaluation research was part of a larger trend of reliance on scientific experts, in order to base decisions on objective measures of program effectiveness. This was not the end of ideology, but the adoption of an ideology which reflected current trends in social policy making.
Stage V. The Qualitative vs. Quantitative Debate.
Evaluation Research has changed from an activity conducted in a few situations to multiple methods applied at every level of government. It is now clearly an integral component of public administration. Past paradigms have not vanished, rather each has been incorporated into succeeding paradigms. Lately there has been some loss of confidence in the quantitative methods of evaluation research (e.g., cost benefit, cost effectiveness, and cost utility analysis), and a call for more direct involvement of the evaluator in the policy process. However, it is unlikely that either approach will eventually dominate the field completely.
WHAT IS EVALUATION?
Evaluation is a process of judging; evaluation research uses the scientific method to control subjectivity. Both examine the relationship between variables, and they both use the same research designs and the same methods of data collection and analysis. However, the differences between evaluation research and pure research are:
EVALUATION RESEARCH |
PURE RESEARCH |
Used in decision-making |
Done for itself |
Program Derived |
Depends on theory |
Judgmental (what should be) |
Refrains from judgement |
Action Setting |
Controlled |
Role Conflicts |
Unbiased, unpressured |
Seldom published |
Always publishable |
Allegiances |
No particular allegiance |
What is Program Evaluation?
Evaluation research, or program evaluation, includes activities designed to support the ongoing decision-making process. It differs from basic research in that it is applied, and its purpose is to evaluate a particular activity. Its purpose is to allow decision-makes to confront the problem of resource allocation among competing social objectives and alternative ways of meeting those goals, and to clarify whether programs should be expanded, changed, or abolished.
Steps in the Process of Evaluation
1. Identify the decision maker(s), their perceptions, how the information will be used, and deadlines.
2. Define the nature and scope of the problem as the decision-makers see it; review the history of the program and problem(s), and any past evaluations.
3. Determine what the program’s objectives are or should be, including intended benefits, how they are to be obtained, intended recipients of services, intended and unintended consequences, and whether multiple objectives conflict.
4. Specify comprehensive measures to quantify how a program’s goal is being met, including tangible measures, side-effects, and distribution.
5. Prepare a detailed plan to cover problems; assumptions; constraints; methods; resources; schedules and deadlines.
6. Assemble the evaluation team, establish lines of communication, and determine the methods of evaluation to be used, considering the level of validity, reliability, significance, efficiency, and timeliness desired.
7. Specify procedures for using the results, including the nature of the reports, who they are sent to, implementation, and follow-up.
Designs for Evaluation Research
Evaluation research can use many types of research design. Social, political, ethical or other human considerations may impede the application of controlled experiments, and so pre-experimental or quasi-experimental designs may be the most appropriate in many instances.
Pre-Experimental. There is little control of extraneous factors.
1) The One-Shot Case study
2) The Pretest-Posttest design
3) Posttest only comparison group design
Quasi-Experimental. There is statistical control of extraneous factors.
4) Contrasted Group Design
5) Planned Variations Design
6) Time Series Design
7) Control Series Design
8) Correlational Design
9) Panel Designs
10) Combined Designs
Experimental Designs. Only experimental designs can control for covariation, time order, nonspuriousness, and rationale. Validity is greater since this design controls for the effects of History, maturation, experimental mortality, instrumentation, testing and statistical regression, as well as interaction effects. Experimental designs ensure greater generalizability of findings.
11) Controlled Experiment
12) Solomon 4-group Design
13) Posttest only Control Group Design
14) Time-Delayed Effect Design
15) Factorial Designs
Why are there demands for evaluation research?
-Changes in the nature of social problems, from individual to institutional
-Changes in the structure and function of public agencies
-Increasing number of professionals trained in research
-Changes in the needs and expectation of the public for government services
Who does evaluation/uses evaluation/benefits from evaluation research?
Program Managers Administrators Agency personnel
Practitioners Recipients/Clients Other agencies
Policy (law)-makers Academics Consultants
Mass media Interest Groups Outside groups
Elected Officials Taxpayers Congress
What are the purposes of evaluation research?
OVERT:
To initiate or terminate a program
To add or drop parts of a program
To extend a program to new areas
To allocate resources among programs
To test a proposed program
COVERT:
Postpone
Duck responsibility
Public Relations
Fulfill grant requirements
Stir up debate
When is evaluation research used?
Policy research/evaluation is not utilized only at the end of the policy making process (i.e., as feedback) but in every stage:
1. Value formation
2. Goal setting
3. Setting attainment criteria
4. Identifying goal-producing activities (Policy evaluation)
5. Program Implementation & operation
6. Assessment of goal attainment (Program Evaluation)
Problems with evaluation research are:
1.There is generally uncertainty over policy goals, which may be diffuse, unclear, and even conflicting. How can you tell if the program works?
2.There are generally many effects: intended and unintended effects; intended and unintended groups are affected; there are intended and unintended costs; and there are intended and unintended benefits.
3.There are methodological problems involved in attempting to establish causality.
4.How can data be obtained that is unbiased? Was the record-keeping changed during the program/during the evaluation?
5.Friction with program administrators and program workers may engender resistance or even sabotage evaluation efforts.
Other Problems:
Turbulent setting of ongoing programs Personality differences
Role differences (admin. vs. evaluator) Conflicting goals, values, ideology
Lack of clear definition of authority Institutional histories
Data collection Changes in record keeping
Selection of participants Controls
Feedback into program operations Feasibility, acceptability
THE ROLE OF EVALUATION IN PUBLIC POLICY
|
IDEAL USES |
ACTUAL USES |
WHO DOES IT? |
The evaluation community systematically produces relevant and timely estimates of impact and cost effectiveness. |
There is little evidence to document whether research is timely and informative; even less to show it is utilized or has an impact on programs or policies. |
SCIENTIFIC MODEL |
Evaluations can explain causal chains and what effects will be produced by program alternatives. |
Limitations of social research mean results are open to interpretation and offer little in the way of alternatives. |
POLICY PROCESS |
Policy goals are clear and measurable; theoretical foundations are sound; adequate resources are committed; agents implement the program effectively; the target population complies; and the evaluators have the necessary skills and resources. |
Evaluators may lack skills and funding; policy may be unclear; theory may be unsound; implementation may be imperfect; target population may be noncompliant; time may be too short; etc. |
UTILIZATION |
Policymakers as evaluation consumers realize the need for research and use it in decision making. Programs are stable to permit evaluation; evaluation criteria are built into the policy programs, as well as instructions for its utilization. |
Policy research and evaluation serve goals, such as symbolic or ritual goals; used to reduce complacency or to calm anxieties or to show an image of government as rational; used as a response to political pressures or conflictive management; reduce the scope for disagreements; opens up policy for negotiations, etc. |
There may be problems with the day-to-day operation of the evaluation, which depend greatly on the skills of the evaluator. Tensions may develop between evaluators and administrators, which require the evaluator to use many strategies. Evaluation should have a high priority in the organization and ideally be conducted by a consultant or third-party organization. Insiders have more knowledge about the organization, but may be less objective, less able to question the status quo, less able to mediate, and less able to devote time to the research. The fragmented multi-centered policy process is used by political actors to help them carry out their own goals (not the evaluator's). Evaluation information has to compete with other non-scientific factors which policymakers consider relevant in policymaking. Neither are crystal clear in pointing the way to a perfect decision. It is doubtful that evaluation impacts policy in other than an incremental fashion. It may not be used in short-term decision making; the evaluator is only an advisor and has no formal power; disagreements between evaluators and administrators may jeopardize the study; staff and resources to carry out recommendations may be lacking; political decisions may be preferred over technical decisions.
Ethical problems may arise with respect to the privacy of the individual, and whether there is a right to know anything about the individual; maintaining the confidentiality of the information; obtaining informed consent and being sure the individual has the capacity to agree to participate in the evaluation; assuring that the data are protected.
What types of failure can evaluation research identify?
IDEAL |
Theory about causal process is correct |
Correct program developed |
Correct program implemented successfully |
desired effects obtained |
THEORY FAILURE |
theory about causal process is incorrect |
Incorrect program developed |
Incorrect program implemented successfully |
undesired effects obtained |
PROGRAM FAILURE |
theory about causal process is correct |
Correct program developed |
correct program not successfully implemented |
undesired effects obtained |
Why policies don't have intended effects:
-Not enough resources were committed
-No administrative mechanism was set up
-Problems are multifaceted and policies are single-minded]
-People respond or adjust to negate the impact of policies
-Policies have incompatible goals that conflict
-Solutions may cost more than the problem
-Some problems don't have solutions
-The 'problem' changes over time
-New problems emerge which impact the area under evaluation
Program evaluation may be a form of social criticism. It can slowly change conventional beliefs of policy-makers over time. It can open up new options in the long run and on a broader scale. Evaluation research is a way to demonstrate the legitimacy of government, since there is no profit margin as in the public sector. Evaluation research can demonstrate the efficiency, effectiveness, and social worth of government. It can use the prestige of the research method to demonstrate the value of government programs.