Sie sind auf Seite 1von 3

MEDICAL EDUCATION

EVALUATION PARADIGMS IN MEDICAL EDUCATION


Zarrin Seema Siddiqui

A BSTRACT
Medical education is regularly challenged with new and innovative ideas in the field of curricula, teaching learning processes and
assessment. Evaluation of these approaches and techniques provide vital information regarding their subsequent use and
application to enhance the quality of learning experiences for students. Therefore, it is essentially important to choose an
evaluation approach/model that provide meaningful and valid information to concerned stakeholders. Here two major paradigm
of evaluation i.e. scientific and naturalistic are discussed emphasizing on their use, strengths and limitations. It is concluded that
no single paradigm is superior to other and it is finally left to the evaluator for making the ultimate choice depending on the
purpose and questions that need to be answered through evaluation.

KEY WORDS: Evaluation. Medical education. Educational intervention.

INTRODUCTION
Medical education worldwide has undergone important behind scientific method is planned assessment of program
changes and shown notable advancement in recent years.1 effects by means of scientific measurements. It is, therefore,
Various interventions have been introduced and a need has referred to a model that is grounded in measuring learning
been recognised to apply evidence-based approaches on the attainment. The behavioural changes are measured against
impact of these interventions.2 Evaluation thus becomes an predefined learning outcomes and provide objective
integral part of these interventions. It implies that there is a information to assist in further development. The paradigm is,
methodology that allows to look at the results of what can be therefore, focussed on the predetermined objectives and the
done in an effective way to influence the actions going final outcomes rather than intervening in the learning
forward.3 Surprisingly, evaluation is one particular aspect of process.5
educational cycle which we, as human being, are constantly
performing in one way or other, it may be a television
programme, a sport event, a new departmental store or just METHODOLOGY
our colleagues to determine their worth, merit or significance. There are two general categories.
In the context of education, however, evaluation is used to 1. TRUE EXPERIMENTAL DESIGN : The subjects are randomly
determine the effectiveness of programs and ascertain that the assigned to program and comparison groups.
objective/outcomes have been achieved. This provides Randomisation is the vital ingredient of this process to
information to program staff and stakeholders in order to ensure that groups are totally comparable and the
identify changes that can further improve program observed differences in the outcomes are not the results of
effectiveness. extraneous factors or pre-existing differences. For example
a study was conducted to determine impact of an
Educational experts have suggested many evaluation models
to achieve the above - mentioned aims each having their own educational intervention on the students’ learning of
clinical research methodology where participants were
limitations and strengths. 4 In this article, two major
paradigms of evaluation, used in medical education i.e. divided into control and experimental groups to compare
the outcomes.6 One of the limitations noted by author was
scientific and naturalistic, are discussed with reference to their
key emphasis, methodology and analysis. The strengths and the exchange of learning material among students in two
groups, which may have influence on results.
limitations associated with each paradigm are also identified.
2. Q UASI EXPERIMENTAL DESIGN: If researcher feels that
randomisation is not possible or practical, quasi
SCIENTIFIC PARADIGM experimental design is recommended which may be of the
following types:
There has been great emphasis on scientific research methods
(a). NON -EQUIVALENT GROUP, POST-TEST ONLY: A study is
in initial stages when efforts were being made to introduce
program evaluation as a separate discipline. The basic idea conducted to compare the knowledge scores of
medical students in problem-based learning (PBL)
Department of Medical Education, University of Western Australia, WA 6009, Perth,
and traditional curriculum on public health topics.
Australia. The results show that PBL students were significantly
Correspondence: Dr. Zarrin Seema Siddiqui, Lecturer, Medical Education, Faculty of
more successful in the knowledge test used.7 Here
Medicine and Dentistry, University of Western Australia, MBDP: M515, 1st Floor ‘N’ only the outcome measure is used for comparison.
Block, SCGH, Nedlands WA 6009, Perth, Australia. E - m a i l : This, however, does not rule out if one group was
zsiddiqui@meddent.uwa.edu.au already better than the other before experiment or
Received July 21, 2005; accepted: February 17, 2006. there might be other influential factors.

JCPSP 2006, Vol. 16 (4): 291-293 291


Zarrin Seema Siddiqui

(b). NON-EQUIVALENT GROUP , PRE-TEST – POST-TEST: Here


the pre- and post-test design partially eliminates the NATURALISTIC PARADIGM
limitation earlier discussed. Still there might be This model is derived from ethnographic methodologies
problems resulting from students in control group developed by anthropologists. The rationale behind this
being exposed to the experimental condition. model was that no other model really captures the context of
(c). P OST-TEST ONLY, CONTROL GROUP: When a large the program that involves students, their families, teaching
number of students or teachers are involved, it is staff and other surrounding elements of the community.
practically impossible and time-consuming to do pre Naturalistic model provides detailed information about
and post-test, therefore, this design is used. individual, groups or institution as they occur naturally.
When compared to experimental model, the naturalistic
(d). T IME SERIES DESIGN: Several measurements are
model considers as well as values the positions of multiple
undertaken from both the groups prior to and after
audiences. It does not solely rely on numerical data and
the experimental treatment. This provides a more
focuses on program activity rather than intent. For example in
reliable evaluation, although, earlier problems may
1996-97, a structured training course was organised by a
still occur if groups differ at the onset of evaluation.
British University for Hungarian Medical Teachers. To assess
(e). A SSESSMENT OF TREATMENT GROUP ONLY: Another the relevance of this course to the needs of the participants, the
form of evaluation where only treatment group is evaluator adapted a naturalistic approach using in-depth
considered yet, without the information that might interviews. These interviews were triangulated using
have incurred in the absence of intervention/
observation and documentary analysis. Since the aim of
experiment, it would be hard to know whether the
evaluation was to construct the meaning of the value of MSc
program has actually any impact.
experience for foreign participants, this seems to be the most
appropriate model.1 2
STRENGTHS
Experimental designs are especially useful in addressing METHODOLOGY
evaluation questions about the effectiveness and impact of
The naturalistic model mainly relies on qualitative data and
programs.8 With the use of comparative data as context for
analysis. Four important forms of qualitative analysis,
interpreting findings, experimental designs increase our
confidence that observed outcomes are the result of a given commonly adapted by the naturalistic model13, are
program or innovation instead of a function of extraneous phenomenological analysis, content analysis, analytic
variable or events. For example, experimental designs may induction and constant comparative analysis.
help to answer questions such as; Attempts are made by the evaluator to look for information
• Would adopting a new integrated educational program that can be identified across multiple data sources and
improve student performance? methods. The categories or themes are identified and
relationship among categories is established. Finally, more
• Is problem-based learning having a positive impact on
evidence to support categories and relationship is collected.
student achievement and faculty satisfaction?
Dornan et al. used phenomenological analysis in their study to
• How is the professional development program influencing
evaluate how clinicians perceive their role in problem-based
teacher’s collegiality and classroom practice?
medical education and how closely those perceptions match
the curriculum they are teaching. 1 4 Yet in another study, the
LIMITATIONS role of autopsy in the modern undergraduate curriculum has
been investigated, using content analysis within a theoretical
The main limitations associated with this paradigm is that sample.15
objectives are sometimes difficult to predict, or may change as
In addition to qualitative data in narrative form, quantitative
the course proceeds and unintended learning may be more
data may also be included in analysis but, nevertheless, the
important than the expected outcome of a learning program. 5
model is highly reliant on the expertise of the evaluator to
The use of randomised controlled trial is specially limited in
interpret the data, determine the significance of results and
medical education for a number of reasons, which involve lack
draw conclusions.
of appropriate resources and funding.9,10 Other limitations are
to precisely define the parameters of the experiment and
control of variables which can literally take place only in
laboratories. A genuine control is impossible. Practical
STRENGTHS
Naturalistic methods have greater validity as they encourage
difficulties in separating groups often result in contamination
multiple data types and resources. This yields rich and timely
of designs.1 1 Withholding intervention to control group has
information about the important aspects of the program
also been subjected to ethical issues. Sample reliability and
implementation, interaction between various stakeholders,
lack of control group as mentioned earlier may also affect the
problems encountered by the program staff etc. This also
results of evaluation. Finally, even when the purpose of
decreases the possibility of missing intended effects as well as
evaluation is to assess the impact of a program, logistical and
offers some degree of flexibility to the evaluator, which
feasibility issues constrain the whole framework.
differentiates it from experimental model.

292 JCPSP 2006, Vol. 16 (4): 291-293


Evaluation paradigms in medical education

LIMITATIONS combination of methods when situation allows depending


upon the purpose for evaluation. However, if the key concern
There is higher degree of reliance on subjectivity and is program’s impact on participant outcomes or if multiple
reliability of human observers thus their personal programs are being considered with regards to their
observations and intuitive analysis can lead to biased opinion. effectiveness for decision-making, experimental designs seem
Similarly, data collection and analysis also becomes laborious appropriate. Otherwise, the evaluation should be enriched
task and may potentially be very expensive. with detailed information to enable timely and responsive
feedback for the stakeholders.
DISCUSSION
These two major paradigms have been in debate for a long REFERENCES
period. One group favours experimental methods for program 1. Yolsal N, Bulut A, Karabey S, Ortayli N, Bahadir G, Aydin Z.
evaluation on the basis that qualitative methods often provide Development of training of trainers programmes and evaluation of their
misleading information which may be detrimental for the effectiveness in Istanbul, Turkey. Med Teach 2003. 25: 319-24.
decision-making process. The other group argues that 2. Reed D, Price EG, Windish DM, Wright SM, Gozu A, Hsu EB, et al.
qualitative methods best serve the purpose of evaluation and Challenges in systemic reviews of educational intervention studies. Ann
reject any merit of experimental or quasi-experimental Intern Med 2005; 142: 1080-9.
methods. They see qualitative model as highly responsive and 3. Wolfensohn J. Evaluation and poverty reduction conference. Washington
emphasized that the entire evaluation can be performed DC: The World Bank, 1999.
through naturalistic methods of information collection.16,17 4. Stufflebeam DL. Evaluation models: new directions for evaluation, San
Francisco: Jossey - Bass, 2001.
An interesting debate between Robert F. Buroch and Milbrey
5. Brigley S, Littlejohns P, Young Y, McEwan J. Continuing medical
W. Mc Laughlin has been reported in literature.18 The occasion education: the question of evaluation. Med Edu 1997; 31: 67-71.
was the annual meeting of the Evaluation Network and the
6. Pelayo - Alvarez M, Albert - Rox X, Gil - Latore F, Gutierrez - Sigler D.
Evaluation Research Society in United States. The issue for Feasibility analysis of a personalised training plan for learning research
discussion was the recommendation to the Congress and the methodology. Med Educ 2000. 34: 139-45.
Department of Education regarding mandatory use of 7. Gurpinar E, Musal B, Aksakoglu G, Ucku R. Comparison of knowledge
experimental methods, where appropriate, by the federal scores of medical students in problem-based learning and traditional
government. Boruch argued that field tests are essentially curriculum on public health topics. BMC Med Educ 2005. 5: 7.
required for an unbiased estimate effects. He further stressed 8. Barry G, Herman J. True and quasi-experimental designs. Practical
that the quality of an evaluation will enhance provided assessment, Res Evaluation, 1997; 5(14).
experimental methods are implemented at federal or state 9. Smith M. The whole is greater: combining qualitative and quantitative
levels. On the contrary, McLaughlin pointed out the rigidity of approaches in evaluation studies, in naturalistic evaluation. New
experimental methods that make an evaluator ask wrong directions for program evaluation. San Francisco: Jossey-Bass, 1986.
questions, use wrong measures and thus resulting in failure to 10. Wilkes M, Bligh J. Evaluating educational interventions. BMJ 1999;
provide valid information for policy and practice. It is clearly 318: 1269-72.
evident from the debate that individuals within each 11. Kember D. To control or not to control: the question of whether
paradigm (naturalistic or scientific) when evaluate the other experimental designs are appropriate for evaluating teaching innovations
paradigm may overlook the strengths of that model while in higher education. Assess Eval Higher Educ 2003; 28: 89-101.
both paradigms have their own pros and cons. It is finally left 12. Bolden K, Willoughby SA, Claridge MT, Lewis AP. Training Hungarian
for the evaluators to make the choice considering a series of primary healthcare teachers: the relevance of a UK postgraduate course
for health educators. Med Educ 2000; 34: 61-5.
questions (Table I) based on Stufflebeam's four essential
attributes for a sound and fair programme evaluation.1 9 13. Payne D. Designing educational project and program evaluations: a
practical overview based on research and experience. Boston: Kluwer
Alternatively an evaluation matrix may be developed where Academic Publishers, 1994.
the set of research questions is tabulated against a selection of
14. Dornan T, Scherpbier A, King N, Boshuizen H. Clinical teachers and
possible analysis tools.20 This enables evaluators to select the problem-based learning: a phenomenological study. Med Educ 2005.
most appropriate and feasible method. 39: 163-70.
15. Burton J. The autopsy in modern undergraduate medical education: a
Table I: Essential questions for evaluation. qualitative study of uses and curriculum considerations. Med Educ
2003. 37: 1073-81.
Utility Will evaluation model be useful in serving the information
needs of the intended users?
16. Parlett M, Hamilton D. Evaluation and illumination: a new approach to
the study of innovative programs, in beyond the numbers game. Berkley,
Feasibility Will the evaluation model follow practical and feasible means CA: McCutchan, 1978.
for collecting evaluative information?
Propriety How the evaluation model will be conducted in an ethical
17. Guba E, Lincoln Y. Effective evaluation: improving the usefulness of
evaluation results through responsive and naturalistic approaches. San
manner, with due regard for the welfare of those involved in
the evaluation, as well as those affected by its results?
Francisco: Jossey- Bass, 1981.
Accuracy Will the evaluation model convey technically adequate 18. Davis B. Boruch and Mclaughlin debate. Eval News 1982; 3: 11-20.
information about the features that determined the worth or 19. Curran V, Christopher J, Lemire F, Collins A, Barrett B. Application of
merit of the programme being evaluated? a responsive evaluation approach in medical education. Med Educ
2003. 37: 256-66.
CONCLUSION 20. Reeves T. Evaluation matrix. available http://mime1.marc.gatech.
edu/MM_Tools/EM.html accessed online 29th September 2005.
There is no single model that is comprehensive enough to
conduct an effective evaluation. It is worthwhile to try a

l l l l l O
l l l l l

JCPSP 2006, Vol. 16 (4): 291-293 293

Das könnte Ihnen auch gefallen