Evaluating E-learning A Guide to the Evaluation of E-learning - Graham Attwell (ed.) Evaluate Europe Handbook Series Volume 2
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
Graham Attwell (ed.) Evaluating E-learning A Guide to the Evaluation of E-learning Evaluate Europe Handbook Series Volume 2
Graham Attwell (ed.) Evaluating e-learning A guide to the evaluation of e-learning Evaluate Europe Handbook Series Volume 2 ISSN 1610-0875 2006 This work is licenced under the Creative Commons Attribution-NonCommercial-ShareA- like License. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc- sa/2.0/de/ or send a letter to Creative Commons, 559 Nathan Abbott Way, Stanford, California 94305, USA. Design and layout: Dirk Stieglitz, Bremen, Germany Printer: Perspektiven-Offset-Druck, Bremen, Germany This brochure is financed by the European Commission within the Leonardo da Vinci-Programme. Disclaimer: The views expressed in this report are those of the authors and do not necessarily reflect those of the European Commission. Neither the European Commission nor any person acting on behalf of the Commission is responsible for the use that might be made of the following information.
Contents Preface . ..................................................................................................... 5 1. Introduction – why do we need new models and tools for the evaluation of e-learning.............................................................. 7 2. Evaluating e-learning – what does the literature tell us?................ 9 3. A Framework for the evaluation of e-learning . ............................ 14 4. Models and theories of evaluation . ................................................ 17 5. Models and tools for the evaluation of e-learning – an overview...................................................................25 6. The SPEAK model and tool ............................................................. 27 7. Tool for the evaluation of the effectiveness of e-learning programmes in small- and medium-sized enterprises (SMEs)........ 30 8. Models and tools for evaluation of e-learning in higher vocational education . ............................................................... 32 9. Policy model and tool ........................................................................38 10. A management oriented approach to the evaluation of e-learning . ...................................................................... 42 11. Individual learning model and tool............................................... 45
Preface Evaluating e-learning A guide to the evaluation of e-learning W ho produced this guide? This guide has been produced as a report on the work of the Models and Instruments for the evaluation of e-learning and ICT supported learning (E-VAL) project. The project took place between 2002 and 2005 and was sponsored by the European Commission Leonardo da Vinci programme. The project was coordinated by Pontydysgu from Wales. The following organisations and individuals took part in the project: Who produced this guide? • Centre for Research and Education Develop- ment (CRED), Wales E-VAL project Jenny Hughes – jenhughes@mac.com • Interactive Institute, Sweden Who is this guide for and how can it be used? Cecilia.Katzeff – cecilia.katzeff@tii.se Laila Abdallah • Institute for Future Studies, Austria Friedrich Scheuermann – office@friedrich- scheuermann.net Klaus Reich – klaus.reich@futurestudies.org • IVLOS, University of Utrecht, Netherlands Wilfried Admiraal – W.F.Admiraal@ivlos. uu.nl Ineke Lam – J.I.Lam@ivlos.uu.nl Yvonne de Jong – Y.C.deJong@ivlos.uu.nl • Nexus, Ireland Brian Dillon – brian@nexus.ie • Pontydysgu, Wales Graham Attwell – graham10@mac.com Claire Middleton • University of Tampere, Finland Tuula.Heiskanen – tuula.heiskanen@uta.fi Riitta Kuusinen
Who is this guide for and how can it be used? sections of the guide and vive versa, those This guide is for anyone interested in the whose focus is more on research, to examine evaluation of e-learning. the different models and tools which have been More specifically it may of interest and developed. It is our strongly held belief that value to teachers and trainers involved in one problem in e-learning, and especially the e-learning, to education, pedagogy and edu- evaluation of e-learning, is the lack of iteration cational technology students and researchers, between theory and practice. to policy makers and planners and to institu- The guide does not claim to be the final tional managers. answer to the issue of evaluating e-learning. Whilst the Models and Instruments for the We claim only to have made some modest evaluation of e-learning and ICT supported progress towards solving a number of impor- learning (E-VAL) project was primarily fo- tant issues. We hope that others will build on cused on research, the project resulted in the our work in developing new insights and new development of a number of new models and models and tools in the forthcoming period. tools, which were tested in the final stages of The project partners are continuing to develop the project. the different models and tools described in This guide provides an overview of the the guide and the evaluate-europe web site research results of the project including the will provide up to date information on those outcomes of an extensive literature review developments. and an overview of different models and ap- The guide itself is only a summary of what proaches to the evaluation of e-learning and a was produced for the project. Access to more short review of the different models and tools detailed papers and to the full models and tools developed through the project. themselves is provided through the project Thus, it may be of interest to those involved web site and different sections of the report in the more theoretical research approaches index links to section to that site. to e-learning and to those who are looking for ways of evaluating e-learning programmes Contents of the guide and courses. 1. Introduction – why do we need new models and The primary focus for the Leonardo da tools for the evaluation of e-learning Vinci programme, which funded the project 2. Evaluating e-learning – what does the literature is vocational education and training. But we tell us? are aware that vocational learning takes place 3. A Framework for the evaluation of e-learning in a variety of different contexts and institu- 4. Models and theories of evaluation tional settings. The use of new technologies is 5. Models and tools for the evaluation contributing to the breakdown of the divides of e-learning – an overview between different sectors of education and 6. The SPEAK Model and Tool leading to a widening of the contexts in which 7. Tool for the evaluation of the effectiveness learning takes place. The models and tools of e-learning programmes in small- and medium- outlined in this guide have been tested in Small sized enterprises (SMEs) and Medium enterprises and in community 8. Models and tools for evaluation of e-learning in education, as well as in the more traditional higher vocational education vocational college and university settings. 9. Policy model and tool The format of this guide has designed to 10. A management oriented approach to the evalua- be flexible, to allow users to dip in and out of tion of e-learning different sections. Nevertheless, we would urge 11. Individual learning model and tool those whose primary interests are the practical and applied evaluation of e-learning to at least quickly look at the more research oriented
Section 1 Introduction – why do we need new models and tools for the evaluation of e-learning T he development of e-learning products and the provision of e-learning oppor- tunities is one of the most rapidly ex- panding areas of education and training. Whether this is through an intranet, the internet, multimedia, interactive TV or com- puter based training, the growth of e-learning is accelerating. However, what is known about these innovative approaches to training has been limited by the shortage of scientifically credible evaluation. Is e-learning effective? In what contexts? For what groups of learners? How do different learners respond? Are there marked differences between different ICT platforms? Does the socio-cultural environ- ment make a difference? Considering the costs of implementing ICT based training, is there a positive return on investment? What are the perceptions of VET professionals? What problems has it created for them? E-learning is also one of the areas that attracts the most research and development Development of e-learning products is one funding. If this investment is to be maxim- of the most rapidly expanding areas of edu- ised, it is imperative that we generate robust cation and training models for the evaluation of e-learning and tools which are flexible in use but consistent Is e-learning effective? in results. “Although recent attention has increased E-learning is also one of the areas that e-learning evaluation, the current re- attracts the most research and development search base for evaluating e-learning funding is inadequate … Due to the initial cost of implementing e-learning programs, it is important to conduct evaluation studies.” (American Society for Training and Development, 2001).
The Capitalisation report on the Leonardo da Vinci 1 programme, one of the biggest sponsors of innovative e-learning projects in European VET, also identified the lack of systematic evaluation as being the major weakness in e-learning projects. However, whilst some have been desperately seeking answers to the question ‘What works and what doesn’t work?’ and looking for ways of improving the quality of e-learning, the response by a large sector of the community of e-learning developers and practitioners has been a growing preoccupation with software and platforms. There has been only limited attention to pedagogy and learning. The devel- opment of models and tools for the evaluation of e-learning can help in improving the quality of e-learning and in informing and shaping future development in policy and practice.
Section 2 Evaluating e-learning – what does the literature tell us? T he Evaluation of e-learning project under- took a collaborative review of literature on the subject. Over one hundred reviews have been added to the project database and may be accessed through the project web site. This is what we found out. Categorising the literature In general there is an overwhelming prepon- derance on the evaluation of technology, usu- ally through ethnographic research methods (questionnaires to participants). However, it is possible to identify identified ten main categories of literature. Case studies of specific e-training programmes For the most part these are descriptive rather than analytic or predictive, predominantly American, mainly located in a Higher Edu- cation rather than vocational training envi- ronment and focused on the ‘virtual class- room’ model. They also tend to be restricted Categorising the literature to particular subject areas, in particular IT, languages and engineering disciplines. (This Comparisons with traditional learning is not necessarily to say that e-learning is restricted to these areas, rather that they are Return on Investment (ROI) reports over-represented in evaluation reports.) Studies on the contribution of evaluation to Comparisons with traditional learning metadata There are some (but surprisingly few) system- atic studies that compare e-learning effective- Issues and gaps ness with traditional learning and which are empirically robust. Those that exist are mainly Pedagogic and curricular approaches small-scale studies, often using a matched pairs design and are frequently of very specific Programme and policy evaluation instances of e-learning in which the e-learn- ing methodologies are idiosyncratic and the conclusions cannot be generalised.
Tools and instruments for evaluation of e-learning ment through traditional pedagogies. An There is an abundance of literature detail- additional problem is that the designers of ing tools for the evaluation of e-learning. these benchmarking systems are often locked However, these are mainly divided into two in to a particular model of e-learning which types. Firstly there are many on-line data limits their transferability. gathering instruments for assessing, typically, the user interface characteristics of software Product evaluation (e.g. student perception questionnaires) or By far the greatest number of `hits’ on evalua- secondly, there are devices to record and tion of e-learning are reports describing (and analyse usage by duration and frequency of extolling the virtues of) particular education log-in, pages accessed, user profile etc. Many software. The vast majority of these reports of these are sophisticated in their design and are commissioned or published by the soft- ingenuity but lack guidance on interpretation ware developers. This is not to question the and analysis. usefulness of these reports or necessarily to doubt their validity but evaluation of ‘de- Return on Investment (ROI) reports contextualised’ software is not an acceptable There are surprisingly few ROI reports, con- substitute for the rigorous evaluation of e- sidering the huge investments into e-learning learning systems. at all levels. The majority of those that exist draw mainly from industry based examples Performance evaluation and are written from an HRD perspective. The Scrivens (2000) in the USA, uses the term conclusion is inevitably that the investment ‘performance evaluation’ for what would, in was cost-effective and represented value-for- European terms, be called student assessment. money but often the savings are defined in Whilst it is true that an examination of student efficiency rather than effectiveness with no performance is a powerful indicator of the ef- long-term impact analysis that takes account fectiveness of e-learning, it is by no means the of unintended outcomes and consequences. only one. Moreover, a survey of reports on per- It is also difficult to compare figures across formance evaluation in the context of e-learning reports because the distinctions between net were mainly concerned with on-line tools and and gross costs, capital and revenue costs, instruments for examining knowledge-based displacement of existing funds, costs over time learner performance and could therefore be etc. are often blurred or missing. Many ROI categorised under that heading. type evaluation reports appear to be justify- ing investment rather than evaluating it and Handbooks for the evaluation of e-learning more geared to an audience of shareholders There is an increasing number of handbooks rather than researchers. for e-learning which focus primarily on evalu- ation. The evaluation methods and tools differ Benchmarking models widely. What they do have in common is that There have been several attempts to generate they recognise the importance of evaluation sets of criteria for quality assuring e-learning. and many propose that evaluation should be However, these tend to be skewed towards an integral part of any e-learning initiatives or proposing quality standards for e-learning development. In this regard, they tend toward a systems and software which often disregard management model of evaluation; the primary key variables in the wider learning environ- aim of the evaluation is to provide feedback ment or are based on criteria associated with to influence e-learning implementation and evaluating traditional learning processes (and future development. which disregard the technology) or criteria Many of the handbooks appear to have been associated with measuring learner achieve- produced by education advisors and advisory 10
services and agencies. They are frequently quite lihood that learning technologists focus on basic and lack a theoretical approach. A sur- the development of tools for evaluation, whist prising number cite The Evaluation Cookbook those coming from evaluation studies are and appear to have borrowed many of their more likely to consider how traditional models ideas from this influential publication. More of evaluation can be applied to e-learning. worrying is the technology centred approach Psychologists are more likely to undertake that many of the guides espouse. comparative studies. What is the importance of this? It suggests that we need teams of Meta-studies evaluators or of researchers from different There are a number of meta-studies of the disciplines to develop cross-disciplinary ap- evaluation of e-learning, all of them based on proaches to the evaluation of e-learning US literature. These are attempting to answer the question of the effectiveness of e-learning Issues and gaps by combining or bringing together the results First, the positive things which have emerged of a series of different studies to provide a form the literature review. larger sample base. Although, obviously, the There appears to be a growing realisation methodology is open to some question, these of the importance of evaluation. This seems studies are interesting and offer a new ap- to be linked to concerns that e-learning is proach to the issue, not succeeding in the way that had been ex- pected. Evaluation is needed to gain a better Studies on the contribution of evaluation understanding of the problems and issues to metadata regarding e-learning. This cannot be called a category of literature There is also an increasing focus on evalu- on evaluation as such, as this literature only ation methodologies and a realisation that includes evaluation as one of the factors to the evaluation of e-learning is complex and be taken into consideration in developing requires the development of new models and metadata. But, it is very important for the approaches. future of e-learning evaluation and will be explored further in the notes below. Evaluation and assessment However there remains confusion between Disciplinary backgrounds evaluation and assessment. To some extent It was notable that there are considerable this can be explained by different understand- differences in the disciplinary background ings of evaluation in different countries and of the authors and in the journals in which by the linguistic confusions between the two they were published or conferences they were processes. However, I think it goes further presented at. These include: than this. There appears to be an over reliance • Education and training journals and con- on assessment or achievement as the basis ferences for evaluation. That is not to say that learner • Journals and conferences relating to the achievement is not an evaluation factor. But use of Information and Communication the assumption that if something has not Technologies for learning been assessed it has not been learnt is surely • Specialist evaluation journals and confer- wrong. Furthermore, such an approach misses ences informal learning and learning not included • Journals and conferences around the psy- in the assessments. chology of learning. The disciplinary background of the research- Pedagogic and curricular approaches ers/authors tends to determine their approach Two big gaps stand out in the literature re- to evaluation. For instance there is more like- viewed. The first is the complete lack of any 11
evaluation – or attempts to evaluate – ped- Perhaps the lack of such studies just reflects agogic approaches of e-learning. This is in a time delay, before evaluation catches up contradiction to the increasing emphasis of with technical and e-learning development. e-learning researchers and developers on the Yet I think it reflects the need for accom- pedagogy of e-learning. panying evaluation where evaluation takes Evaluation approaches still tend to focus on place alongside technical development and the functionality of the learning technologies innovation. Secondly, it suggests to me that and not on the learning which is facilitated a discourse is needed between technical de- by the functionality. If pedagogy is such an velopers and innovators and evaluators with important factor in e-learning, then an impor- a better understanding or both of what the tant role for evaluation is to assist researchers other is doing. through providing an understanding of the impact of different pedagogic approaches. Programme and policy evaluation At the same time, in the European literature The second gap in the literature review is there is often an assumption that construc- programme and policy evaluation. Every tivism is the proven and best pedagogy for country and most large institutions have e-learning. Furthermore, it is often hard to active policy driven programmes to develop know what the authors mean or understand e-learning. However, there are few programme by constructivism. In the way the term is level evaluation reports, and still less models being used constructivism is only a means or theoretical approaches for evaluating e- of describing in one category all the varie- learning policy. This is both surprising and ties of creative, perceptive and innovative worrying. Without evaluation, how do we approaches to facilitating the acquisition of know which policy approaches are working new knowledge. and which are not? I remain suspicious that Linked to this lack of clear focus on peda- e-learning is merely seen by policy makers as a gogy is a similar gap in the evaluation of cur- ‘good thing’ and that money is being invested ricular approaches. Why is this important? with little understanding of where or why. In the technical developers world a debate has been raging for the last three years over Metadata something called learning objects. Learning As I mentioned earlier there are a growing objects have been seen as one of the main number of references to evaluation in the answers to the problem of sustainable con- technical literature around metadata. I think tent creation. A learning object is any digital, this is a significant development. Metadata reproducible and addressable resource used is simply data about data. However, it is cru- to perform learning activities or learning cial in allowing computers to know what support activities, made available for others exists on other computers and providing a to use. The problem is that is order to develop machine-readable description of learning reusability, objects need to be granular. This resources. At first it had been assumed that means that e-learning developers are pro- creators of learning resources would provide moting modularity as a curricular process, the metadata according to some kind of agreed driven by technical need. Many critics are common schema or standard. More recently, concerned that modularisation and granular is a realisation that in education many dif- learning object will challenge the coherence of ferent people have an interest and role in learning programmes. Since this is such a key providing the metadata associated with any debate in fort e-learning development, surely given object – developers, teachers, trainers, evaluation should focus on what is happening curriculum developers, technical developers, and whether modular programmes built from librarians and archivist, students and trainees learning object can work. and evaluators. Clearly, it would be absurd 12
to expect material developers or creators to add an objective description of the quality of their own learning materials and evaluation may play a critical role in describing quality. Technically the debate is around distributed metadata and how all the different data which becomes naturally associated with an object or learning materials in the course of their development, deployment and use, can be found and aggregated. Once more, I think this growing debate shows a necessity for evaluators to work along- side technical developers and at a more theo- retical level for a discourse around ideas. 13
Section 3 A framework for the evaluation of e-learning F rom a baseline of practice of attempting to evaluate many e-learning programmes, one of the biggest problems has proved to be handling the number of variables which potentially impact on the effectiveness of the programme and deciding what constitutes dependent, independent and irrelevant vari- ables in a given situation. Literature reviews and the study of existing evaluation practice, suggests that many evalua- tion tools and schema tend to disregard – con- sciously or otherwise many of these variables. Much of existing practice is overly focused on the technology – and on learner reaction to the use of technology. Socio-economic factors such as class or gender are seldom considered and even learning environment variables such as the subject environment are all too often ignored. Not only does this result in limitations in the data available on the use of ICT in learning but the limited recognition of the different variables can distort analysis of the weaknesses (and strengths) in current e-learn- ing provision. The evaluation of e-learning project has What framework? developed a more comprehensive framework. Over several e-learning evaluation projects, What factors are being disregarded or five major clusters of variables have emerged; edited out of the framework? individual learner variables, environmental variables, technology variables contextual vari- How might this framework be used? ables and pedagogic variables. Each of these can be disaggregated into more precise groups and further disaggregated until individual variables can be identified and isolated. Of course we recognise that no single evalu- ation model or tool, much less evaluation study, can address every variable. But, we believe in approaching and designing any evaluation it is important to be conscious of 14
what factors are being disregarded or edited may be context determined) for mapping and out of the framework. coding existing work into the effectiveness, Individual learner variables include efficiency and economy of e-learning irrespec- • physical characteristics (e.g. age, sex, physi- tive of whether this is an evaluation or an cal abilities) independent research study. Methodologies • learning history, (negative/positive experi- are cross-referenced against the variables ence, level of attainment, duration, recency being studied and major areas of omission etc.) can be identified that in turn will suggest a • learner attitude (positive/negative) future research agenda. • learner motivation (high/low) Secondly we are using the clusters of vari- • familiarity with the technology ables can be sued for proposing and testing Learning environment variables include hypotheses. Any one cluster can act as the • the immediate (physical) learning environ- dependent variable; the other four then oper- ment ate as independent variables. For example, at • the organisational or institutional environ- the micro level, part of the Eval project has ment tested the hypothesis that the effectiveness of • the subject environment different e-learning pedagogies will depend Contextual variables include on particular individual learning histories. • socio-economic factors (e.g. class, gen- Another survey explored whether the effec- der,) tiveness of particular technologies depends on • the political context (e.g. who is fund- gender. At a macro level we are also interested ing/paying for the e-learning and for what in whether the presence (or absence) of some reason?) individual variables or clusters of variables are • cultural background (e.g. how highly is more significant than others in determining learning/e-learning valued?) the effectiveness of e-learning and, if so, can • geographic location (e.g. country, language, they be weighted in some way? Is the profile of urban/rural) the learner more significant than the nature of Technology variables include the learning environment? Is the effectiveness • hardware of the technological solution outweighed or • software, enhanced by particular environmental vari- • connectivity, ables? Which is more important – getting the • the media software right or the learner support right? • mode of delivery, Can we use statistical techniques such as fac- Pedagogic variables include tor analysis to see which variables ‘cluster’ • Level and nature of learner support sys- together and impact on each other? tems We were not able to test every variable in • accessibility issues. the limited time and resources available to us • Methodologies through he project. However, the research we • Flexibility were able to undertake proved the value of the • Learner autonomy framework as a tool for research and confirmed • Selection and recruitment the validity of the framework design. • Assessment and examination Thirdly, we have found it a useful framework • Accreditation and certification for evaluating and researching the effective- ness of specific e-learning projects and pro- How might this framework be used ? grammes. The evaluation of e-learning, and Firstly the framework can be used to develop a research into the evaluation of e-learning, has robust classification system with clearly identi- been dominated by descriptive ethnographic fied levels of aggregation, (which themselves studies, rather than interpretation and analyses 15
and there is a predominance of ethno-meth- odological approaches, in particular, heavily contextualised case studies. The relatively small number of empirical studies has focussed on a limited number of variables. The best of these have controlled for variables other than those under study; the worst have simply discounted them. As the databank of research results is built up, particularly as the different variables are ‘weighted’, it becomes easier to identify the irrelevant variables and allow for the impact of others. It also allows predictions to be made which can short circuit the search for an ap- propriate evaluation methodology. 16
Section 4 Models and theories of evaluation E volution of evaluation theory Evaluation as a formal activity that we would recognise, has existed for a surprisingly long time. One of the earliest recorded was the evaluation of the effective- ness of lime-juice in preventing scurvy in sailors – commissioned by the British navy in the 18th century! The French make even earlier claim and say that the Norman armies conducted an evaluation of the relative ef- fectiveness of the crossbow and the longbow. Unfortunately, on the basis of the evaluation findings, the management decision was to go for the crossbow and the rest, as they say, is history! However, evaluation has only become a rec- ognised area of academic study since about the 1960’s. It is probably true to say that evaluation started as a field of practice and the theory was derived from it. As it evolved, so ideological disputes developed alongside disagreements Evolution of evaluation theory on definitions, terminology, ethics and so on. FitzPatrick, Sanders and Worthern in 2004 Philosophical /ideological difference identified nearly 60 different models in the 30 years between 1960 and 1990 alone. This Differences based on defining value or proliferation of models was bewildering for worth the practitioner, especially as many of these models and the tools they generated had no Differences according to discipline or field of obvious theoretical perspective. application Why is this a problem? Why should practi- tioners need a theoretical framework? Simply, Differences in practice a ‘good’ theory will set out the assumptions that it is making and on which its logic is A classification schemata for evaluation predicated. Different theories make different approaches assumption and generate models that will be based on different pre-conceptions and definitions of evaluation, which in turn lead to very different practices. 17
Deriving a taxonomy of evaluation approaches and conclusions. Thus, the evaluation pro- Many researchers have tried to make sense cedures are ‘internal’ to each evaluator and of this huge diversity of models and theories are not explicitly understood or reproducible and to find some way of classifying them. by others. However, even they could not agree so now Until 20 years ago, objectivism in evaluation we have the problem of trying to classify the was a goal to be aspired to. However, the same classification systems! criticisms levelled at the usefulness of positiv- All this is by way of saying that what follows ism in the social sciences in general were also is only one framework for distinguishing be- applied to objectivism in evaluation. tween different theories of evaluation and you Campbell (1984) summed it up: may well come across others. This framework, “twenty years ago positivism dominated which we find comprehensive and useable, was the philosophy of science...Today the tide devised by FitzPatrick, Sanders and Worthen has completely turned among the theo- (2004) who we have quoted at length. rists of science in philosophy, sociology, and elsewhere. Logical positivism is al- Philosophical/ideological differences most universally rejected.” Approaches to evaluation may differ funda- This point of view has been upheld by many mentally because their underpinning philoso- writers on evaluation and even if it is not phy or ideological base is different. That is, universally subscribed to, probably represents different evaluation theories will be based on a general trend. The major argument is that different assumptions about the way the world unlike traditional scientific research, evalu- works and so the models and practices based ation has to deal with complex phenomena on those theories will be different as well. By in real world settings, take into account a and large, we can locate them on a continuum multiplicity of stakeholders, unstable and from objectivist to subjectivist. unpredictable systems and requires a high Objectivism is equivalent to the empirical level of human interactivity. tradition in scientific research (positivism) The other criticism is that objectivism de- and focuses on data collection and analysis pends for its validity on its ‘scientific’ meth- techniques that produce results that are re- odology and is only credible if you happen to producible and verifiable by other evaluators value that methodology. We would argue that and to generate conclusions that are evidence objectivism conceals hidden values and biases based and which can be ‘scientifically’ jus- of which many evaluators are unaware – even tified. So the evaluation is ‘external’ to the the choice of data collection techniques and evaluator who is simply someone technically instruments is not value-neutral but this is not competent and proficient in the application recognised or else ignored by many so-called of procedures. objective evaluations. Subjectivism is based on: Despite the reaction of the theorists, how- “... an appeal to experience rather than ever, the message does not seem to have filtered to scientific method. Knowledge [of the through to the client base and the majority of evaluator] is conceived as being largely evaluation consumers, particularly in educa- tacit rather than scientific.” tion (and the public sector in general), are still (House 1980 in FitzPatrick, Sanders wedded to the idea of objective evaluation and and Worthen 2004) ‘finding the facts’. The validity of a subjectivist evaluation de- The major criticism is that subjectivist evalu- pends on the relevance of the evaluators’ ation often leads to contradictory conclusions background, their experience and expertise, that cannot be reconciled because the processes the keenness of their perceptions and their which led to the conclusions is largely inside the insightfulness in generating interpretations evaluators head and so cannot be replicated. 18
Differences based on defining value or worth there are other major divides based on meth- We can also distinguish between different odological differences that are not necessarily theoretical approaches depending on how rooted in a particular philosophical approach. they define value and make judgements, rather For example, many evaluators (both theoreti- than on their philosophical differences. This cians and practitioners) and also many clients time the continuum extends from ‘utilitarian’ tend to view qualitative and quantitative ap- to ‘intuitionist-pluralist’. proaches as different paradigms. We do not ‘Utilitarianism’ is a philosophy based on subscribe to this view, believing that this is maximising happiness in society. Utilitarian not a fundamental divide but simply a way of approaches to evaluation are based on the describing evaluation approaches by types of premise that the best programmes are those data that are used. Nevertheless, we recognise that achieve the greatest good for the great- this as an important distinction for others and est number. The evaluator will try and assess one that impacts on the overall evaluation overall impact in terms of total group gains methodology and the tools used. by using average outcome scores against the criteria selected to determine worth. Again, Differences according to discipline or field governments and the public sector tend to of application be adherents of this type of evaluation as Evaluation is a relatively young field and still it lends itself to large-scale comparisons of draws heavily on methodologies adapted from programmes and mass aggregation of data. anthropology, sociology, psychology, philoso- Managers and public programme administra- phy, economics and mathematics. One of the tors tend to be the main audiences. consequences is that evaluation approaches According to FitzPatrick et al, the intui- can be grouped around their parent discipline tionist-pluralist approach is at the other end so we tend to find ‘mathematical approaches’ of the spectrum and is based on the premise or ‘sociological approaches’. More recently the that value depends on the impact of a pro- search for new models has widened its net and gramme on each individual and the ‘greatest evaluation theorists such as Smith (1981) are good’ is that which maximises the benefits for trawling newer disciplines such as investiga- all stakeholders. This evaluation focus will be tive journalism, photography, storytellling, on the distribution of gains (for example by philosophical analysis, forensic pathology cultural or sub-cultural demographic groups and literary criticism for new ideas. such as ethnicity or gender or age) or distri- Evaluation theory has also developed in a bution of benefit across stakeholders (e.g. social context and practitioners work in dif- learners, administrators, delivery agencies, ferent cultures, different sectors, with different funding bodies, the public). There can be no target groups and different audiences. Conse- common index of ‘good’ but rather a plural- quently, different approaches and models have ity of criteria and judges. The evaluator is no tended to emerge based on these factors. For longer an impartial ‘averager’ but a portrayer of example, ‘education programme’ evaluation different values and needs. The merit or worth has developed along a different trajectory of any programme depends on the values and than, for example, the health services. Whilst perspectives of whoever is judging it and all many writers would argue that this is not a stakeholders are legitimate judges. true theoretical divide, ‘theory-in-practice’ is a powerful determinant of evaluation ap- Methodological differences proach and also stakeholders perceptions and Although there is a strong correlation between expectations of the evaluation process. an evaluator’s ideological approach and the methodology and techniques they will use (because of necessity one drives the other), 19
Differences in practice process of evaluation. Others claim that not The above distinctions are all based (loosely) only is this unnecessary but can, on occasions, on theoretical divisions. However, FitzPatrick be unhelpful. et al also point out that differences in approach can be practice-driven. A classification system Firstly, evaluators disagree about whether The above analysis is interesting and helps evaluators should simply provide information understand the major theoretical divides in so that decision makers can make the value evaluation. However, it does not get us far in judgements. Others, would say that the evalu- terms of systematically examining the varia- ator’s report to decision makers is incomplete tion between particular evaluation approaches if it does not contain a value judgement. because although those approaches could be Secondly, evaluators differ in their percep- positioned on each of the above ‘dimensions’, tion of their own role and their place in the their location would vary from one dimension evaluation process. Who has authority and to another. The next section tries to provide responsibility for the evaluation and to whom some answers. should the evaluator be accountable and an- Many evaluation theorists have attempted swerable? If one evaluator sees his role as a this but we are going to stick with the solu- ‘critical friend’ and another as ‘inspector’ or tion put forward by Fitzpatrick, Sanders and ‘judge’, then this will obviously influence the Worthen (1983). We are proposing to use their way they conduct an evaluation and also the work – with some modifications – partly in conclusions they draw. the interests of consistency (having referenced Thirdly, evaluators will be limited by their them heavily so far) and partly because they prior experience both in evaluation and also set out very clearly the thinking and rationale by their own discipline or professional back- underpinning their classification system. ground. Evaluation skills and knowledge are For the purist it is a less than satisfactory cumulative. Previous exposure to frequently taxonomy as the approaches do not neces- recurring problems will affect the way an sarily differ from one another along the same evaluator works. On the one hand it will prob- dimension. However, they are pragmatic as ably mean the evaluator is quicker to detect they conveniently represent the major clusters problems, to identify issues of concern and of models and approaches in use today. make more insightful judgements. On the other hand, it will also mean that the evaluator’s A classification schemata for evaluation approaches perceptions in a new situation are unlikely to FitzPatrick et al identify 5 major clusters of be ‘neutral’. evaluation approaches: Fourthly, evaluators have different views • Objectives oriented approaches about what skills and expertise they should • Management oriented approaches possess. Evaluators are frequently chosen on • Consumer oriented approaches the basis of their expertise or practitioner • Expertise oriented approaches base in the field being evaluated rather than • Participant oriented approaches on the basis of their skills and experience as However, to this we propose to add Van der an evaluator. This is gradually changing but Knapp’s ‘learning oriented approach’. as evaluation is becoming increasingly profes- These 6 categories fall more or less along sionalised and recognised as a specialist area a continuum from utilitarian to intuition- in its own right, so professional evaluators ist-pluralist so there is some logical basis in are becoming specialised within the area. addition to its convenience and accessibility Some evaluators would argue that specialist (see figure 1). knowledge of the field being evaluated is a pre-requisite for the credibility of the whole 20
Objectivist subjectivist Rationalist – positivist Utilitarian intuitionist-pluralist Naturalistic & Objectives Management Consumer Learning Expert participant oriented oriented oriented oriented oriented oriented A Continuum of Evaluation Models Figure 1: A Continuum of Evaluation Models Objectives orientated evaluation approaches needs of managers, policy-makers, adminis- Objectives-orientated evaluation is based on trators and practitioners. the idea that the purposes, goals or targets of Developers of this approach have tradition- a project are determined at the start and the ally relied on a systems approach to evaluation evaluation process should establish whether in which decisions are made about inputs, these have actually been achieved – and, if not, processes and outputs based on logic models why not. It is very similar to another approach and cybernetic theory. However, more recent known as ‘a systems approach’ to evaluation developments have highlighted different levels and both are very popular with public sector of decision and decision makers and have fo- agencies who are concerned with justifying cussed on who will use the evaluation results, expenditure and performance measurement. how they will use them and what aspect(s) of It is sometimes called ‘goal-driven’ evaluation, the system they are making decisions about. in contrast with other approaches, which are Not surprisingly, it is the model preferred called ‘goal-free’. by many managers and management com- There are many examples of objectives mittees but the downside is that the needs of orientated models; the earliest is probably other stakeholders are ignored. Tyler’s and more recently, Provus’s Discrep- Stufflebeam’s CIPP model is one of the ancy Model. most popular in management-orientated The disadvantages are that this sort of ap- evaluation. proach can miss important outcomes if they were not included in the original objectives Consumer orientated approaches nor does it challenge the value of the objec- Consumer orientated approaches to evaluation tives themselves adopt the perspective of the end user of what- Management orientated evaluation approaches ever service or product is being provided. For The management-orientated approach to this reason they tend to be summative, rather evaluation is meant to serve decision mak- than formative and are concerned primarily ers. Its rationale is that evaluation information with product evaluation. Consumer-orientated is an essential part of good decision making evaluation relies heavily on criteria referenced and that the evaluator can be most effective evaluation techniques such as benchmarking or by focussing the evaluation products on the kite marking and is understandably popular with standards agencies and ‘watchdog’ organisations. 21
Michael Scrivens ‘Key Evaluation Checklist’ is included it because it is an approach that we probably the best-known example. personally use more than any other. The major disadvantage of a consumer-ori- The operating principle is that the purpose entated approach is that it is a ‘backward- of evaluation is to contribute to some form of mapping’ approach and does not help make collective or organisational learning. Different predictions about future impacts. It also tends models within this overall approach are based to play down the nature of human interaction on different theories and types of learning with the products being evaluated. including ‘corrective’ or behavioural learn- ing, cognitive learning and social learning. Expertise orientated approaches The outputs and processes of the evaluation Expertise orientated evaluation is based on the form the inputs of the learning. notion of ‘connoisseurship’ and criticism and The pioneer of work in this field was Peter relies on the subjective professional judgement Van der Knaap. More recently we have ex- and expert knowledge of the evaluator. This is tended the approach to include evaluation the oldest form of evaluation and is still very as a contributor to knowledge creation in an popular despite its limitations. organisation. Expertise-orientated evaluation may be for- Learning-orientated evaluation approaches mal or informal, based on individual expertise are still not widespread but are beginning to or, more usually, on the collective expertise of gather momentum in the social agency sector, a panel. The opinions of multiple experts is in education establishments and in voluntary popularly believed to minimise bias, though organisations. this does not always follow! It relies far less The main limitations of this approach is that on external tools and instruments than other it does not lend itself to ‘mass surveys’ as it forms of evaluation and more on the experi- relies heavily on personal interaction between ence and wisdom of the evaluator. the evaluator and the project team and the Many public systems are based on expertise evaluator’s understanding of the learning needs orientated evaluation – for example the jury of the organisation. Also, within this overall system, school inspection system, licensing approach there are very disparate models, agencies, review boards, the refereeing system some requiring a high level of commitment for academic journals, national commissions to the process, which may be lacking. and enquiries and so on. Many organisations expect this type of Participant-orientated evaluation approaches evaluation if they employ an external evalu- An increasingly popular approach that dif- ator and the notion of evaluation by ‘peer fers fundamentally from all the others as it review’ is still the dominant model in most takes the needs of project participants as its professional associations. starting point. This is not to say that the other The disadvantages are obviously the high reli- approaches ignore participant needs but that ance on the assumed expertise of the evaluator for the most part benefits for participants and a lack of explicit and published standards. represent the end point of the evaluation and Also, the credibility of results is affected by not the beginning. the status of the evaluator but equally the Participants are not simply the direct ben- credibility of the evaluator is often affected eficiary target group of a project but will by the results. also include other stakeholders and potential beneficiaries. Thus, an educational project for Learning-orientated evaluation approaches women returners would include the learners This is a relatively new group of approaches themselves, the project staff, the management and not one that was included in FitzPatrick team and the funders but may also include et al’s classification. Nevertheless we have the wider community, the learners families, 22
the schools attended by the learners’ children, Boyle, P. G., Planning Better Programs, New childcare agencies or whatever. York: MacGraw-Hill Book Company, Participant-orientated evaluation does 1981. not usually follow a formal plan drawn up C.L.Taylor and Carl E. Beeman2 (1992) Evalu- in advance; rather it looks for patterns in ation for Accountability: An Overview, the data as the evaluation progresses. Data is University of Florida gathered in a variety of ways, using a range Describing Innovation as a prerequisite for of techniques and culled from many different innovation, Hughes and Attwell 2000, sources. Understandings grow from observa- www.evlauate-europe.net tion and bottom up investigation rather than Donald Clarke (1997) Instructional system de- rational deductive processes. The evaluator’s velopment: evaluation phase, http://www. role is to represent multiple realities and values nwlink.com/~donclark/hrd/sat6.html, ac- rather than singular perspectives. cessed November 4, 2003 Participant-orientated evaluation includes Israel, Glenn D. and C. L. Taylor. 1990. „Can many sub-groups that share all or some of the Response-Order Bias Evaluations?“ Evalu- above characteristics including Responsive ation and Program Planning. 13(4):365- Evaluation, Naturalistic Evaluation, Utilisa- 371 tion Focussed evaluation and Empowerment Mendenhall, W., and R.L. Scheaffer. 1973. Evaluation. Of all the models, probably the best Mathematical statistics with applications. known and one of the most useful is Stakes Duxbury Press, North Scituate, MA. Countenance Framework. Patton, M. Q. (1986) Utilization-focused evalu- Criticisms of this approach are many; bu- ation. Newbury Park, CA: Sage. reaucrats tend to hate it because of its lack of Project Star, User‘s Guide to Evaluation for ‘objectivity’ and because the outputs of the National Service Programs, http://www. evaluation are unpredictable. It is difficult to projectstar.org/star/Generic/usersguide. cost and control. Without a very experienced htm, accessed November 4, 2003 evaluator to facilitate the process, it can de- Rossi, P. H., & Freeman, H. E. (1989). Evalu- generate from an ‘organic’ approach to one ation: A systematic approach. Newbury which is chaotic and lacking in focus. Also, Park, CA: Sage Publications. there may be concentration on process at the Ryans, D. G. (1960). Characteristics of teach- expense of outputs. ers. Washington, DC: American Council on Education References Administration on Children, Youth and Fami- Scrivens M. (1991), The evaluation thesaurus, liesThe Program Manager‘s Guide to Evalu- Sage ation, http://www.acf.dhhs.gov/programs/ Weiss The Interface Between Evaluation and core/pubs_reports/foreword_pmguide. Public Policy, in Evaluation: The Interna- html, accessed 4 November, 2003 tional Journal of Theory, Research, and Allum, N.C Bauer, M.W., Gasgell, G. and. Practice (1999) (2000) Quality, Quantity and Knowledge interests, Avoiding confusion in Bauer and Gasgell, Qualitative research, Lon- mdon, Sag Boone, E. J., Developing Programs in Adult Education, Englewood Cliffs, New Jersey: Prentice-Hall Inc., 1985. 23
Section 5 New models and tools for the evaluation of e-learning – an overview T he remainder of this guide is given over to a description of the different models and tools developed through the project. We do not suggest this is a comprehensive selection – but it does illustrate the different models and tools which can be deployed. Cen- tral to our project’s findings are that evaluating e-learning is no different than evaluating any other form of learning – but that there are many variables and that the models and tools must take account of the different aims of the evaluations and contexts in which e-learning is taking place. The following models and tools have been developed through the project and are de- scribed in more detail in the following sec- tion. Models and tools for evaluation of e-Learning in higher vocational education The learner-benchmarking tool developed for the evaluation of e-learning in higher vocational education is a consumer tool in order to evaluate the use of Virtual Learn- Models and tools for evaluation of e-Learn- ing Environments (VLE). It is based on an ing in higher vocational education on-line questionnaire including items on the appraisal of the VLE used by the teachers in Models and tools for evaluating e-learning their courses. policy The Learner Evaluation tool is focused on the guidance or scaffolding teachers provide Tool for the evaluation of the effectiveness within the VLE and is based on constructivist of e-learning programmes in small and principles. medium sized enterprises (SMEs) A management oriented approach to the evaluation Evalact of e-learning The management oriented approach to the The SPEAK model and tool evaluation of e-learning is a model and tool designed to provide formative feedback for 24
decision making in developing and imple- The SPEAK system has been designed to in- menting an e-learning programme. It is based corporate a number of inter-related functions on the CIPP model. There are three variants within a logical framework. of the tool for a) Educational institutions; • Strategic Planning: producing information b) small and medium enterprises; d) Large to assist both staff and management to make companies informed decisions. • Evaluation: ongoing internal review; cross Models and tools for evaluating e-learning policy project periodic assessments; with a facility The models and tools for evaluating e-learn- for external inputs. ing policy are based on the framework for • Knowledge: networking based on sharing the evaluation of e-learning. The five differ- information between institutes, countries, ent groups of variables are further broken regions and themes. down to provide differentiating factors or The SPEAK model has been developed as an criteria against which a policy can be evalu- electronic tool distributed by a CD ROM. ated. The tool is designed for uses in policy Data can be aggregated and queries through evaluation at any level – national, regional or a server application. institutional. Tool for the evaluation of the effectiveness of e-learning programmes in small and medium sized enterprises (SMEs) The model and tool for the evaluation of the effectiveness of e-Learning programs in SMEs has been designed to provide an easy-to- use instrument to carry out a retrospective evaluation of an e-learning program. This retrospective analysis should enable enter- prises to detect weaknesses and strengths of their learning program with regard to its organisational, pedagogic and technological implications. Evalact Evalact is an electronic tool, designed to allow facilitate the evaluation of individual learning. It allows the creation of on-line questionnaires and provides rich feedback and statistical evidence for evaluators. It also allows the comparison of different evaluation instances and provides graphical interfaces for assisting in the analysis of evaluation data. The SPEAK model and tool SPEAK is a model and tool for the self evalu- ation of learning in a group and community context. It is designed both to facilitate group discussion and self evaluation and to provide accumulative data for programme evaluation. 25
You can also read