Methodological perspectives: Standardized (summative) or contextualized (formative) evaluation?
A critical issue in educational evaluation is whether evaluations should focus on standardized (summative, often quantitative) or contextualized (formative or often qualitative) evidence. The author of this article advises readers to beware of false dichotomies. The big issue is not whether evaluati...
Main Author: | |
---|---|
Format: | Article |
Language: | English |
Published: |
Arizona State University
2018-04-01
|
Series: | Education Policy Analysis Archives |
Subjects: | |
Online Access: | https://epaa.asu.edu/ojs/article/view/3813 |
id |
doaj-97481880d5294a26b5e9fc58f656f0e3 |
---|---|
record_format |
Article |
spelling |
doaj-97481880d5294a26b5e9fc58f656f0e32020-11-25T03:24:20ZengArizona State UniversityEducation Policy Analysis Archives1068-23412018-04-0126010.14507/epaa.26.38131743Methodological perspectives: Standardized (summative) or contextualized (formative) evaluation?Richard J. Shavelson0Stanford University (Emeritus)A critical issue in educational evaluation is whether evaluations should focus on standardized (summative, often quantitative) or contextualized (formative or often qualitative) evidence. The author of this article advises readers to beware of false dichotomies. The big issue is not whether evaluations should be “standardized” or “contextualized” but rather whether the evidence collected rigorously addresses the policy and/or practice questions driving the evaluation. The questions asked, in turn, lead to evaluation designs which may be standardized (summative), contextualized (formative) or both. Three general questions drive research and evaluation: (1) Descriptive—What’s Happening? (2) Causal—Is there a systematic effect? and (3) Process or mechanism—Why or how is it happening? Depending on the nature of the question, summative and/or formative data might be collected. Equally important are politics, measurement methods and modeling in conducting evaluations. Ignore these matters at your peril. Concrete examples show how assumptions and misperceptions can upend or change the outcomes of evaluation; they are drawn from political, measurement and statistical modeling contexts.https://epaa.asu.edu/ojs/article/view/3813Summative EvaluationFormative EvaluationEvaluation MethodsPolitics of Evaluation |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Richard J. Shavelson |
spellingShingle |
Richard J. Shavelson Methodological perspectives: Standardized (summative) or contextualized (formative) evaluation? Education Policy Analysis Archives Summative Evaluation Formative Evaluation Evaluation Methods Politics of Evaluation |
author_facet |
Richard J. Shavelson |
author_sort |
Richard J. Shavelson |
title |
Methodological perspectives: Standardized (summative) or contextualized (formative) evaluation? |
title_short |
Methodological perspectives: Standardized (summative) or contextualized (formative) evaluation? |
title_full |
Methodological perspectives: Standardized (summative) or contextualized (formative) evaluation? |
title_fullStr |
Methodological perspectives: Standardized (summative) or contextualized (formative) evaluation? |
title_full_unstemmed |
Methodological perspectives: Standardized (summative) or contextualized (formative) evaluation? |
title_sort |
methodological perspectives: standardized (summative) or contextualized (formative) evaluation? |
publisher |
Arizona State University |
series |
Education Policy Analysis Archives |
issn |
1068-2341 |
publishDate |
2018-04-01 |
description |
A critical issue in educational evaluation is whether evaluations should focus on standardized (summative, often quantitative) or contextualized (formative or often qualitative) evidence. The author of this article advises readers to beware of false dichotomies. The big issue is not whether evaluations should be “standardized” or “contextualized” but rather whether the evidence collected rigorously addresses the policy and/or practice questions driving the evaluation. The questions asked, in turn, lead to evaluation designs which may be standardized (summative), contextualized (formative) or both. Three general questions drive research and evaluation: (1) Descriptive—What’s Happening? (2) Causal—Is there a systematic effect? and (3) Process or mechanism—Why or how is it happening? Depending on the nature of the question, summative and/or formative data might be collected. Equally important are politics, measurement methods and modeling in conducting evaluations. Ignore these matters at your peril. Concrete examples show how assumptions and misperceptions can upend or change the outcomes of evaluation; they are drawn from political, measurement and statistical modeling contexts. |
topic |
Summative Evaluation Formative Evaluation Evaluation Methods Politics of Evaluation |
url |
https://epaa.asu.edu/ojs/article/view/3813 |
work_keys_str_mv |
AT richardjshavelson methodologicalperspectivesstandardizedsummativeorcontextualizedformativeevaluation |
_version_ |
1724602167358128128 |