COE HomeCollege ProgramsResearchOutreachReportsPeopleAlumniNewsSearch
Educational Research Reports
Understand Complex Student Outcomes
November 2000

The Article
In this article, Mary Kennedy, professor in the Department of Teacher Education, deals with the difficulty for education researchers of accurately gauging complex student learning as a result of policy changes. She analyzes and compares various assessment research by looking at such things as standardized tests, open-ended teacher interviews and observations. She examines and compares the various assessment tools.

Discussion
Education policy researchers face a difficult task in documenting student learning based on the implementation of new policies or changes to existing policies. It is extremely difficult, she says, to chart a clear path from policy change to student outcome. The key is that there is often no adequate measure of the outcomes researchers seek. There simply is no widely available and inexpensive (and generally agreed-upon) indicator or indicators of complex student learning. That puts researchers at a disadvantage when trying to gauge the influence of a particular policy or program. Instead, researchers have had to rely on more easily documented intermediate outcomes. Intermediate outcomes include such things as the impact a policy change has had on teachers’ attitudes and assuming that such an effect will lead to further changes in student outcomes. Kennedy breaks down these intermediate outcomes, known as approximations, into four levels, and shows that they vary widely in terms of how well they actually approximate an indicator of complex student learning. The first level is classroom observation and standardized tests. Standardized tests, she finds, have the advantage of being inexpensive, convenient, and perceived as valid indicators of student learning. Yet they also have been criticized as too narrow to gauge complex learning. Observations, on the other hand, are capable of documenting the intellectual complexity of the work students do in class. But like standardized tests, observations have shortcomings, including the fact that they cannot document the extent to which students have learned to handle complex problems. In addition, they are costly to conduct. The second level involves situated descriptions of teaching. In this case, researchers rely on teachers to obtain “as situated a description as possible of the teachers’ own teaching practice.” Common forms of situated descriptions are the teacher log and the vignette. Kennedy finds that this form of assessment also has pluses and minuses. The third level is the non-situated testimony about practice. This level relies on teacher questionnaires and interviews that ask about teaching practices but are not situated in specific teaching episodes. These types of questions elicit general responses such as “I always treat my students equally.” While used widely in survey research, Kennedy notes that non-situated testimony has only a “slight” relationship to complex student learning. The final level is testimony about effects of policies and programs. Kennedy finds this level even more removed than from non-situated testimony. In this level, teachers testify about whether a particular policy or program was helpful or whether it influenced teachers in some way. Although helpful in understanding teachers’ satisfaction or dissatisfaction with a policy or program, these approximations are “even more susceptible to self-serving biases than second- and third-level approximations are.” Kennedy concludes by drawing on the research to point out that the best first-level approximation available to researchers would be a classroom observation that focuses on the nature of intellectual work students do in class. Because of the expense, she suspects many researchers cannot afford this type of assessment. In such cases, she advises researchers that “it might be more fruitful to invest time in developing and field-testing questions that are closely situated, so that the distance between these questions and closer approximations is reduced as much as possible.”

Citation
Kennedy, M.M. (2000). Approximations to indicators of student outcomes. Education Evaluation and Policy Analysis, 21(4), 345-363.


<back to 2000 ed-Research Reports

| College of Education | MSU | Contact Us |