Assessing Student Performance in Science
David Haury

Haury, D.(1993, July). Assessing student performance in science. ERIC CSMEE Digest. ERIC Clearinghouse for Science, Mathematics, and Environmental Education, Columbus, Ohio. ED359068


Return to: | Readings in Educational Psychology | Educational Psychology Interactive |


With increased student achievement in science being a national goal (AMERICA 2000, 1991), do we know what students are learning? Given the emerging national standards in science education National Committee on Science Education Standards and Assessment (NCSESA), 1993, how will we determine whether students measure up to the standards? Radical changes are underway for school science curricula Rutherford & Ahlgren, 1990; The National Science Teachers Association (NSTA), 1992, but are complementary changes in assessment in progress? States are developing student assessments based on science frameworks or guides (Blank & Engler, 1992; Davis & Armstrong, 1991), but do we know how to assess student performance in all the domains of interest and concern? Assessment of student performance is emerging as a crucial ingredient in the recipe for ongoing improvement of school science. As programmatic change is occurring, there is a need to align student assessment practices with curricular aims, instructional practices, and performance standards. In short, "What we teach must be valued; what we test is what must be taught" (Iris Carl as quoted in McKinney, 1993).

FUNCTIONS ON ASSESSMENT

Before considering alternative approaches to assessing student performance, it is important to consider the various functions that assessment serves. Various reasons for assessing student performance have been described in both specific terms (Kober, 1993, pp. 57-58; Raizen & Kaser, 1989) and general terms (Meng & Doran, 1993), with distinctions being made between assessment for reporting purposes and for purposes of diagnosis and program evaluation. Eventually, as National Science Education Standards are developed, guidelines for assessing both student achievement and program effectiveness will be developed (NCSESA, 1993).

In this digest, the focus is on assessment in the service of instruction, for helping students, teachers, and parents monitor learning. Assessment in this context must be unobtrusive and tailored to measure specific learning outcomes, not necessarily norm-referenced and generalizable across schools, states, and countries (Haertel, 1991). What are the issues and methods of assessment in the context of classroom instruction?

NEW FORMS OF ASSESSMENT

Jorgensen (1993) cited the plethora of new labels for assessment strategies in contrast to the scarcity of expertise, procedures, and guidelines as evidence of a paradigm shift in assessment. There is an urgency to develop innovative forms of assessment that provide valid indicators of student understanding and other learning outcomes, but there is widely divided opinion as to how to proceed. Even though teachers know that students who are most knowledgeable in science are not necessarily the ones that get the highest grades, most continue to depend on multiple-choice test scores to determine grades (Baron, 1990). Furthermore, there is evidence that standardized testing strongly influences instructional planning among teachers (Herman & Golan, 1992). So, "one thing must be made clear from the outset: assessment encompasses more than testing, and much more than standardized testing. It includes such techniques as systematic teacher observation and so-called 'authentic' assessment, in which the tasks assessed more closely parallel the learning activities and outcomes that are desirable in the science classroom" (Kober, 1993).

Among the new labels being used today is performance-based assessment. Though there are a variety of definitions, it is clear that performance-based assessment does not include multiple-choice testing or related paper-and-pencil approaches. According to Jorgensen (1993), "performance-based assessment requires that the student complete, demonstrate, or perform the actual behavior of interest. There is a minimal degree of inference involved." Baron (1991) has provided a list of characteristics of performance assessment tasks, with a notable blending of content with process, and major concepts with specific problems. As Kober (1993) has mentioned, "in this type of assessment, students may work together or separately, using the equipment, materials, and procedures they would use in good, hands-on science instruction."

GUIDELINES FOR SCHOOLS AND CLASSROOMS

Today's assessment strategies must be aligned with the emerging vision of "science for all," with all students engaged in science experiences that "teach the nature and process of science as well as the subject matter" (NCSESA, 1993). A first step in considering assessment methods, then, is to become familiar with the wide range of student outcomes that are being endorsed by science teachers (NSTA, 1992), scientists (Rutherford & Ahlgren, 1990), and the National Research Council (NCSESA, 1993). It is also necessary to consider the diverse needs, interests, and abilities of students, particularly girls, minorities, students with disabilities, and those with limited English proficiency. "Assessment should be context dependent; reflect the nature of the subject matter; and address the unique cultural aspects of class, school, and community among culturally diverse populations" (Tippins & Dana, 1992). As White and Gunstone (1992) have stated, "a limited range of tests promotes limited forms of understanding. As an example of how alternative assessment strategies can enable students to show what they know in a variety of knowledge domains, consider the approach taken in one urban school (Dana, Lorsbach, Hook, & Briscoe, 1991). Concept mapping and journal writing techniques are used to document conceptual change among students, and student presentation and interview techniques allow learners to communicate their understanding in ways that rely less on reading and writing skills. For additional samples of techniques being used, see the Appendix of Kulm and Malcom (1991).

Among the promising alternative assessment techniques are the use of scoring rubrics to monitor skill development and the use of portfolios to assemble evidence of skill attainment. Scoring rubrics can be used to clarify for both students and teachers how valued skills are being measured (Nott, Reeve, & Reeve, 1992). Portfolios documenting student accomplishments can take a variety of forms, with student products, collected data, or other evidence of performance being used as information for self, peer, or teacher evaluation (Collins, 1992).

It should be acknowledged that there are drawbacks to performance assessments. Staff development will be required, performance assessments take more time than conventional methods, standardization is difficult, and the results may not be generalizable from one context to another. These problems reinforce the importance of practitioners, assessment specialists, and assessment "consumers" being clear on the purposes of specific assessment activities. There is no one approach to assessment that will best serve all functions, knowledge domains, and learners.

Resources

References

-----

David Haury is an Associate Professor of Science Education at The Ohio State University and the Director of the ERIC Clearinghouse for Science, Mathematics, and Environmental Education.

-----

THIS DIGEST WAS CREATED BY ERIC, THE EDUCATIONAL RESOURCES INFORMATION CENTER. FOR MORE INFORMATION ABOUT ERIC, CONTACT ACCESS ERIC 1-800-LET-ERIC

This digest was funded by the Office of Educational Research and Improvement, U. S. Department of Education under contract no. RI88062006. Opinions expressed in this digest do not necessarily reflect the positions or policies of OERI or the Department of Education.

This digest is in the public domain and may be freely reproduced.