Application of new automated scoring technologies, such as natural language processing and machine learning, makes it possible to provide automated feedback on students' short written responses. Even though many studies investigated the automated feedback in the computer-mediated learning environments, most of them focused on the multiple-choice items instead of the constructed response items. This study focuses on the latter and investigates a formative feedback system integrated into an online science curriculum module teaching climate change.
As teacher education shifts to focus on teaching beginners to do the work of teaching, assessments need to shift to focus on assessing practice. We focus on one teaching practice, eliciting student thinking, in the context of elementary mathematics. We describe assessments in two contexts (field and simulation). For each assessment, we describe the eliciting of three prospective teachers what could be seen about the skills of group of prospective teachers (N = 44).
We report on the use of bilingual constructed response science assessments in the context of a research and development partnership with secondary school science teachers. Given the power that assessments have in today’s education systems, our project provided a series of workshops for teachers where they explored students’ emergent reform-oriented science meaning-making in our project-designed assessments.
This paper illustrates how the combination of teacher and computer guidance can strengthen collaborative revision and identifies opportunities for teacher guidance in a computer-supported collaborative learning environment. We took advantage of natural language processing tools embedded in an online, collaborative environment to automatically score student responses using human-designed knowledge integration rubrics. We used the automated explanation scores to assign adaptive guidance to the students and to provide real-time information to the teacher on students’ learning.
Test scoring procedures should align with the intended uses and interpretations of test results. In this paper, we examine three test scoring procedures for an operational assessment of early numeracy, the Early Grade Mathematics Assessment (EGMA). Current test specifications call for subscores to be reported for each of the eight subtests on the EGMA. This test scoring procedures has been criticized as being difficult for stakeholders to use and interpret, thereby impacting the overall usefulness of the EGMA for informing decisions.
Dig deeper into classroom artifacts using research-based learning progressions to enhance your analysis and response to student work, even when most students solve a problem correctly.
Ebby, C. B., Hulbert, E. T., and Fletcher, N. (2019). What can we learn from correct answers? Teaching Children Mathematics, 25(6), 346-353.
Ambitious efforts are taking place to implement a new vision for science education in the United States, in both Next Generation Science Standards (NGSS)-adopted states and those states creating their own, often related, standards. Inservice and pre-service teacher educators are involved in supporting teacher shifts in practice toward the new standards. With these efforts, it will be important to document shifts in science instruction toward the goals of NGSS and broader science education reform.
This paper describes HASbot, an automated text scoring and real‐time feedback system designed to support student revision of scientific arguments. Students submit open‐ended text responses to explain how their data support claims and how the limitations of their data affect the uncertainty of their explanations. HASbot automatically scores these text responses and returns the scores with feedback to students. Data were collected from 343 middle‐ and high‐school students taught by nine teachers across seven states in the United States.
Cirillo, M. & Hummer, J. (2019). Addressing misconceptions in secondary geometry proof. Mathematics Teacher, 112(6).
The purpose of this study was to develop and validate a survey of opportunities to participate (OtP) in science that will allow educators and researchers to closely approximate the types of learning opportunities students have in science classrooms. Additionally, we examined whether and how opportunity gaps in science learning may exist across schools with different socioeconomic levels. The OtP in science survey consists of four dimensions that include acquiring foundational knowledge, planning an investigation, conducting an investigation, and using evidence to communicate findings.