Baltimore , Maryland
June 25, 2023
June 25, 2023
June 28, 2023
Computers in Education Division (COED)
7
10.18260/1-2--42888
https://peer.asee.org/42888
141
Matthew Norris is a PhD student and Graduate Research Assistant in the Department of Engineering Education at Virginia Tech.
Hamidreza is a Ph.D. student in Engineering Education and has a master's degree in industrial engineering at Virginia Tech (VT). He has worked in the industry as a research and development engineer. He is currently a data analyst in TLOS (Technology-Enhanced Learning And Online Strategies) at VT. His expertise is in quantitative research. His primary research interest is motivation, co-curricular activities, and professional development in engineering education.
Andrew Katz is an assistant professor in the Department of Engineering Education at Virginia Tech. He leads the Improving Decisions in Engineering Education Agents and Systems (IDEEAS) Lab, a group that uses multi-modal data to characterize, understand, a
Jacob Grohs is an Assistant Professor in Engineering Education at Virginia Tech with Affiliate Faculty status in Biomedical Engineering and Mechanics and the Learning Sciences and Technologies at Virginia Tech. He holds degrees in Engineering Mechanics (
In engineering education contexts, assessing socio-technical skills, such as systems thinking or risk assessment, is a complicated and difficult task. In many cases self-report scales are the most common, and sometimes only, existing available assessment tools to evaluate students’ abilities. Scenario-based and case-based assessments have risen as an alternative to counteract the response shift biases that afflict many self-report scale pre/post tests. Unfortunately, while these scenario-based assessments may offer more reliable measures of students’ socio-technical skills, the process of scoring scenario responses is time intensive even with trained raters and detailed scoring guides. Additionally, the time to score each scenario response plateaus once raters reach high proficiency, limiting scenario-based assessments’ usefulness as formative assessment tools. Larger pools of textual data also create more challenges for intrarater reliability as raters’ interpretation of scoring guides may drift over time. To address this barrier, we have created a natural language processing system that augments scoring by preprocessing textual responses and assigning in line with a developed scoring guide. Specifically, we take responses to a scenario-based assessment and a detailed scoring guide accompanying the assessment and utilize term extraction to categorize common terms from the response based on categories from the scoring guide. Responses containing phrases that meet these scoring categories are then identified and extracted from the raw text and presented alongside that raw text to the human rater. Such a system can speed up the scoring process by performing a first pass of responses and identifying the presence of certain content specified by the accompanying scoring guide. The human rater is then able to briefly check the accuracy of the system’s categorization and assign a score. This accelerates the scoring process by identifying and highlighting salient content from the raw text and narrowing the range of prospective scores that a rater must consider. The system also improves consistency by applying the same categorization across the collection of text simultaneously. This is in contrast with a single person or a team that would analyze responses sequentially with the aid of the scoring guide, creating the potential for inconsistencies across time as the rater builds familiarity with the guide. In this paper we describe the system’s architecture, data processing steps, and preliminary results. We demonstrate the utility of the system by applying it on an open-ended question from a scenario-based assessment targeting systems thinking in domain general contexts. This instance of the scenario was administered to undergraduate students across disciplines as part of both a statistics and introductory humanities course. Given students’ numerous undergraduate disciplines and knowledge domains, this presented a varied and challenging dataset. Our preliminary results suggest that pre-processing of textual content can improve the speed and reliability of scoring when compared to unassisted human scoring with the same scoring guide. As natural language processing methods continue to advance, applications to augment textually focused assessment like scenario and case-based should continue to be explored.
Norris, M., & Taimoory, H., & Katz, A., & Grohs, J. R. (2023, June), Board 65: Work in Progress: Using Natural Language Processing to Facilitate Scoring of Scenario-Based Assessments Paper presented at 2023 ASEE Annual Conference & Exposition, Baltimore , Maryland. 10.18260/1-2--42888
ASEE holds the copyright on this document. It may be read by the public free of charge. Authors may archive their work on personal websites or in institutional repositories with the following citation: © 2023 American Society for Engineering Education. Other scholars may excerpt or quote from these materials with the same citation. When excerpting or quoting from Conference Proceedings, authors should, in addition to noting the ASEE copyright, list all the original authors and their institutions and name the host city of the conference. - Last updated April 1, 2015