June 23, 2013
June 23, 2013
June 26, 2013
23.121.1 - 23.121.12
A Teaching Assistant Training Protocol for Improving Feedback on Open-Ended Engineering Problems in Large ClassesTeaching Assistants (TAs) are vital to teaching large classes. TAs often function as students’primary contact within a large course, and, in many cases, they evaluate the majority of studentwork on assignments. For TAs, evaluating students’ work on open-ended problems ischallenging because students produce a variety of solutions that the TA must interpret toaccurately apply a given rubric. Reliable evaluation of student work is desirable. This paperexplores a TA training protocol for identifying TAs who are in need of additional guidance onhow to evaluate students’ work on open-ended engineering problems in a large class.Model-Eliciting Activities (MEAs) are authentic, open-ended, client-driven, engineering-basedmathematical modeling problems. Teams of students develop a written document describingtheir generalizable procedure (mathematical model) for solving a given problem and similarproblems. MEAs have been conducted at a midwestern university since 2002 in the large first-year engineering courses. In Fall 2012, the enrollment was approximately 1700 students, thecourse was staffed by 9 Graduate TAs, 70 Undergraduate TAs, 11 faculty, and 3 staff. All TAs(both graduate and undergraduate) are the primary point-of-contact for students and evaluatealmost all class assignments, including MEA solutions. The use of undergraduate TAs,including sophomore juniors, and seniors, to evaluate student work on MEAs is relatively new.Ensuring that each of the approximately 80 TAs is adequately prepared to reliably evaluatestudent work is a significant challenge.TAs, currently, engage in approximately 5 hours of training per MEA taught in the two-semestersequence required first-year engineering courses. The focus of this training is on evaluating andproviding feedback on students work on the MEAs slated for implementation in a givensemester. After the face-to-face training, TAs practice evaluating prototypical student solutionsand compare their evaluations to those of an expert. Following each evaluation, the TAs areasked to critically reflect on how they can improve their subsequent evaluations. Uponcompletion of the training sequence, the TAs’ evaluations of the sample solutions are examinedby the course coordinator to identifying which TAs need additional help to better align theirevaluations and feedback to an expert’s. Identified TAs are then given additional guidance onhow to improve their evaluations.This paper will do the following: (1) explore the history and need for TA training, (2) describethe context in which training occurs, (3) describe the training process and protocol in detail, (4)examine, quantitatively and qualitatively, historical and current data to explore the effectivenessof the TA training protocol, and (5) identify future changes that should be made to improve howTAs are being trained. Such an overview and analysis will provide insights for others whostruggle to bring open-ended problem solving in to large courses because of the demands forreliable evaluation of student work.
ASEE holds the copyright on this document. It may be read by the public free of charge. Authors may archive their work on personal websites or in institutional repositories with the following citation: © 2013 American Society for Engineering Education. Other scholars may excerpt or quote from these materials with the same citation. When excerpting or quoting from Conference Proceedings, authors should, in addition to noting the ASEE copyright, list all the original authors and their institutions and name the host city of the conference. - Last updated April 1, 2015