File(s) under permanent embargo
Can software improve marker accuracy at detecting contract cheating? A pilot study of the Turnitin authorship investigate alpha
journal contributionposted on 2019-09-23, 00:00 authored by Phillip DawsonPhillip Dawson, Wendy Sutherland-SmithWendy Sutherland-Smith, M Ricksen
Contract cheating happens when students outsource their assessed work to a third party. One approach that has been suggested for improving contract cheating detection is comparing students’ assignment submissions with their previous work, the rationale being that changes in style may indicate a piece of work has been written by somebody else. This approach is time consuming, but recent advances in machine learning and natural language processing suggest that it may be well suited to computerization. We trialed an early alpha version of Turnitin’s Authorship Investigate tool, which compares students’ submissions against their previous work. Twenty-four experienced markers from five units of study were asked to make decisions about the presence of contract cheating in bundles of 20 student assignments, which included 14 legitimate assignments and six purchased from contract cheating sites. We asked markers to determine if each assignment was contract cheating, then provided them with an Authorship Investigate report and let them change their decision. Marker accuracy at detecting contract cheating increased significantly, from 48% to 59% after using the report, with no significant difference in false positives. These findings suggest that software may be an effective component of institutional strategies to address contract cheating.