Logo des Repositoriums
  • English
  • Deutsch
Anmelden
Keine TU-ID? Klicken Sie hier für mehr Informationen.
  1. Startseite
  2. Publikationen
  3. Publikationen der Technischen Universität Darmstadt
  4. Zweitveröffentlichungen (aus DeepGreen)
  5. Cheating Automatic Short Answer Grading with the Adversarial Usage of Adjectives and Adverbs
 
  • Details
2024
Zweitveröffentlichung
Artikel
Verlagsversion

Cheating Automatic Short Answer Grading with the Adversarial Usage of Adjectives and Adverbs

File(s)
Download
Hauptpublikation
s40593-023-00361-2.pdf
CC BY 4.0 International
Format: Adobe PDF
Size: 1.47 MB
TUDa URI
tuda/12465
URN
urn:nbn:de:tuda-tuprints-283790
DOI
10.26083/tuprints-00028379
Autor:innen
Filighera, Anna ORCID 0000-0001-5519-9959
Ochs, Sebastian
Steuer, Tim ORCID 0000-0002-3141-712X
Tregel, Thomas ORCID 0000-0003-0715-3889
Kurzbeschreibung (Abstract)

Automatic grading models are valued for the time and effort saved during the instruction of large student bodies. Especially with the increasing digitization of education and interest in large-scale standardized testing, the popularity of automatic grading has risen to the point where commercial solutions are widely available and used. However, for short answer formats, automatic grading is challenging due to natural language ambiguity and versatility. While automatic short answer grading models are beginning to compare to human performance on some datasets, their robustness, especially to adversarially manipulated data, is questionable. Exploitable vulnerabilities in grading models can have far-reaching consequences ranging from cheating students receiving undeserved credit to undermining automatic grading altogether—even when most predictions are valid. In this paper, we devise a black-box adversarial attack tailored to the educational short answer grading scenario to investigate the grading models’ robustness. In our attack, we insert adjectives and adverbs into natural places of incorrect student answers, fooling the model into predicting them as correct. We observed a loss of prediction accuracy between 10 and 22 percentage points using the state-of-the-art models BERT and T5. While our attack made answers appear less natural to humans in our experiments, it did not significantly increase the graders’ suspicions of cheating. Based on our experiments, we provide recommendations for utilizing automatic grading systems more safely in practice.

Freie Schlagworte

Assessment

Adversarial attacks

Automatic short answe...

Cheating

Natural language proc...

Autograder

Academic dishonesty

Constructed response

Fairness

Sprache
Englisch
Fachbereich/-gebiet
18 Fachbereich Elektrotechnik und Informationstechnik > Institut für Datentechnik > Multimedia Kommunikation
DDC
000 Allgemeines, Informatik, Informationswissenschaft > 004 Informatik
Institution
Universitäts- und Landesbibliothek Darmstadt
Ort
Darmstadt
Titel der Zeitschrift / Schriftenreihe
International Journal of Artificial Intelligence in Education
Startseite
616
Endseite
646
Jahrgang der Zeitschrift
34
Heftnummer der Zeitschrift
2
ISSN
1560-4306
Verlag
Springer
Ort der Erstveröffentlichung
New York
Publikationsjahr der Erstveröffentlichung
2024
Verlags-DOI
10.1007/s40593-023-00361-2
PPN
542395061
Ergänzende Ressourcen (Supplement)
https://github.com/SebOchs/adversarial_insertions.git

  • TUprints Leitlinien
  • Cookie-Einstellungen
  • Impressum
  • Datenschutzbestimmungen
  • Webseitenanalyse
Diese Webseite wird von der Universitäts- und Landesbibliothek Darmstadt (ULB) betrieben.