Automated Assessment of Comprehension Strategies from Self-Explanations Using LLMs

Text comprehension is an essential skill in today’s information-rich world, and self-explanation practice helps students improve their understanding of complex texts. This study was centered on leveraging open-source Large Language Models (LLMs), specifically FLAN-T5, to automatically assess the com...

Full description

Bibliographic Details
Main Authors: Bogdan Nicula, Mihai Dascalu, Tracy Arner, Renu Balyan, Danielle S. McNamara
Format: Article
Language:English
Published: MDPI AG 2023-10-01
Series:Information
Subjects:
Online Access:https://www.mdpi.com/2078-2489/14/10/567
_version_ 1827720847071117312
author Bogdan Nicula
Mihai Dascalu
Tracy Arner
Renu Balyan
Danielle S. McNamara
author_facet Bogdan Nicula
Mihai Dascalu
Tracy Arner
Renu Balyan
Danielle S. McNamara
author_sort Bogdan Nicula
collection DOAJ
description Text comprehension is an essential skill in today’s information-rich world, and self-explanation practice helps students improve their understanding of complex texts. This study was centered on leveraging open-source Large Language Models (LLMs), specifically FLAN-T5, to automatically assess the comprehension strategies employed by readers while understanding Science, Technology, Engineering, and Mathematics (STEM) texts. The experiments relied on a corpus of three datasets (N = 11,833) with self-explanations annotated on 4 dimensions: 3 comprehension strategies (i.e., bridging, elaboration, and paraphrasing) and overall quality. Besides FLAN-T5, we also considered GPT3.5-turbo to establish a stronger baseline. Our experiments indicated that the performance improved with fine-tuning, having a larger LLM model, and providing examples via the prompt. Our best model considered a pretrained FLAN-T5 XXL model and obtained a weighted F1-score of 0.721, surpassing the 0.699 F1-score previously obtained using smaller models (i.e., RoBERTa).
first_indexed 2024-03-10T21:11:07Z
format Article
id doaj.art-bfaeca1d096f4e6491985b5ae61a3bfc
institution Directory Open Access Journal
issn 2078-2489
language English
last_indexed 2024-03-10T21:11:07Z
publishDate 2023-10-01
publisher MDPI AG
record_format Article
series Information
spelling doaj.art-bfaeca1d096f4e6491985b5ae61a3bfc2023-11-19T16:48:17ZengMDPI AGInformation2078-24892023-10-01141056710.3390/info14100567Automated Assessment of Comprehension Strategies from Self-Explanations Using LLMsBogdan Nicula0Mihai Dascalu1Tracy Arner2Renu Balyan3Danielle S. McNamara4Computer Science and Engineering Department, National University of Science and Technology POLITEHNICA of Bucharest, 313 Splaiul Independentei, 060042 Bucharest, RomaniaComputer Science and Engineering Department, National University of Science and Technology POLITEHNICA of Bucharest, 313 Splaiul Independentei, 060042 Bucharest, RomaniaDepartment of Psychology, Arizona State University, P.O. Box 871104, Tempe, AZ 85287, USAMath/CIS Department, SUNY Old Westbury, Old Westbury, NY 11568, USADepartment of Psychology, Arizona State University, P.O. Box 871104, Tempe, AZ 85287, USAText comprehension is an essential skill in today’s information-rich world, and self-explanation practice helps students improve their understanding of complex texts. This study was centered on leveraging open-source Large Language Models (LLMs), specifically FLAN-T5, to automatically assess the comprehension strategies employed by readers while understanding Science, Technology, Engineering, and Mathematics (STEM) texts. The experiments relied on a corpus of three datasets (N = 11,833) with self-explanations annotated on 4 dimensions: 3 comprehension strategies (i.e., bridging, elaboration, and paraphrasing) and overall quality. Besides FLAN-T5, we also considered GPT3.5-turbo to establish a stronger baseline. Our experiments indicated that the performance improved with fine-tuning, having a larger LLM model, and providing examples via the prompt. Our best model considered a pretrained FLAN-T5 XXL model and obtained a weighted F1-score of 0.721, surpassing the 0.699 F1-score previously obtained using smaller models (i.e., RoBERTa).https://www.mdpi.com/2078-2489/14/10/567language modelslarge language modelsself-explanationself-explanation strategies
spellingShingle Bogdan Nicula
Mihai Dascalu
Tracy Arner
Renu Balyan
Danielle S. McNamara
Automated Assessment of Comprehension Strategies from Self-Explanations Using LLMs
Information
language models
large language models
self-explanation
self-explanation strategies
title Automated Assessment of Comprehension Strategies from Self-Explanations Using LLMs
title_full Automated Assessment of Comprehension Strategies from Self-Explanations Using LLMs
title_fullStr Automated Assessment of Comprehension Strategies from Self-Explanations Using LLMs
title_full_unstemmed Automated Assessment of Comprehension Strategies from Self-Explanations Using LLMs
title_short Automated Assessment of Comprehension Strategies from Self-Explanations Using LLMs
title_sort automated assessment of comprehension strategies from self explanations using llms
topic language models
large language models
self-explanation
self-explanation strategies
url https://www.mdpi.com/2078-2489/14/10/567
work_keys_str_mv AT bogdannicula automatedassessmentofcomprehensionstrategiesfromselfexplanationsusingllms
AT mihaidascalu automatedassessmentofcomprehensionstrategiesfromselfexplanationsusingllms
AT tracyarner automatedassessmentofcomprehensionstrategiesfromselfexplanationsusingllms
AT renubalyan automatedassessmentofcomprehensionstrategiesfromselfexplanationsusingllms
AT daniellesmcnamara automatedassessmentofcomprehensionstrategiesfromselfexplanationsusingllms