TY - GEN
T1 - A Comparison Between Transformers and Foundation Models in Sentiment Analysis of Student Evaluation of Teaching
AU - Vega, Ines
AU - Valencia, Jose
AU - Arcos, Angel
AU - Navarrete, Danny
AU - Baldeon-Calisto, Maria
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Student evaluation of teaching (SET) serves as a crucial tool for improving educational quality, enabling students to articulate their opinions about instructors. However, manually evaluating student feedback is time-consuming, subjective, and prone to error. Sentiment analysis, which automatically classifies texts using computational algorithms, presents a promising alternative for this task. In this work, we conduct a comparative analysis of sentiment analysis on SET between three Transformer networks and three Foundation models on a dataset from an Ecuadorian university. Our experiments demonstrate that Transformer models trained on the dataset of interest have a better overall performance than general-purpose Foundation models. Furthermore, among the models examined, DistilBERT emerges as the top performer, achieving an accuracy of 84.90% and an F-1 score of 0.836. In comparison, among the Foundation models, Google Bard achieves the highest accuracy and F-1 score with 78.3% and 0.767, respectively. This work contributes valuable insights to the realm of higher education evaluation, showcasing the potential of advanced NLP techniques to expedite and enhance the SET process, ultimately paving the way for continuous improvement in educational settings.
AB - Student evaluation of teaching (SET) serves as a crucial tool for improving educational quality, enabling students to articulate their opinions about instructors. However, manually evaluating student feedback is time-consuming, subjective, and prone to error. Sentiment analysis, which automatically classifies texts using computational algorithms, presents a promising alternative for this task. In this work, we conduct a comparative analysis of sentiment analysis on SET between three Transformer networks and three Foundation models on a dataset from an Ecuadorian university. Our experiments demonstrate that Transformer models trained on the dataset of interest have a better overall performance than general-purpose Foundation models. Furthermore, among the models examined, DistilBERT emerges as the top performer, achieving an accuracy of 84.90% and an F-1 score of 0.836. In comparison, among the Foundation models, Google Bard achieves the highest accuracy and F-1 score with 78.3% and 0.767, respectively. This work contributes valuable insights to the realm of higher education evaluation, showcasing the potential of advanced NLP techniques to expedite and enhance the SET process, ultimately paving the way for continuous improvement in educational settings.
KW - Artificial Intelligence
KW - Foundation Models
KW - Sentiment Analysis
KW - Student Evaluation of Teaching
KW - Transformers Models
UR - http://www.scopus.com/inward/record.url?scp=85194080122&partnerID=8YFLogxK
U2 - 10.1109/ISDFS60797.2024.10527264
DO - 10.1109/ISDFS60797.2024.10527264
M3 - Contribución a la conferencia
AN - SCOPUS:85194080122
T3 - 12th International Symposium on Digital Forensics and Security, ISDFS 2024
BT - 12th International Symposium on Digital Forensics and Security, ISDFS 2024
A2 - Varol, Asaf
A2 - Karabatak, Murat
A2 - Varol, Cihan
A2 - Tuba, Eva
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 12th International Symposium on Digital Forensics and Security, ISDFS 2024
Y2 - 29 April 2024 through 30 April 2024
ER -