分级(工程)
数学教育
心理学
质量(理念)
高等教育
教育学
计算机科学
认识论
政治学
工程类
哲学
土木工程
法学
作者
Afnan Almegren,Hassan Saleh Mahdi,Abduljalil Nasr Hazaea,Jamal Kaid Mohammed Ali,Rehan Almegren
标识
DOI:10.1080/14703297.2024.2437122
摘要
This study aimed to explore how artificial intelligence (AI) tools compare with humans in evaluating the essays written by students in a writing course. Using a dataset of 30 essays written by English as a foreign language (EFL) students, the evaluations by the AI tools were compared with those of human evaluators, to examine whether the AI evaluations differed with respect to the quality of the entire essay or specific categories (i.e., content, vocabulary, organization, and accuracy). The results indicated that the AI tools provided high-quality feedback to students across all categories despite differences regarding essay quality. Additionally, AI tools differed in the scores they assigned, consistently grading lower than human raters across multiple evaluation categories while providing more detailed feedback than human raters. The scores assigned by each AI tool for student essays across various assessment categories did not differ significantly from the overall scores assigned by AI tools.
科研通智能强力驱动
Strongly Powered by AbleSci AI