GPT-4기반 과학적 탐구 글쓰기 역량의 평가 자동화를 위한 언어 모델의 투명성과 시각화에 관한 연구
Automated Robot Agent for Evaluating the Scientific Inquiry Capabilities via GPT-4: A Study on the Transparency and Visualization in Large Language Model
Citations

WEB OF SCIENCE

0
Citations

SCOPUS

0

초록

This study investigates GPT-4’s ability to assess scientific inquiry writing tasks, focusing on its capacity to justify evaluations with evidence. Scientific inquiry involves problem definition, hypothesis formulation, data analysis, and experimental design. Using scenario-based scientific questions, GPT-4’s responses were compared with human evaluators to analyze its reasoning and limitations. Results indicate that GPT-4 can provide valid reasoning under specific conditions but struggles with complex data interpretation. This research highlights GPT-4’s potential and limitations as an automated evaluation tool, addressing inconsistencies in human evaluation and suggesting future directions for AI-assisted educational assessments.

키워드

Automated AgentLarge Language ModelChatGPTScientific InquiryCapability EvaluationEvidence Extraction
제목
GPT-4기반 과학적 탐구 글쓰기 역량의 평가 자동화를 위한 언어 모델의 투명성과 시각화에 관한 연구
제목 (타언어)
Automated Robot Agent for Evaluating the Scientific Inquiry Capabilities via GPT-4: A Study on the Transparency and Visualization in Large Language Model
저자
윤지나이효정이성혜박소영동서연심주용
DOI
10.7746/jkros.2025.20.2.200
발행일
2025-06
저널명
로봇학회 논문지
20
2
페이지
200 ~ 209