Yayın:
Is conversational ai ready for engineering licensure? A chatgpt performance benchmark and prompting strategy evaluation

dc.contributor.buuauthorGENÇ, OLCAY
dc.contributor.departmentMühendislik Fakültesi
dc.contributor.departmentİnşaat Mühendisliği Ana Bilim Dalı
dc.contributor.researcheridAFH-5568-2022
dc.date.accessioned2025-11-06T16:39:28Z
dc.date.issued2025-10-18
dc.description.abstractThe integration of artificial intelligence (AI) into specialized fields like civil engineering presents transformative opportunities. This study provides a comprehensive quantitative benchmark of the ChatGPT-4 model's performance on the Fundamentals of Engineering (FE) Civil Exam. A two-phase methodology was employed. First, the model's baseline performance was evaluated on a dataset of 100 representative exam questions using zero-shot prompting. Second, a follow-up experiment investigated the impact of advanced prompting strategies, including one-shot, on the 51 incorrectly answered questions. The initial findings reveal a significant performance disparity: ChatGPT-4 achieved high accuracy in conceptual domains like "Ethics and Professional Practice" (100%) but struggled in calculation-intensive areas such as "Statics" (14%), with an overall accuracy of 49%. The subsequent prompt engineering experiment, however, demonstrated that providing a single example (one-shot prompting) was the most effective strategy, correctly answering 30 of the 51 previously failed questions. These combined results offer critical evidence that while current LLMs have inherent limitations in analytical reasoning, their effectiveness can be substantially enhanced through strategic user interaction. The study concludes that AI should be implemented as a powerful supplemental tool, with an educational focus on teaching students how to effectively guide these models to achieve desired outcomes.
dc.identifier.doi10.1080/13467581.2025.2574556
dc.identifier.issn1346-7581
dc.identifier.scopus2-s2.0-105019232462
dc.identifier.urihttps://doi.org/10.1080/13467581.2025.2574556
dc.identifier.urihttps://hdl.handle.net/11452/56566
dc.identifier.wos001594517800001
dc.indexed.wosWOS.SCI
dc.indexed.wosWOS.AHCI
dc.language.isoen
dc.publisherTaylor & francis ltd
dc.relation.journalJournal of asian architecture and building engineering
dc.subjectArtificial intelligence in education
dc.subjectChatGPT-4
dc.subjectCivil Engineering education
dc.subjectconversational AI
dc.subjectLLM
dc.subjecteducational technology integration
dc.subjectArts & Humanities
dc.subjectScience & Technology
dc.subjectTechnology
dc.subjectArchitecture
dc.titleIs conversational ai ready for engineering licensure? A chatgpt performance benchmark and prompting strategy evaluation
dc.typeArticle
dspace.entity.typePublication
local.contributor.departmentMühendislik Fakültesi/İnşaat Mühendisliği Ana Bilim Dalı
local.indexed.atWOS
local.indexed.atScopus
relation.isAuthorOfPublication2d57a04e-3183-4474-a6b0-1e54038d3d1c
relation.isAuthorOfPublication.latestForDiscovery2d57a04e-3183-4474-a6b0-1e54038d3d1c

Dosyalar

Orijinal seri

Şimdi gösteriliyor 1 - 1 / 1
Küçük Resim
Ad:
Genc_2025.pdf
Boyut:
1.51 MB
Format:
Adobe Portable Document Format