Using Large Language Models to Analyze Interviews for Driver Psychological Assessment: A Performance Comparison of ChatGPT and Google-Gemini

1Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

This study examines the application of large language models (LLMs) in analyzing subjective driver perceptions during tunnel driving simulations, comparing the effectiveness of questionnaires and interviews. Building on previous research involving driver simulations, we recruited 29 new participants, collected their perceptions via questionnaires, and conducted follow-up interviews. The interview data were analyzed using three LLMs: GPT-3.5, GPT-4, and Google-Gemini. The results revealed that while GPT-4 provides more in-depth and accurate analysis, it is significantly slower than GPT-3.5. Conversely, Google-Gemini demonstrated a balance between analysis quality and speed, outperforming the other models overall. Despite the challenge of occasional misunderstandings, LLMs still have the potential to enhance the efficiency and accuracy of subjective data analysis in transportation research.

Cite

CITATION STYLE

APA

Sun, R., Hu, X., Shao, Y., Luo, Z., Liu, B., & Cheng, Y. (2025). Using Large Language Models to Analyze Interviews for Driver Psychological Assessment: A Performance Comparison of ChatGPT and Google-Gemini. Symmetry, 17(10). https://doi.org/10.3390/sym17101713

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free