Abstract
Background: Mentalization, which is integral to human cognitive processes, pertains to the interpretation of one’s own and others’ mental states, including emotions, beliefs, and intentions. With the advent of artificial intelligence (AI) and the prominence of large language models in mental health applications, questions persist about their aptitude in emotional comprehension. The prior iteration of the large language model from OpenAI, ChatGPT-3.5, demonstrated an advanced capacity to interpret emotions from textual data, surpassing human benchmarks. Given the introduction of ChatGPT-4, with its enhanced visual processing capabilities, and considering Google Bard’s existing visual functionalities, a rigorous assessment of their proficiency in visual mentalizing is warranted. Objective: The aim of the research was to critically evaluate the capabilities of ChatGPT-4 and Google Bard with regard to their competence in discerning visual mentalizing indicators as contrasted with their textual-based mentalizing abilities. Methods: The Reading the Mind in the Eyes Test developed by Baron-Cohen and colleagues was used to assess the models’ proficiency in interpreting visual emotional indicators. Simultaneously, the Levels of Emotional Awareness Scale was used to evaluate the large language models’ aptitude in textual mentalizing. Collating data from both tests provided a holistic view of the mentalizing capabilities of ChatGPT-4 and Bard. Results: ChatGPT-4, displaying a pronounced ability in emotion recognition, secured scores of 26 and 27 in 2 distinct evaluations, significantly deviating from a random response paradigm (P
Author supplied keywords
- AI
- ChatGPT
- LLM
- LLMs
- RMET
- Reading the Mind in the Eyes Test
- algorithm
- algorithms
- artificial intelligence
- early detection
- early warning
- emotional awareness
- emotional comprehension
- emotional cue
- emotional cues
- empathy
- large language model
- large language models
- machine learning
- mental disease
- mental diseases
- mental health
- mental illness
- mental illnesses
- mentalization
- mentalizing
- practical model
- practical models
- predictive analytics
- predictive model
- predictive models
- predictive system
Cite
CITATION STYLE
Elyoseph, Z., Refoua, E., Asraf, K., Lvovsky, M., Shimoni, Y., & Hadar-Shoval, D. (2024). Capacity of Generative AI to Interpret Human Emotions From Visual and Textual Data: Pilot Evaluation Study. JMIR Mental Health, 11(1). https://doi.org/10.2196/54369
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.