logo
ResearchBunny Logo
Evaluating the capacity of large language models to interpret emotions in images

Computer Science

Evaluating the capacity of large language models to interpret emotions in images

H. Alrasheed, A. Alghihab, et al.

Discover how GPT-4 can streamline emotional stimulus selection by rating visual images on valence and arousal, closely approximating human judgments under zero-shot conditions while noting challenges with subtler cues. This research was conducted by Hend Alrasheed, Adwa Alghihab, Alex Pentland, and Sharifa Alghowinem.

00:00
00:00
~3 min • Beginner • English
Abstract
The integration of artificial intelligence, specifically large language models (LLMs), in emotional stimulus selection and validation offers a promising avenue for enhancing emotion comprehension frameworks. Traditional methods in this domain are often labor-intensive and susceptible to biases, highlighting the need for more efficient and scalable alternatives. This study evaluates the capability of GPT-4, in recognizing and rating emotions from visual stimuli, focusing on two primary emotional dimensions: valence (positive, neutral, or negative) and arousal (calm, neutral, or stimulated). By comparing the performance of GPT-4 against human evaluations using the well-established Geneva Affective PicturE Database (GAPED), we aim to assess the model's efficacy as a tool for automating the selection and validation of emotional elicitation stimuli. Our findings indicate that GPT-4 closely approximates human ratings under zero-shot learning conditions, although it encounters some difficulties in accurately classifying subtler emotional cues. These results underscore the potential of LLMs to streamline the emotional stimulus selection and validation process, thereby reducing the time and labor associated with traditional methods.
Publisher
PLOS ONE
Published On
Jun 03, 2025
Authors
Hend Alrasheed, Adwa Alghihab, Alex Pentland, Sharifa Alghowinem
Tags
large language models
GPT-4
emotional stimulus selection
valence
arousal
GAPED
zero-shot learning
Listen, Learn & Level Up
Over 10,000 hours of research content in 25+ fields, available in 12+ languages.
No more digging through PDFs, just hit play and absorb the world's latest research in your language, on your time.
listen to research audio papers with researchbunny