Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Artif. Intell.

Sec. Medicine and Public Health

Visually grounded emotion regulation via diffusion models and user-driven reappraisal

Provisionally accepted
  • 1Leibniz Institute for Resilience Research, Mainz, Germany
  • 2Department of Psychiatry, Psychotherapy and Psychosomatic Medicine, University Medical Center Halle, Halle (Saale), Germany
  • 3German Center for Mental Health (DZPG), Site Halle-Jena-Magdeburg, Halle (Saale), Germany
  • 4Department of Psychiatry and Psychotherapy, University Medical Center of the Johannes Gutenberg-University Mainz, Mainz, Germany
  • 5School of Life Sciences, Technical University of Munich, Freising, Germany

The final, formatted version of the article will be published soon.

Cognitive reappraisal is a key strategy in emotion regulation, involving reinterpretation of emotionally charged stimuli to alter affective responses. Despite its central role in clinical and cognitive science, real-world reappraisal interventions remain cognitively demanding, abstract, and primarily verbal in nature. This reliance on higher-order cognitive and linguistic processes can be especially impaired in individuals with trauma, depression, or dissociative symptoms, limiting the effectiveness of standard approaches. Here, we propose a novel, visually based augmentation of cognitive reappraisal by integrating large-scale text-to-image diffusion models into the emotional regulation process. Specifically, we introduce a system wherein users reinterpret emotionally negative im-ages via spoken reappraisals, which are then transformed into supportive, emotionally congruent visualizations using stable diffusion models with a fine-tuned IP-adapter module. This generative transformation visually instantiates users' cognitive reappraisals while maintaining structural similarity to the original stimuli, thus externalizing and reinforcing regulatory intent. To evaluate this approach, we conducted a within-subjects experiment (N = 20) using a modified cognitive emotion regulation (CER) task. Participants reappraised or described aversive images from the international affective picture system (IAPS), with or without AI-generated visual feedback. Re-sults indicate that AI-assisted reappraisal significantly reduced negative affect relative to both non-AI reappraisal and control conditions. Further analyses show that sentiment alignment between participant reappraisals and generated images correlates with affective relief, suggesting that multimodal coherence enhances regulatory efficacy. Our findings highlight the feasibility of using generative visual support for cognitive reappraisal. This work opens a new interdisciplinary direction at the intersection of generative AI, affective computing, and therapeutic technology design.

Keywords: Affective Computing, affective responses, AI-generated visual feedback, aversive images, cognitive emotion regulation (CER) task, cognitive reappraisal, diffusion models, Emotion Regulation

Received: 23 Aug 2025; Accepted: 13 Jan 2026.

Copyright: © 2026 Pinzuti, Tüscher and Ferreira Castro. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: André Ferreira Castro

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.