Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Neurosci.

Sec. Visual Neuroscience

Volume 19 - 2025 | doi: 10.3389/fnins.2025.1622194

Exploiting Facial Emotion Recognition System for Ambient Assisted Living Technologies Triggered by Interpreting the User's Emotional State

Provisionally accepted
  • 1Department of Psychology, Sapienza University of Rome, Italy, Rome, Italy
  • 2Sapienza University of Rome, Rome, Italy
  • 3eCampus University, Novedrate, Italy
  • 4University of Djelfa, Algeria, Djelfa, Algeria
  • 5Department of Computer Sciences, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia
  • 6School of ICT, Faculty of Engineering, Design and Information & Communications Technology (EDICT), Bahrain Polytechnic, Isa Town, Bahrain
  • 7Applied Science Research Center. Applied Science Private University, Amman, Jordan, Amman, Jordan
  • 8Department of Computer, Automation and Management Engineering, Sapienza University of Rome, Italy, Rome, Italy
  • 9Institute for Systems Analysis and Computer Science, Italian National Research Council, Rome, Italy
  • 10Department of Artificial Intelligence, Czestochowa University of Technology, Czestochowa, Poland

The final, formatted version of the article will be published soon.

Introduction: Facial Emotion Recognition (FER) enables smart environments and robots to adapt their behaviour to a user's affective state. Translating those recognized emotions into ambient cues, such as colored lighting, can improve comfort and engagement in Ambient Assisted Living (AAL) settings.We design a FER pipeline that combines a Spatial Transformer Network for poseinvariant region focusing with a novel Multiple Self-Attention (MSA) block comprising parallel attention heads and learned fusion weights. The MSA-enhanced block is inserted into a compact VGG-style backbone trained on the FER+ dataset using weighted sampling to counteract class 1 Russo et al.imbalance. The resulting soft-max probabilities are linearly blended with prototype hues derived from a simplified Plutchik wheel to drive RGB lighting in real time.The proposed VGGFac-STN-MSA model achieves 82.54% test accuracy on FER+, outperforming a CNN baseline and the reproduced Deep-Emotion architecture. Ablation shows that MSA contributes +1% accuracy. Continuous colour blending yields smooth, intensity-aware lighting transitions in a proof-of-concept demo. Discussion: Our attention scheme is architecture-agnostic, adds minimal computational overhead, and markedly boosts FER accuracy on low-resolution faces. Coupling the probability distribution directly to the RGB gamut provides a fine-grained, perceptually meaningful channel for affect-adaptive AAL systems.

Keywords: facial emotion recognition, Spatial transformer network, Self-attention, Ambient Assisted Living, human-robot interaction

Received: 02 May 2025; Accepted: 30 Jul 2025.

Copyright: © 2025 Russo, TIBERMACINE, Randieri, Rabehi, H. Alharbi, El-kenawy and Napoli. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Imad Eddine TIBERMACINE, Sapienza University of Rome, Rome, Italy

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.